123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348 |
- // CodeMirror, copyright (c) by Marijn Haverbeke and others
- // Distributed under an MIT license: https://codemirror.net/LICENSE
- /**
- * Link to the project's GitHub page:
- * https://github.com/pickhardt/coffeescript-codemirror-mode
- */
- ;(function (mod) {
- if (typeof exports == 'object' && typeof module == 'object')
- // CommonJS
- mod(require('../../lib/codemirror'))
- else if (typeof define == 'function' && define.amd)
- // AMD
- define(['../../lib/codemirror'], mod)
- // Plain browser env
- else mod(CodeMirror)
- })(function (CodeMirror) {
- 'use strict'
- CodeMirror.defineMode('coffeescript', function (conf, parserConf) {
- var ERRORCLASS = 'error'
- function wordRegexp(words) {
- return new RegExp('^((' + words.join(')|(') + '))\\b')
- }
- var operators = /^(?:->|=>|\+[+=]?|-[\-=]?|\*[\*=]?|\/[\/=]?|[=!]=|<[><]?=?|>>?=?|%=?|&=?|\|=?|\^=?|\~|!|\?|(or|and|\|\||&&|\?)=)/
- var delimiters = /^(?:[()\[\]{},:`=;]|\.\.?\.?)/
- var identifiers = /^[_A-Za-z$][_A-Za-z$0-9]*/
- var atProp = /^@[_A-Za-z$][_A-Za-z$0-9]*/
- var wordOperators = wordRegexp(['and', 'or', 'not', 'is', 'isnt', 'in', 'instanceof', 'typeof'])
- var indentKeywords = ['for', 'while', 'loop', 'if', 'unless', 'else', 'switch', 'try', 'catch', 'finally', 'class']
- var commonKeywords = ['break', 'by', 'continue', 'debugger', 'delete', 'do', 'in', 'of', 'new', 'return', 'then', 'this', '@', 'throw', 'when', 'until', 'extends']
- var keywords = wordRegexp(indentKeywords.concat(commonKeywords))
- indentKeywords = wordRegexp(indentKeywords)
- var stringPrefixes = /^('{3}|\"{3}|['\"])/
- var regexPrefixes = /^(\/{3}|\/)/
- var commonConstants = ['Infinity', 'NaN', 'undefined', 'null', 'true', 'false', 'on', 'off', 'yes', 'no']
- var constants = wordRegexp(commonConstants)
- // Tokenizers
- function tokenBase(stream, state) {
- // Handle scope changes
- if (stream.sol()) {
- if (state.scope.align === null) state.scope.align = false
- var scopeOffset = state.scope.offset
- if (stream.eatSpace()) {
- var lineOffset = stream.indentation()
- if (lineOffset > scopeOffset && state.scope.type == 'coffee') {
- return 'indent'
- } else if (lineOffset < scopeOffset) {
- return 'dedent'
- }
- return null
- } else {
- if (scopeOffset > 0) {
- dedent(stream, state)
- }
- }
- }
- if (stream.eatSpace()) {
- return null
- }
- var ch = stream.peek()
- // Handle docco title comment (single line)
- if (stream.match('####')) {
- stream.skipToEnd()
- return 'comment'
- }
- // Handle multi line comments
- if (stream.match('###')) {
- state.tokenize = longComment
- return state.tokenize(stream, state)
- }
- // Single line comment
- if (ch === '#') {
- stream.skipToEnd()
- return 'comment'
- }
- // Handle number literals
- if (stream.match(/^-?[0-9\.]/, false)) {
- var floatLiteral = false
- // Floats
- if (stream.match(/^-?\d*\.\d+(e[\+\-]?\d+)?/i)) {
- floatLiteral = true
- }
- if (stream.match(/^-?\d+\.\d*/)) {
- floatLiteral = true
- }
- if (stream.match(/^-?\.\d+/)) {
- floatLiteral = true
- }
- if (floatLiteral) {
- // prevent from getting extra . on 1..
- if (stream.peek() == '.') {
- stream.backUp(1)
- }
- return 'number'
- }
- // Integers
- var intLiteral = false
- // Hex
- if (stream.match(/^-?0x[0-9a-f]+/i)) {
- intLiteral = true
- }
- // Decimal
- if (stream.match(/^-?[1-9]\d*(e[\+\-]?\d+)?/)) {
- intLiteral = true
- }
- // Zero by itself with no other piece of number.
- if (stream.match(/^-?0(?![\dx])/i)) {
- intLiteral = true
- }
- if (intLiteral) {
- return 'number'
- }
- }
- // Handle strings
- if (stream.match(stringPrefixes)) {
- state.tokenize = tokenFactory(stream.current(), false, 'string')
- return state.tokenize(stream, state)
- }
- // Handle regex literals
- if (stream.match(regexPrefixes)) {
- if (stream.current() != '/' || stream.match(/^.*\//, false)) {
- // prevent highlight of division
- state.tokenize = tokenFactory(stream.current(), true, 'string-2')
- return state.tokenize(stream, state)
- } else {
- stream.backUp(1)
- }
- }
- // Handle operators and delimiters
- if (stream.match(operators) || stream.match(wordOperators)) {
- return 'operator'
- }
- if (stream.match(delimiters)) {
- return 'punctuation'
- }
- if (stream.match(constants)) {
- return 'atom'
- }
- if (stream.match(atProp) || (state.prop && stream.match(identifiers))) {
- return 'property'
- }
- if (stream.match(keywords)) {
- return 'keyword'
- }
- if (stream.match(identifiers)) {
- return 'variable'
- }
- // Handle non-detected items
- stream.next()
- return ERRORCLASS
- }
- function tokenFactory(delimiter, singleline, outclass) {
- return function (stream, state) {
- while (!stream.eol()) {
- stream.eatWhile(/[^'"\/\\]/)
- if (stream.eat('\\')) {
- stream.next()
- if (singleline && stream.eol()) {
- return outclass
- }
- } else if (stream.match(delimiter)) {
- state.tokenize = tokenBase
- return outclass
- } else {
- stream.eat(/['"\/]/)
- }
- }
- if (singleline) {
- if (parserConf.singleLineStringErrors) {
- outclass = ERRORCLASS
- } else {
- state.tokenize = tokenBase
- }
- }
- return outclass
- }
- }
- function longComment(stream, state) {
- while (!stream.eol()) {
- stream.eatWhile(/[^#]/)
- if (stream.match('###')) {
- state.tokenize = tokenBase
- break
- }
- stream.eatWhile('#')
- }
- return 'comment'
- }
- function indent(stream, state, type) {
- type = type || 'coffee'
- var offset = 0,
- align = false,
- alignOffset = null
- for (var scope = state.scope; scope; scope = scope.prev) {
- if (scope.type === 'coffee' || scope.type == '}') {
- offset = scope.offset + conf.indentUnit
- break
- }
- }
- if (type !== 'coffee') {
- align = null
- alignOffset = stream.column() + stream.current().length
- } else if (state.scope.align) {
- state.scope.align = false
- }
- state.scope = {
- offset: offset,
- type: type,
- prev: state.scope,
- align: align,
- alignOffset: alignOffset,
- }
- }
- function dedent(stream, state) {
- if (!state.scope.prev) return
- if (state.scope.type === 'coffee') {
- var _indent = stream.indentation()
- var matched = false
- for (var scope = state.scope; scope; scope = scope.prev) {
- if (_indent === scope.offset) {
- matched = true
- break
- }
- }
- if (!matched) {
- return true
- }
- while (state.scope.prev && state.scope.offset !== _indent) {
- state.scope = state.scope.prev
- }
- return false
- } else {
- state.scope = state.scope.prev
- return false
- }
- }
- function tokenLexer(stream, state) {
- var style = state.tokenize(stream, state)
- var current = stream.current()
- // Handle scope changes.
- if (current === 'return') {
- state.dedent = true
- }
- if (((current === '->' || current === '=>') && stream.eol()) || style === 'indent') {
- indent(stream, state)
- }
- var delimiter_index = '[({'.indexOf(current)
- if (delimiter_index !== -1) {
- indent(stream, state, '])}'.slice(delimiter_index, delimiter_index + 1))
- }
- if (indentKeywords.exec(current)) {
- indent(stream, state)
- }
- if (current == 'then') {
- dedent(stream, state)
- }
- if (style === 'dedent') {
- if (dedent(stream, state)) {
- return ERRORCLASS
- }
- }
- delimiter_index = '])}'.indexOf(current)
- if (delimiter_index !== -1) {
- while (state.scope.type == 'coffee' && state.scope.prev) state.scope = state.scope.prev
- if (state.scope.type == current) state.scope = state.scope.prev
- }
- if (state.dedent && stream.eol()) {
- if (state.scope.type == 'coffee' && state.scope.prev) state.scope = state.scope.prev
- state.dedent = false
- }
- return style
- }
- var external = {
- startState: function (basecolumn) {
- return {
- tokenize: tokenBase,
- scope: { offset: basecolumn || 0, type: 'coffee', prev: null, align: false },
- prop: false,
- dedent: 0,
- }
- },
- token: function (stream, state) {
- var fillAlign = state.scope.align === null && state.scope
- if (fillAlign && stream.sol()) fillAlign.align = false
- var style = tokenLexer(stream, state)
- if (style && style != 'comment') {
- if (fillAlign) fillAlign.align = true
- state.prop = style == 'punctuation' && stream.current() == '.'
- }
- return style
- },
- indent: function (state, text) {
- if (state.tokenize != tokenBase) return 0
- var scope = state.scope
- var closer = text && '])}'.indexOf(text.charAt(0)) > -1
- if (closer) while (scope.type == 'coffee' && scope.prev) scope = scope.prev
- var closes = closer && scope.type === text.charAt(0)
- if (scope.align) return scope.alignOffset - (closes ? 1 : 0)
- else return (closes ? scope.prev : scope).offset
- },
- lineComment: '#',
- fold: 'indent',
- }
- return external
- })
- // IANA registered media type
- // https://www.iana.org/assignments/media-types/
- CodeMirror.defineMIME('application/vnd.coffeescript', 'coffeescript')
- CodeMirror.defineMIME('text/x-coffeescript', 'coffeescript')
- CodeMirror.defineMIME('text/coffeescript', 'coffeescript')
- })
|