瀏覽代碼

Merge pull request #98 from ALANVF/language-scala

Alexandru Dima 4 年之前
父節點
當前提交
2845e42207
共有 3 個文件被更改,包括 403 次插入206 次删除
  1. 3 3
      src/scala/scala.contribution.ts
  2. 155 130
      src/scala/scala.test.ts
  3. 245 73
      src/scala/scala.ts

+ 3 - 3
src/scala/scala.contribution.ts

@@ -8,8 +8,8 @@ import { registerLanguage } from '../_.contribution';
 
 registerLanguage({
 	id: 'scala',
-	extensions: ['.scala', '.sc'],
-	aliases: ['Scala', 'scala'],
-	mimetypes: ['text/x-scala-source', 'text/x-scala'],
+	extensions: ['.scala', '.sc', '.sbt'],
+	aliases: ['Scala', 'scala', 'SBT', 'Sbt', 'sbt', 'Dotty', 'dotty'],
+	mimetypes: ['text/x-scala-source', 'text/x-scala', 'text/x-sbt', 'text/x-dotty'],
 	loader: () => import('./scala')
 });

+ 155 - 130
src/scala/scala.test.ts

@@ -8,6 +8,20 @@
 import { testTokenization } from '../test/testRunner';
 
 testTokenization('scala', [
+
+	[{
+		line: 'var a = 1',
+		tokens: [
+			{startIndex: 0, type: 'keyword.scala'},
+			{startIndex: 3, type: 'white.scala'},
+			{startIndex: 4, type: 'variable.scala'},
+			{startIndex: 5, type: 'white.scala'},
+			{startIndex: 6, type: 'operator.scala'},
+			{startIndex: 7, type: 'white.scala'},
+			{startIndex: 8, type: 'number.scala'}
+		]
+	}],
+
 	// Comments - single line
 	[{
 		line: '//',
@@ -19,7 +33,7 @@ testTokenization('scala', [
 	[{
 		line: '    // a comment',
 		tokens: [
-			{ startIndex: 0, type: '' },
+			{ startIndex: 0, type: 'white.scala' },
 			{ startIndex: 4, type: 'comment.scala' }
 		]
 	}],
@@ -28,9 +42,7 @@ testTokenization('scala', [
 	[{
 		line: '/* //*/ a',
 		tokens: [
-			{ startIndex: 0, type: 'comment.scala' },
-			{ startIndex: 7, type: '' },
-			{ startIndex: 8, type: 'identifier.scala' }
+			{ startIndex: 0, type: 'comment.scala' }
 		]
 	}],
 
@@ -51,11 +63,11 @@ testTokenization('scala', [
 	[{
 		line: '/almost a comment',
 		tokens: [
-			{ startIndex: 0, type: 'delimiter.scala' },
+			{ startIndex: 0, type: 'operator.scala' },
 			{ startIndex: 1, type: 'identifier.scala' },
-			{ startIndex: 7, type: '' },
+			{ startIndex: 7, type: 'white.scala' },
 			{ startIndex: 8, type: 'identifier.scala' },
-			{ startIndex: 9, type: '' },
+			{ startIndex: 9, type: 'white.scala' },
 			{ startIndex: 10, type: 'identifier.scala' }
 		]
 	}],
@@ -64,12 +76,12 @@ testTokenization('scala', [
 		line: '1 / 2; /* comment',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 1, type: '' },
-			{ startIndex: 2, type: 'delimiter.scala' },
-			{ startIndex: 3, type: '' },
+			{ startIndex: 1, type: 'white.scala' },
+			{ startIndex: 2, type: 'operator.scala' },
+			{ startIndex: 3, type: 'white.scala' },
 			{ startIndex: 4, type: 'number.scala' },
 			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
+			{ startIndex: 6, type: 'white.scala' },
 			{ startIndex: 7, type: 'comment.scala' }
 		]
 	}],
@@ -77,18 +89,18 @@ testTokenization('scala', [
 	[{
 		line: 'val x: Int = 1; // my comment // is a nice one',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'keyword.Int.scala' },
-			{ startIndex: 10, type: '' },
-			{ startIndex: 11, type: 'delimiter.scala' },
-			{ startIndex: 12, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 10, type: 'white.scala' },
+			{ startIndex: 11, type: 'operator.scala' },
+			{ startIndex: 12, type: 'white.scala' },
 			{ startIndex: 13, type: 'number.scala' },
 			{ startIndex: 14, type: 'delimiter.scala' },
-			{ startIndex: 15, type: '' },
+			{ startIndex: 15, type: 'white.scala' },
 			{ startIndex: 16, type: 'comment.scala' }
 		]
 	}],
@@ -104,17 +116,17 @@ testTokenization('scala', [
 	[{
 		line: 'val x: Int = /* a simple comment */ 1;',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'keyword.Int.scala' },
-			{ startIndex: 10, type: '' },
-			{ startIndex: 11, type: 'delimiter.scala' },
-			{ startIndex: 12, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 10, type: 'white.scala' },
+			{ startIndex: 11, type: 'operator.scala' },
+			{ startIndex: 12, type: 'white.scala' },
 			{ startIndex: 13, type: 'comment.scala' },
-			{ startIndex: 35, type: '' },
+			{ startIndex: 35, type: 'white.scala' },
 			{ startIndex: 36, type: 'number.scala' },
 			{ startIndex: 37, type: 'delimiter.scala' }
 		]
@@ -123,20 +135,21 @@ testTokenization('scala', [
 	[{
 		line: 'val x: Int = /* a simple comment */ 1; */',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'keyword.Int.scala' },
-			{ startIndex: 10, type: '' },
-			{ startIndex: 11, type: 'delimiter.scala' },
-			{ startIndex: 12, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 10, type: 'white.scala' },
+			{ startIndex: 11, type: 'operator.scala' },
+			{ startIndex: 12, type: 'white.scala' },
 			{ startIndex: 13, type: 'comment.scala' },
-			{ startIndex: 35, type: '' },
+			{ startIndex: 35, type: 'white.scala' },
 			{ startIndex: 36, type: 'number.scala' },
 			{ startIndex: 37, type: 'delimiter.scala' },
-			{ startIndex: 38, type: '' }
+			{ startIndex: 38, type: 'white.scala' },
+			{ startIndex: 39, type: 'operator.scala' },
 		]
 	}],
 
@@ -144,9 +157,9 @@ testTokenization('scala', [
 		line: 'x = /**/;',
 		tokens: [
 			{ startIndex: 0, type: 'identifier.scala' },
-			{ startIndex: 1, type: '' },
-			{ startIndex: 2, type: 'delimiter.scala' },
-			{ startIndex: 3, type: '' },
+			{ startIndex: 1, type: 'white.scala' },
+			{ startIndex: 2, type: 'operator.scala' },
+			{ startIndex: 3, type: 'white.scala' },
 			{ startIndex: 4, type: 'comment.scala' },
 			{ startIndex: 8, type: 'delimiter.scala' }
 		]
@@ -156,9 +169,9 @@ testTokenization('scala', [
 		line: 'x = /*/;',
 		tokens: [
 			{ startIndex: 0, type: 'identifier.scala' },
-			{ startIndex: 1, type: '' },
-			{ startIndex: 2, type: 'delimiter.scala' },
-			{ startIndex: 3, type: '' },
+			{ startIndex: 1, type: 'white.scala' },
+			{ startIndex: 2, type: 'operator.scala' },
+			{ startIndex: 3, type: 'white.scala' },
 			{ startIndex: 4, type: 'comment.scala' }
 		]
 	}],
@@ -184,15 +197,15 @@ testTokenization('scala', [
 	[{
 		line: 'val x: Int = /* start a comment',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'keyword.Int.scala' },
-			{ startIndex: 10, type: '' },
-			{ startIndex: 11, type: 'delimiter.scala' },
-			{ startIndex: 12, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 10, type: 'white.scala' },
+			{ startIndex: 11, type: 'operator.scala' },
+			{ startIndex: 12, type: 'white.scala' },
 			{ startIndex: 13, type: 'comment.scala' },
 		]
 	}, {
@@ -204,7 +217,7 @@ testTokenization('scala', [
 		line: 'and end it */ 2;',
 		tokens: [
 			{ startIndex: 0, type: 'comment.scala' },
-			{ startIndex: 13, type: '' },
+			{ startIndex: 13, type: 'white.scala' },
 			{ startIndex: 14, type: 'number.scala' },
 			{ startIndex: 15, type: 'delimiter.scala' }
 		]
@@ -214,17 +227,17 @@ testTokenization('scala', [
 	[{
 		line: '/** start of Scala Doc',
 		tokens: [
-			{ startIndex: 0, type: 'comment.doc.scala' }
+			{ startIndex: 0, type: 'comment.scala' }
 		]
 	}, {
 		line: 'a comment between without a star',
 		tokens: [
-			{ startIndex: 0, type: 'comment.doc.scala' }
+			{ startIndex: 0, type: 'comment.scala' }
 		]
 	}, {
 		line: 'end of multiline comment*/',
 		tokens: [
-			{ startIndex: 0, type: 'comment.doc.scala' }
+			{ startIndex: 0, type: 'comment.scala' }
 		]
 	}],
 
@@ -232,37 +245,37 @@ testTokenization('scala', [
 	[{
 		line: 'package test; object Program { def main(args: Array[String]): Unit = {} }',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.package.scala' },
-			{ startIndex: 7, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 7, type: 'white.scala' },
 			{ startIndex: 8, type: 'identifier.scala' },
 			{ startIndex: 12, type: 'delimiter.scala' },
-			{ startIndex: 13, type: '' },
-			{ startIndex: 14, type: 'keyword.object.scala' },
-			{ startIndex: 20, type: '' },
-			{ startIndex: 21, type: 'identifier.scala' },
-			{ startIndex: 28, type: '' },
+			{ startIndex: 13, type: 'white.scala' },
+			{ startIndex: 14, type: 'keyword.scala' },
+			{ startIndex: 20, type: 'white.scala' },
+			{ startIndex: 21, type: 'type.scala' },
+			{ startIndex: 28, type: 'white.scala' },
 			{ startIndex: 29, type: 'delimiter.curly.scala' },
-			{ startIndex: 30, type: '' },
-			{ startIndex: 31, type: 'keyword.def.scala' },
-			{ startIndex: 34, type: '' },
+			{ startIndex: 30, type: 'white.scala' },
+			{ startIndex: 31, type: 'keyword.scala' },
+			{ startIndex: 34, type: 'white.scala' },
 			{ startIndex: 35, type: 'identifier.scala' },
 			{ startIndex: 39, type: 'delimiter.parenthesis.scala' },
-			{ startIndex: 40, type: 'identifier.scala' },
-			{ startIndex: 44, type: 'delimiter.scala' },
-			{ startIndex: 45, type: '' },
-			{ startIndex: 46, type: 'identifier.scala' },
-			{ startIndex: 51, type: 'delimiter.square.scala' },
-			{ startIndex: 52, type: 'identifier.scala' },
-			{ startIndex: 58, type: 'delimiter.square.scala' },
+			{ startIndex: 40, type: 'variable.scala' },
+			{ startIndex: 44, type: 'operator.scala' },
+			{ startIndex: 45, type: 'white.scala' },
+			{ startIndex: 46, type: 'type.scala' },
+			{ startIndex: 51, type: 'operator.square.scala' },
+			{ startIndex: 52, type: 'type.scala' },
+			{ startIndex: 58, type: 'operator.square.scala' },
 			{ startIndex: 59, type: 'delimiter.parenthesis.scala' },
-			{ startIndex: 60, type: 'delimiter.scala' },
-			{ startIndex: 61, type: '' },
-			{ startIndex: 62, type: 'keyword.Unit.scala' },
-			{ startIndex: 66, type: '' },
-			{ startIndex: 67, type: 'delimiter.scala' },
-			{ startIndex: 68, type: '' },
+			{ startIndex: 60, type: 'operator.scala' },
+			{ startIndex: 61, type: 'white.scala' },
+			{ startIndex: 62, type: 'type.scala' },
+			{ startIndex: 66, type: 'white.scala' },
+			{ startIndex: 67, type: 'operator.scala' },
+			{ startIndex: 68, type: 'white.scala' },
 			{ startIndex: 69, type: 'delimiter.curly.scala' },
-			{ startIndex: 71, type: '' },
+			{ startIndex: 71, type: 'white.scala' },
 			{ startIndex: 72, type: 'delimiter.curly.scala' }
 		]
 	}],
@@ -455,7 +468,7 @@ testTokenization('scala', [
 		line: '3_.1415F',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 1, type: 'identifier.scala' },
+			{ startIndex: 1, type: 'keyword.scala' },
 			{ startIndex: 2, type: 'delimiter.scala' },
 			{ startIndex: 3, type: 'number.float.scala' }
 		]
@@ -465,7 +478,7 @@ testTokenization('scala', [
 		line: '3._1415F',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 1, type: 'delimiter.scala' },
+			{ startIndex: 1, type: 'operator.scala' },
 			{ startIndex: 2, type: 'identifier.scala' }
 		]
 	}],
@@ -482,7 +495,7 @@ testTokenization('scala', [
 		line: '52_',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 2, type: 'identifier.scala' }
+			{ startIndex: 2, type: 'keyword.scala' }
 		]
 	}],
 
@@ -506,7 +519,7 @@ testTokenization('scala', [
 		line: '0x52_',
 		tokens: [
 			{ startIndex: 0, type: 'number.hex.scala' },
-			{ startIndex: 4, type: 'identifier.scala' }
+			{ startIndex: 4, type: 'keyword.scala' } // TODO
 		]
 	}],
 
@@ -514,7 +527,7 @@ testTokenization('scala', [
 		line: '23.5L',
 		tokens: [
 			{ startIndex: 0, type: 'number.float.scala' },
-			{ startIndex: 4, type: 'identifier.scala' }
+			{ startIndex: 4, type: 'type.scala' }
 		]
 	}],
 
@@ -522,7 +535,7 @@ testTokenization('scala', [
 		line: '0+0',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 1, type: 'delimiter.scala' },
+			{ startIndex: 1, type: 'operator.scala' },
 			{ startIndex: 2, type: 'number.scala' }
 		]
 	}],
@@ -531,7 +544,7 @@ testTokenization('scala', [
 		line: '100+10',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 3, type: 'delimiter.scala' },
+			{ startIndex: 3, type: 'operator.scala' },
 			{ startIndex: 4, type: 'number.scala' }
 		]
 	}],
@@ -540,9 +553,9 @@ testTokenization('scala', [
 		line: '0 + 0',
 		tokens: [
 			{ startIndex: 0, type: 'number.scala' },
-			{ startIndex: 1, type: '' },
-			{ startIndex: 2, type: 'delimiter.scala' },
-			{ startIndex: 3, type: '' },
+			{ startIndex: 1, type: 'white.scala' },
+			{ startIndex: 2, type: 'operator.scala' },
+			{ startIndex: 3, type: 'white.scala' },
 			{ startIndex: 4, type: 'number.scala' }
 		]
 	}],
@@ -551,16 +564,18 @@ testTokenization('scala', [
 	[{
 		line: 'val s: String = "I\'m a Scala String";',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'identifier.scala' },
-			{ startIndex: 13, type: '' },
-			{ startIndex: 14, type: 'delimiter.scala' },
-			{ startIndex: 15, type: '' },
-			{ startIndex: 16, type: 'string.scala' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 13, type: 'white.scala' },
+			{ startIndex: 14, type: 'operator.scala' },
+			{ startIndex: 15, type: 'white.scala' },
+			{ startIndex: 16, type: 'string.quote.scala' },
+			{ startIndex: 17, type: 'string.scala' },
+			{ startIndex: 35, type: 'string.quote.scala' },
 			{ startIndex: 36, type: 'delimiter.scala' }
 		]
 	}],
@@ -568,21 +583,25 @@ testTokenization('scala', [
 	[{
 		line: 'val s: String = "concatenated" + " String" ;',
 		tokens: [
-			{ startIndex: 0, type: 'keyword.val.scala' },
-			{ startIndex: 3, type: '' },
-			{ startIndex: 4, type: 'identifier.scala' },
-			{ startIndex: 5, type: 'delimiter.scala' },
-			{ startIndex: 6, type: '' },
-			{ startIndex: 7, type: 'identifier.scala' },
-			{ startIndex: 13, type: '' },
-			{ startIndex: 14, type: 'delimiter.scala' },
-			{ startIndex: 15, type: '' },
-			{ startIndex: 16, type: 'string.scala' },
-			{ startIndex: 30, type: '' },
-			{ startIndex: 31, type: 'delimiter.scala' },
-			{ startIndex: 32, type: '' },
-			{ startIndex: 33, type: 'string.scala' },
-			{ startIndex: 42, type: '' },
+			{ startIndex: 0, type: 'keyword.scala' },
+			{ startIndex: 3, type: 'white.scala' },
+			{ startIndex: 4, type: 'variable.scala' },
+			{ startIndex: 5, type: 'operator.scala' },
+			{ startIndex: 6, type: 'white.scala' },
+			{ startIndex: 7, type: 'type.scala' },
+			{ startIndex: 13, type: 'white.scala' },
+			{ startIndex: 14, type: 'operator.scala' },
+			{ startIndex: 15, type: 'white.scala' },
+			{ startIndex: 16, type: 'string.quote.scala' },
+			{ startIndex: 17, type: 'string.scala' },
+			{ startIndex: 29, type: 'string.quote.scala' },
+			{ startIndex: 30, type: 'white.scala' },
+			{ startIndex: 31, type: 'operator.scala' },
+			{ startIndex: 32, type: 'white.scala' },
+			{ startIndex: 33, type: 'string.quote.scala' },
+			{ startIndex: 34, type: 'string.scala' },
+			{ startIndex: 41, type: 'string.quote.scala' },
+			{ startIndex: 42, type: 'white.scala' },
 			{ startIndex: 43, type: 'delimiter.scala' }
 		]
 	}],
@@ -590,25 +609,30 @@ testTokenization('scala', [
 	[{
 		line: '"quote in a string"',
 		tokens: [
-			{ startIndex: 0, type: 'string.scala' }
+			{ startIndex: 0, type: 'string.quote.scala' },
+			{ startIndex: 1, type: 'string.scala' },
+			{ startIndex: 18, type: 'string.quote.scala' }
 		]
 	}],
 
 	[{
 		line: '"escaping \\"quotes\\" is cool"',
 		tokens: [
-			{ startIndex: 0, type: 'string.scala' },
+			{ startIndex: 0, type: 'string.quote.scala' },
+			{ startIndex: 1, type: 'string.scala' },
 			{ startIndex: 10, type: 'string.escape.scala' },
 			{ startIndex: 12, type: 'string.scala' },
 			{ startIndex: 18, type: 'string.escape.scala' },
-			{ startIndex: 20, type: 'string.scala' }
+			{ startIndex: 20, type: 'string.scala' },
+			{ startIndex: 28, type: 'string.quote.scala' },
 		]
 	}],
 
 	[{
 		line: '"\\"',
 		tokens: [
-			{ startIndex: 0, type: 'string.invalid.scala' }
+			{ startIndex: 0, type: 'string.quote.scala' },
+			{ startIndex: 1, type: 'string.escape.scala' }
 		]
 	}],
 
@@ -616,7 +640,7 @@ testTokenization('scala', [
 	[{
 		line: '@',
 		tokens: [
-			{ startIndex: 0, type: '' }
+			{ startIndex: 0, type: 'operator.scala' }
 		]
 	}],
 
@@ -632,18 +656,19 @@ testTokenization('scala', [
 		tokens: [
 			{ startIndex: 0, type: 'annotation.scala' },
 			{ startIndex: 7, type: 'delimiter.parenthesis.scala' },
-			{ startIndex: 8, type: 'string.scala' },
+			{ startIndex: 8, type: 'string.quote.scala' },
+			{ startIndex: 9, type: 'string.scala' },
+			{ startIndex: 19, type: 'string.quote.scala' },
 			{ startIndex: 20, type: 'delimiter.parenthesis.scala' }
 		]
 	}],
 
 	[{
-		line: '@ AnnotationWithKeywordAfter private',
+		line: '@AnnotationWithKeywordAfter private',
 		tokens: [
 			{ startIndex: 0, type: 'annotation.scala' },
-			{ startIndex: 28, type: '' },
-			{ startIndex: 29, type: 'keyword.private.scala' }
+			{ startIndex: 27, type: 'white.scala' },
+			{ startIndex: 28, type: 'keyword.modifier.scala' }
 		]
 	}]
 ]);
-

+ 245 - 73
src/scala/scala.ts

@@ -9,8 +9,13 @@ import IRichLanguageConfiguration = monaco.languages.LanguageConfiguration;
 import ILanguage = monaco.languages.IMonarchLanguage;
 
 export const conf: IRichLanguageConfiguration = {
-	// the default separators except `@$`
-	wordPattern: /(-?\d*\.\d\w*)|([^\`\~\!\#\%\^\&\*\(\)\-\=\+\[\{\]\}\\\|\;\:\'\"\,\.\<\>\/\?\s]+)/g,
+	/*
+	 * `...` is allowed as an identifier.
+	 * $ is allowed in identifiers.
+	 * unary_<op> is allowed as an identifier.
+	 * <name>_= is allowed as an identifier.
+	 */
+	wordPattern: /(unary_[@~!#%^&*()\-=+\\|:<>\/?]+)|([a-zA-Z_$][\w$]*?_=)|(`[^`]+`)|([a-zA-Z_$][\w$]*)/g,
 	comments: {
 		lineComment: '//',
 		blockComment: ['/*', '*/'],
@@ -33,121 +38,288 @@ export const conf: IRichLanguageConfiguration = {
 		{ open: '(', close: ')' },
 		{ open: '"', close: '"' },
 		{ open: '\'', close: '\'' },
-		{ open: '<', close: '>' },
-	]
+	],
+	folding: {
+		markers: {
+			start: new RegExp("^\\s*//\\s*(?:(?:#?region\\b)|(?:<editor-fold\\b))"),
+			end: new RegExp("^\\s*//\\s*(?:(?:#?endregion\\b)|(?:</editor-fold>))")
+		}
+	}
 };
 
 export const language = <ILanguage>{
-	defaultToken: '',
 	tokenPostfix: '.scala',
 
-	// Reserved words are defined in lexical syntax at:
-	// https://www.scala-lang.org/files/archive/spec/2.13/01-lexical-syntax.html
+	// We can't easily add everything from Dotty, but we can at least add some of its keywords
 	keywords: [
-		'abstract', 'case', 'catch', 'class', 'def',
-		'do', 'else', 'extends', 'false', 'final',
-		'finally', 'for', 'forSome', 'if', 'implicit',
-		'import', 'lazy', 'macro', 'match', 'new',
-		'null', 'object', 'override', 'package', 'private',
-		'protected', 'return', 'sealed', 'super', 'this',
-		'throw', 'trait', 'try', 'true', 'type',
-		'val', 'var', 'while', 'with', 'yield',
-		'Double', 'Float', 'Long', 'Int', 'Short',
-		'Byte', 'Char', 'Unit', 'Boolean'
+		'asInstanceOf',
+		'catch', 'class', 'classOf',
+		'def', 'do',
+		'else', 'extends',
+		'finally', 'for', 'foreach', 'forSome',
+		'if', 'import', 'isInstanceOf',
+		'macro', 'match',
+		'new',
+		'object',
+		'package',
+		'return',
+		'throw', 'trait', 'try', 'type',
+		'until',
+		'val', 'var',
+		'while', 'with',
+		'yield',
+
+		// Dotty-specific:
+		'given', 'enum', 'then'
 	],
 
-	// Extended the operators that are defined in lexical syntax at:
-	// https://www.scala-lang.org/files/archive/spec/2.13/01-lexical-syntax.html
-	operators: [
-		'=', '>', '<', '!', '~', '?', ':',
-		'==', '<=', '>=', '!=', '&&', '||', '++', '--',
-		'+', '-', '*', '/', '&', '|', '^', '%', '<<',
-		'>>', '>>>', '+=', '-=', '*=', '/=', '&=', '|=',
-		'^=', '%=', '<<=', '>>=', '>>>=',
-		'_', '=>', '<-', '<:', '<%', '>:', '#', '@'
+	// Dotty-specific:
+	softKeywords: [
+		'as', 'export', 'extension', 'end',
+		'derives', 'on'
 	],
 
+	constants: [
+		'true', 'false', 'null',
+		'this', 'super'
+	],
+
+	modifiers: [
+		'abstract', 'final', 'implicit', 'lazy', 'override',
+		'private', 'protected', 'sealed'
+	],
+
+	// Dotty-specific:
+	softModifiers: [
+		'inline', 'opaque', 'open', 'transparent', 'using'
+	],
+
+	name: /(?:[a-z_$][\w$]*|`[^`]+`)/,
+	type: /(?:[A-Z][\w$]*)/,
+
 	// we include these common regular expressions
-	symbols: /[=><!~?:&|+\-*\/\^%]+/,
-	escapes: /\\(?:[abfnrtv\\"']|x[0-9A-Fa-f]{1,4}|u[0-9A-Fa-f]{4}|U[0-9A-Fa-f]{8})/,
+	symbols: /[=><!~?:&|+\-*\/^\\%@#]+/,
 	digits: /\d+(_+\d+)*/,
 	hexdigits: /[[0-9a-fA-F]+(_+[0-9a-fA-F]+)*/,
 
+	// C# style strings
+	escapes: /\\(?:[btnfr\\"']|x[0-9A-Fa-f]{1,4}|u[0-9A-Fa-f]{4}|U[0-9A-Fa-f]{8})/,
+
+	fstring_conv: /[bBhHsScCdoxXeEfgGaAt]|[Tn](?:[HIklMSLNpzZsQ]|[BbhAaCYyjmde]|[RTrDFC])/,
+
 	// The main tokenizer for our languages
 	tokenizer: {
 		root: [
+			// strings
+			[/\braw"""/, {token: 'string.quote', bracket: '@open', next: '@rawstringt'}],
+			[/\braw"/, {token: 'string.quote', bracket: '@open', next: '@rawstring'}],
+
+			[/\bs"""/, {token: 'string.quote', bracket: '@open', next: '@sstringt'}],
+			[/\bs"/, {token: 'string.quote', bracket: '@open', next: '@sstring'}],
+
+			[/\bf""""/, {token: 'string.quote', bracket: '@open', next: '@fstringt'}],
+			[/\bf"/, {token: 'string.quote', bracket: '@open', next: '@fstring'}],
+
+			[/"""/, {token: 'string.quote', bracket: '@open', next: '@stringt'}],
+			[/"/, {token: 'string.quote', bracket: '@open', next: '@string'}],
+
+			// numbers
+			[/(@digits)[eE]([\-+]?(@digits))?[fFdD]?/, 'number.float', '@allowMethod'],
+			[/(@digits)\.(@digits)([eE][\-+]?(@digits))?[fFdD]?/, 'number.float', '@allowMethod'],
+			[/0[xX](@hexdigits)[Ll]?/, 'number.hex', '@allowMethod'],
+			[/(@digits)[fFdD]/, 'number.float', '@allowMethod'],
+			[/(@digits)[lL]?/, 'number', '@allowMethod'],
+
+			[/\b_\*/, 'key'],
+			[/\b(_)\b/, 'keyword', '@allowMethod'],
+
 			// identifiers and keywords
-			[/[a-zA-Z_$][\w$]*/, {
-				cases: {
-					'@keywords': { token: 'keyword.$0' },
-					'@default': 'identifier'
-				}
-			}],
+			[/\bimport\b/, 'keyword', '@import'],
+			[/\b(case)([ \t]+)(class)\b/, ['keyword.modifier', 'white', 'keyword']],
+			[/\bcase\b/, 'keyword', '@case'],
+			[/\bva[lr]\b/, 'keyword', '@vardef'],
+			[/\b(def)([ \t]+)((?:unary_)?@symbols|@name(?:_=)|@name)/, ['keyword', 'white', 'identifier']],
+			[/@name(?=[ \t]*:(?!:))/, 'variable'],
+			[/(\.)(@name|@symbols)/, ['operator', {token: '@rematch', next: '@allowMethod'}]],
+			[/([{(])(\s*)(@name(?=\s*=>))/, ['@brackets', 'white', 'variable']],
+			[/@name/, {cases: {
+				'@keywords': 'keyword',
+				'@softKeywords': 'keyword',
+				'@modifiers': 'keyword.modifier',
+				'@softModifiers': 'keyword.modifier',
+				'@constants': {token: 'constant', next: '@allowMethod'},
+				'@default': {token: 'identifier', next: '@allowMethod'}
+			}}],
+			[/@type/, 'type', '@allowMethod'],
 
 			// whitespace
-			{ include: '@whitespace' },
-
-			// delimiters and operators
-			[/[{}()\[\]]/, '@brackets'],
-			[/[<>](?!@symbols)/, '@brackets'],
-			[/@symbols/, {
-				cases: {
-					'@operators': 'delimiter',
-					'@default': ''
-				}
-			}],
+			{include: '@whitespace'},
 
 			// @ annotations.
-			[/@\s*[a-zA-Z_\$][\w\$]*/, 'annotation'],
+			[/@[a-zA-Z_$][\w$]*(?:\.[a-zA-Z_$][\w$]*)*/, 'annotation'],
 
-			// numbers
-			[/(@digits)[eE]([\-+]?(@digits))?[fFdD]?/, 'number.float'],
-			[/(@digits)\.(@digits)([eE][\-+]?(@digits))?[fFdD]?/, 'number.float'],
-			[/0[xX](@hexdigits)[Ll]?/, 'number.hex'],
-			[/(@digits)[fFdD]/, 'number.float'],
-			[/(@digits)[lL]?/, 'number'],
+			// delimiters and operators
+			[/[{(]/, '@brackets'],
+			[/[})]/, '@brackets', '@allowMethod'],
+			[/\[/, 'operator.square'],
+			[/](?!\s*(?:va[rl]|def|type)\b)/, 'operator.square', '@allowMethod'],
+			[/]/, 'operator.square'],
+			[/([=-]>|<-|>:|<:|:>|<%)(?=[\s\w()[\]{},\."'`])/, 'keyword'],
+			[/@symbols/, 'operator'],
 
 			// delimiter: after number because of .\d floats
-			[/[;,.]/, 'delimiter'],
+			[/[;,\.]/, 'delimiter'],
 
-			// strings
-			[/"([^"\\]|\\.)*$/, 'string.invalid'],  // non-terminated string
-			[/"/, 'string', '@string'],
+			// symbols
+			[/'[a-zA-Z$][\w$]*(?!')/, 'attribute.name'],
 
 			// characters
-			[/'[^\\']'/, 'string'],
-			[/(')(@escapes)(')/, ['string', 'string.escape', 'string']],
+			[/'[^\\']'/, 'string', '@allowMethod'],
+			[/(')(@escapes)(')/, ['string', 'string.escape', {token: 'string', next: '@allowMethod'}]],
 			[/'/, 'string.invalid']
 		],
 
-		whitespace: [
-			[/[ \t\r\n]+/, ''],
-			[/\/\*\*(?!\/)/, 'comment.doc', '@scaladoc'],
+		import: [
+			[/;/, 'delimiter', '@pop'],
+			[/^|$/, '', '@pop'],
+			[/[ \t]+/, 'white'],
+			[/[\n\r]+/, 'white', '@pop'],
 			[/\/\*/, 'comment', '@comment'],
-			[/\/\/.*$/, 'comment'],
+			[/@name|@type/, 'type'],
+			[/[(){}]/, '@brackets'],
+			[/[[\]]/, 'operator.square'],
+			[/[\.,]/, 'delimiter'],
+		],
+
+		allowMethod: [
+			[/^|$/, '', '@pop'],
+			[/[ \t]+/, 'white'],
+			[/[\n\r]+/, 'white', '@pop'],
+			[/\/\*/, 'comment', '@comment'],
+			[/(?==>[\s\w([{])/, 'keyword', '@pop'],
+			[/(@name|@symbols)(?=[ \t]*[[({"'`]|[ \t]+(?:[+-]?\.?\d|\w))/, {
+				cases: {
+				'@keywords': {token: 'keyword', next: '@pop'},
+				'->|<-|>:|<:|<%': {token: 'keyword', next: '@pop'},
+				'@default': {token: '@rematch', next: '@pop'}
+				}
+			}],
+			["", "", "@pop"]
 		],
 
 		comment: [
 			[/[^\/*]+/, 'comment'],
-			// [/\/\*/, 'comment', '@push' ],    // nested comment not allowed :-(
-			// [/\/\*/,    'comment.invalid' ],    // this breaks block comments in the shape of /* //*/
+			[/\/\*/, 'comment', '@push'], // nested comment
 			[/\*\//, 'comment', '@pop'],
 			[/[\/*]/, 'comment']
 		],
-		//Identical copy of comment above, except for the addition of .doc
-		scaladoc: [
-			[/[^\/*]+/, 'comment.doc'],
-			// [/\/\*/, 'comment.doc', '@push' ],    // nested comment not allowed :-(
-			[/\/\*/, 'comment.doc.invalid'],
-			[/\*\//, 'comment.doc', '@pop'],
-			[/[\/*]/, 'comment.doc']
+
+		case: [
+			[/\b_\*/, 'key'],
+			[/\b(_|true|false|null|this|super)\b/, 'keyword', '@allowMethod'],
+			[/\bif\b|=>/, 'keyword', '@pop'],
+			[/`[^`]+`/, 'identifier', '@allowMethod'],
+			[/@name/, 'variable', '@allowMethod'],
+			[/:::?|\||@(?![a-z_$])/, 'keyword'],
+			{include: '@root'}
+		],
+
+		vardef: [
+			[/\b_\*/, 'key'],
+			[/\b(_|true|false|null|this|super)\b/, 'keyword'],
+			[/@name/, 'variable'],
+			[/:::?|\||@(?![a-z_$])/, 'keyword'],
+			[/=|:(?!:)/, 'operator', '@pop'],
+			[/$/, 'white', '@pop'],
+			{include: '@root'}
 		],
 
 		string: [
-			[/[^\\"]+/, 'string'],
+			[/[^\\"\n\r]+/, 'string'],
 			[/@escapes/, 'string.escape'],
 			[/\\./, 'string.escape.invalid'],
-			[/"/, 'string', '@pop']
+			[/"/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+		],
+
+		stringt: [
+			[/[^\\"\n\r]+/,	'string'],
+			[/@escapes/, 'string.escape'],
+			[/\\./, 'string.escape.invalid'],
+			[/"(?=""")/, 'string'],
+			[/"""/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/"/, 'string']
+		],
+
+		fstring: [
+			[/@escapes/, 'string.escape'],
+			[/"/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/\$\$/, 'string'],
+			[/(\$)([a-z_]\w*)/, ['operator', 'identifier']],
+			[/\$\{/, 'operator', '@interp'],
+			[/%%/, 'string'],
+			[/(%)([\-#+ 0,(])(\d+|\.\d+|\d+\.\d+)(@fstring_conv)/, ['metatag', 'keyword.modifier', 'number', 'metatag']],
+			[/(%)(\d+|\.\d+|\d+\.\d+)(@fstring_conv)/, ['metatag', 'number', 'metatag']],
+			[/(%)([\-#+ 0,(])(@fstring_conv)/, ['metatag', 'keyword.modifier', 'metatag']],
+			[/(%)(@fstring_conv)/, ['metatag', 'metatag']],
+			[/./, 'string']
+		],
+
+		fstringt: [
+			[/@escapes/, 'string.escape'],
+			[/"(?=""")/, 'string'],
+			[/"""/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/\$\$/, 'string'],
+			[/(\$)([a-z_]\w*)/, ['operator', 'identifier']],
+			[/\$\{/, 'operator', '@interp'],
+			[/%%/, 'string'],
+			[/(%)([\-#+ 0,(])(\d+|\.\d+|\d+\.\d+)(@fstring_conv)/, ['metatag', 'keyword.modifier', 'number', 'metatag']],
+			[/(%)(\d+|\.\d+|\d+\.\d+)(@fstring_conv)/, ['metatag', 'number', 'metatag']],
+			[/(%)([\-#+ 0,(])(@fstring_conv)/, ['metatag', 'keyword.modifier', 'metatag']],
+			[/(%)(@fstring_conv)/, ['metatag', 'metatag']],
+			[/./, 'string']
+		],
+
+		sstring: [
+			[/@escapes/, 'string.escape'],
+			[/"/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/\$\$/, 'string'],
+			[/(\$)([a-z_]\w*)/, ['operator', 'identifier']],
+			[/\$\{/, 'operator', '@interp'],
+			[/./, 'string']
+		],
+
+		sstringt: [
+			[/@escapes/, 'string.escape'],
+			[/"(?=""")/, 'string'],
+			[/"""/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/\$\$/, 'string'],
+			[/(\$)([a-z_]\w*)/, ['operator', 'identifier']],
+			[/\$\{/, 'operator', '@interp'],
+			[/./, 'string']
+		],
+
+		interp: [
+			[/{/, 'operator', '@push'],
+			[/}/, 'operator', '@pop'],
+			{include: '@root'}
+		],
+
+		rawstring: [
+			[/[^"]/, 'string'],
+			[/"/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}]
+		],
+
+		rawstringt: [
+			[/[^"]/, 'string'],
+			[/"(?=""")/, 'string'],
+			[/"""/, {token: 'string.quote', bracket: '@close', switchTo: '@allowMethod'}],
+			[/"/, 'string']
+		],
+
+		whitespace: [
+			[/[ \t\r\n]+/, 'white'],
+			[/\/\*/, 'comment', '@comment'],
+			[/\/\/.*$/, 'comment'],
 		],
 	},
 };