Press n or j to go to the next uncovered block, b, p or k for the previous block.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 | 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 23x 23x 23x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 3x 3x 3x 1x 3x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 2x 1x 1x 1x 3x 3x 3x 3x 3x 3x 3x 3x 3x 3x 1x 1x 1x 1x 1x 1x 1x 1x 1x 1x 3x 1x 3x 2x 2x 3x 1x 3x 2x 2x 3x 3x 3x 3x 3x 3x 3x 1x 1x 1x | import * as LServer from "vscode-languageserver-types"; import {Position} from "../position"; import {VirtualPosition} from "../virtual_position"; import {Comment, Punctuation, StringToken, StringTemplate, StringTemplateBegin, StringTemplateEnd, StringTemplateMiddle} from "../abap/1_lexer/tokens"; import {TokenNodeRegex} from "../abap/nodes"; import * as Statements from "../abap/2_statements/statements"; import {IRegistry} from "../_iregistry"; import {ITextDocumentRange} from "./_interfaces"; import {LSPUtils} from "./_lsp_utils"; const SOURCE_ABAP = "source.abap"; const BLOCK_ABAP = "storage.type.block.abap"; interface Token { line: number, startChar: number, length: number, tokenType: string, tokenModifiers: string[], } export class SemanticHighlighting { private readonly reg: IRegistry; private static readonly tokenTypes: string[] = []; private static tokenTypeMap: {[name: string]: number}; public constructor(reg: IRegistry) { this.reg = reg; SemanticHighlighting.initLegend(); } public static semanticTokensLegend(): LServer.SemanticTokensLegend { // https://code.visualstudio.com/api/language-extensions/semantic-highlight-guide#semantic-token-scope-map // https://microsoft.github.io/language-server-protocol/specifications/specification-3-17/#semanticTokenTypes this.initLegend(); return { tokenTypes: SemanticHighlighting.tokenTypes, tokenModifiers: [], }; } private static initLegend() { if (SemanticHighlighting.tokenTypes.length === 0) { SemanticHighlighting.tokenTypeMap = {}; SemanticHighlighting.tokenTypeMap[SOURCE_ABAP] = SemanticHighlighting.tokenTypes.length; SemanticHighlighting.tokenTypes.push(SOURCE_ABAP); SemanticHighlighting.tokenTypeMap[BLOCK_ABAP] = SemanticHighlighting.tokenTypes.length; SemanticHighlighting.tokenTypes.push(BLOCK_ABAP); for (const t in LServer.SemanticTokenTypes) { SemanticHighlighting.tokenTypeMap[t] = SemanticHighlighting.tokenTypes.length; SemanticHighlighting.tokenTypes.push(t); } } } // https://microsoft.github.io/language-server-protocol/specifications/specification-3-17/#textDocument_semanticTokens public semanticTokensRange(range: ITextDocumentRange): LServer.SemanticTokens { const file = LSPUtils.getABAPFile(this.reg, range.textDocument.uri); if (file === undefined) { return {data: []}; } const rangeStartPosition = new Position(range.start.line + 1, range.start.character + 1); const rangeEndPosition = new Position(range.end.line + 1, range.end.character + 1); const tokens: Token[] = []; for (const s of file.getStatements()) { if (s.getFirstToken().getStart() instanceof VirtualPosition) { continue; } else if (s.getFirstToken().getStart().isAfter(rangeEndPosition)) { break; } else if (s.getLastToken().getEnd().isBefore(rangeStartPosition)) { continue; } const statementInstance = s.get(); for (const t of s.getTokenNodes()) { const tokenInstance = t.get(); let tokenType: string = LServer.SemanticTokenTypes.keyword; if (tokenInstance instanceof Punctuation) { tokenType = SOURCE_ABAP; } else if (statementInstance instanceof Statements.Public || statementInstance instanceof Statements.Private || statementInstance instanceof Statements.Protected || statementInstance instanceof Statements.ClassDefinition || statementInstance instanceof Statements.ClassImplementation || statementInstance instanceof Statements.MethodImplementation || statementInstance instanceof Statements.EndMethod || statementInstance instanceof Statements.EndClass || statementInstance instanceof Statements.Interface || statementInstance instanceof Statements.EndInterface || statementInstance instanceof Statements.Form || statementInstance instanceof Statements.EndForm) { tokenType = BLOCK_ABAP; } else if (tokenInstance instanceof StringToken || tokenInstance instanceof StringTemplate || tokenInstance instanceof StringTemplateBegin || tokenInstance instanceof StringTemplateEnd || tokenInstance instanceof StringTemplateMiddle) { tokenType = LServer.SemanticTokenTypes.string; } else if (tokenInstance instanceof Comment) { tokenType = LServer.SemanticTokenTypes.comment; } else if (t instanceof TokenNodeRegex) { tokenType = SOURCE_ABAP; } const token = t.getFirstToken(); tokens.push({ line: token.getStart().getRow() - 1, startChar: token.getStart().getCol() - 1, length: token.getStr().length, tokenType: tokenType, tokenModifiers: [], }); } } return {data: this.encodeTokens(tokens)}; } private encodeTokens(tokens: Token[]): number[] { const ret: number[] = []; let prevLine: number | undefined = undefined; let prevChar: number | undefined = undefined; for (const t of tokens) { if (prevLine === undefined) { ret.push(t.line); } else { ret.push(t.line - prevLine); } if (prevLine === t.line && prevChar) { ret.push(t.startChar - prevChar); } else { ret.push(t.startChar); // todo, delta? } ret.push(t.length); ret.push(SemanticHighlighting.tokenTypeMap[t.tokenType]); ret.push(0); // no modifier logic implemented yet prevLine = t.line; prevChar = t.startChar; } return ret; } } |