UNPKG

langium

Version:

A language engineering tool for the Language Server Protocol

460 lines 20.5 kB
/****************************************************************************** * Copyright 2021 TypeFox GmbH * This program and the accompanying materials are made available under the * terms of the MIT License, which is available in the project root. ******************************************************************************/ import { CompletionItemKind, CompletionList, Position } from 'vscode-languageserver'; import * as ast from '../../languages/generated/ast.js'; import { assignMandatoryProperties, getContainerOfType } from '../../utils/ast-utils.js'; import { findDeclarationNodeAtOffset, findLeafNodeBeforeOffset } from '../../utils/cst-utils.js'; import { getEntryRule, getExplicitRuleType } from '../../utils/grammar-utils.js'; import { stream } from '../../utils/stream.js'; import { findFirstFeatures, findNextFeatures } from './follow-element-computation.js'; export function mergeCompletionProviderOptions(options) { const triggerCharacters = Array.from(new Set(options.flatMap(option => option?.triggerCharacters ?? []))); const allCommitCharacters = Array.from(new Set(options.flatMap(option => option?.allCommitCharacters ?? []))); return { triggerCharacters: triggerCharacters.length > 0 ? triggerCharacters : undefined, allCommitCharacters: allCommitCharacters.length > 0 ? allCommitCharacters : undefined }; } export class DefaultCompletionProvider { constructor(services) { this.scopeProvider = services.references.ScopeProvider; this.grammar = services.Grammar; this.completionParser = services.parser.CompletionParser; this.nameProvider = services.references.NameProvider; this.lexer = services.parser.Lexer; this.nodeKindProvider = services.shared.lsp.NodeKindProvider; this.fuzzyMatcher = services.shared.lsp.FuzzyMatcher; this.grammarConfig = services.parser.GrammarConfig; this.astReflection = services.shared.AstReflection; this.documentationProvider = services.documentation.DocumentationProvider; } async getCompletion(document, params, _cancelToken) { const items = []; const contexts = this.buildContexts(document, params.position); const acceptor = (context, value) => { const completionItem = this.fillCompletionItem(context, value); if (completionItem) { items.push(completionItem); } }; const distinctionFunction = (element) => { if (ast.isKeyword(element.feature)) { return element.feature.value; } else { return element.feature; } }; const completedFeatures = []; for (const context of contexts) { await Promise.all(stream(context.features) .distinct(distinctionFunction) .exclude(completedFeatures) .map(e => this.completionFor(context, e, acceptor))); // Do not try to complete the same feature multiple times completedFeatures.push(...context.features); // We might want to stop computing completion results if (!this.continueCompletion(items)) { break; } } return CompletionList.create(this.deduplicateItems(items), true); } /** * The completion algorithm could yield the same reference/keyword multiple times. * * This methods deduplicates these items afterwards before returning to the client. * Unique items are identified as a combination of `kind`, `label` and `detail`. */ deduplicateItems(items) { return stream(items).distinct(item => `${item.kind}_${item.label}_${item.detail}`).toArray(); } findFeaturesAt(document, offset) { const text = document.getText({ start: Position.create(0, 0), end: document.positionAt(offset) }); const parserResult = this.completionParser.parse(text); const tokens = parserResult.tokens; // If the parser didn't parse any tokens, return the next features of the entry rule if (parserResult.tokenIndex === 0) { const parserRule = getEntryRule(this.grammar); const firstFeatures = findFirstFeatures({ feature: parserRule.definition, type: getExplicitRuleType(parserRule) }); if (tokens.length > 0) { // We have to skip the first token // The interpreter will only look at the next features, which requires every token after the first tokens.shift(); return findNextFeatures(firstFeatures.map(e => [e]), tokens); } else { return firstFeatures; } } const leftoverTokens = [...tokens].splice(parserResult.tokenIndex); const features = findNextFeatures([parserResult.elementStack.map(feature => ({ feature }))], leftoverTokens); return features; } *buildContexts(document, position) { const cst = document.parseResult.value.$cstNode; if (!cst) { return; } const textDocument = document.textDocument; const text = textDocument.getText(); const offset = textDocument.offsetAt(position); const partialContext = { document, textDocument, offset, position }; // Data type rules need special handling, as their tokens are irrelevant for completion purposes. // If we encounter a data type rule at the current offset, we jump to the start of the data type rule. const dataTypeRuleOffsets = this.findDataTypeRuleStart(cst, offset); if (dataTypeRuleOffsets) { const [ruleStart, ruleEnd] = dataTypeRuleOffsets; const parentNode = findLeafNodeBeforeOffset(cst, ruleStart)?.astNode; yield { ...partialContext, node: parentNode, tokenOffset: ruleStart, tokenEndOffset: ruleEnd, features: this.findFeaturesAt(textDocument, ruleStart), }; } // For all other purposes, it's enough to jump to the start of the current/previous token const { nextTokenStart, nextTokenEnd, previousTokenStart, previousTokenEnd } = this.backtrackToAnyToken(text, offset); let astNodeOffset = nextTokenStart; if (offset <= nextTokenStart && previousTokenStart !== undefined) { // This check indicates that the cursor is still before the next token, so we should use the previous AST node (if it exists) astNodeOffset = previousTokenStart; } const astNode = findLeafNodeBeforeOffset(cst, astNodeOffset)?.astNode; let performNextCompletion = true; if (previousTokenStart !== undefined && previousTokenEnd !== undefined && previousTokenEnd === offset) { // This context aims to complete the current feature yield { ...partialContext, node: astNode, tokenOffset: previousTokenStart, tokenEndOffset: previousTokenEnd, features: this.findFeaturesAt(textDocument, previousTokenStart), }; // The completion after the current token should be prevented in case we find out that the current token definitely isn't completed yet // This is usually the case when the current token ends on a letter. performNextCompletion = this.performNextTokenCompletion(document, text.substring(previousTokenStart, previousTokenEnd), previousTokenStart, previousTokenEnd); if (performNextCompletion) { // This context aims to complete the immediate next feature (if one exists at the current cursor position) // It uses the previous cst start/offset for that. yield { ...partialContext, node: astNode, tokenOffset: previousTokenEnd, tokenEndOffset: previousTokenEnd, features: this.findFeaturesAt(textDocument, previousTokenEnd), }; } } if (!astNode) { const parserRule = getEntryRule(this.grammar); if (!parserRule) { throw new Error('Missing entry parser rule'); } // This context aims to perform completion for the grammar start (usually when the document is empty) yield { ...partialContext, tokenOffset: nextTokenStart, tokenEndOffset: nextTokenEnd, features: findFirstFeatures(parserRule.definition) }; } else if (performNextCompletion) { // This context aims to complete the next feature, using the next cst start/end yield { ...partialContext, node: astNode, tokenOffset: nextTokenStart, tokenEndOffset: nextTokenEnd, features: this.findFeaturesAt(textDocument, nextTokenStart), }; } } performNextTokenCompletion(document, text, _offset, _end) { // This regex returns false if the text ends with a letter. // We don't want to complete new text immediately after a keyword, ID etc. // We only care about the last character in the text, so we use $ here. // The \P{L} used here is a Unicode category that matches any character that is not a letter return /\P{L}$/u.test(text); } findDataTypeRuleStart(cst, offset) { let containerNode = findDeclarationNodeAtOffset(cst, offset, this.grammarConfig.nameRegexp); // Identify whether the element was parsed as part of a data type rule let isDataTypeNode = Boolean(getContainerOfType(containerNode?.grammarSource, ast.isParserRule)?.dataType); if (isDataTypeNode) { while (isDataTypeNode) { // Use the container to find the correct parent element containerNode = containerNode?.container; isDataTypeNode = Boolean(getContainerOfType(containerNode?.grammarSource, ast.isParserRule)?.dataType); } if (containerNode) { return [containerNode.offset, containerNode.end]; } } return undefined; } /** * Indicates whether the completion should continue to process the next completion context. * * The default implementation continues the completion only if there are currently no proposed completion items. */ continueCompletion(items) { return items.length === 0; } /** * This method returns two sets of token offset information. * * The `nextToken*` offsets are related to the token at the cursor position. * If there is none, both offsets are simply set to `offset`. * * The `previousToken*` offsets are related to the last token before the current token at the cursor position. * They are `undefined`, if there is no token before the cursor position. */ backtrackToAnyToken(text, offset) { const tokens = this.lexer.tokenize(text).tokens; if (tokens.length === 0) { // If we don't have any tokens in our document, just return the offset position return { nextTokenStart: offset, nextTokenEnd: offset }; } let previousToken; for (const token of tokens) { if (token.startOffset >= offset) { // We are between two tokens // Return the current offset as the next token index return { nextTokenStart: offset, nextTokenEnd: offset, previousTokenStart: previousToken ? previousToken.startOffset : undefined, previousTokenEnd: previousToken ? previousToken.endOffset + 1 : undefined }; } if (token.endOffset >= offset) { // We are within a token // Return the current and previous token offsets as normal return { nextTokenStart: token.startOffset, nextTokenEnd: token.endOffset + 1, previousTokenStart: previousToken ? previousToken.startOffset : undefined, previousTokenEnd: previousToken ? previousToken.endOffset + 1 : undefined }; } previousToken = token; } // We have run into the end of the file // Return the current offset as the next token index return { nextTokenStart: offset, nextTokenEnd: offset, previousTokenStart: previousToken ? previousToken.startOffset : undefined, previousTokenEnd: previousToken ? previousToken.endOffset + 1 : undefined }; } completionFor(context, next, acceptor) { if (ast.isKeyword(next.feature)) { return this.completionForKeyword(context, next.feature, acceptor); } else if (ast.isCrossReference(next.feature) && context.node) { return this.completionForCrossReference(context, next, acceptor); } // Don't offer any completion for other elements (i.e. terminals, datatype rules) // We - from a framework level - cannot reasonably assume their contents. // Adopters can just override `completionFor` if they want to do that anyway. } completionForCrossReference(context, next, acceptor) { const assignment = getContainerOfType(next.feature, ast.isAssignment); let node = context.node; if (assignment && node) { if (next.type) { // When `type` is set, it indicates that we have just entered a new parser rule. // The cross reference that we're trying to complete is on a new element that doesn't exist yet. // So we create a new synthetic element with the correct type information. node = { $type: next.type, $container: node, $containerProperty: next.property }; assignMandatoryProperties(this.astReflection, node); } let reference; if (next.feature.isMulti) { reference = { $refText: '', items: [] }; } else { reference = { $refText: '', ref: undefined }; } const refInfo = { reference, container: node, property: assignment.feature }; try { for (const candidate of this.getReferenceCandidates(refInfo, context)) { acceptor(context, this.createReferenceCompletionItem(candidate, refInfo, context)); } } catch (err) { console.error(err); } } } /** * Override this method to change how the stream of candidates is determined for a reference. * This way completion-specific modifications and refinements can be added to the proposals computation * beyond the rules being implemented in the scope provider, e.g. filtering. * * @param refInfo Information about the reference for which the candidates are requested. * @param _context Information about the completion request including document, cursor position, token under cursor, etc. * @returns A stream of all elements being valid for the given reference. */ getReferenceCandidates(refInfo, _context) { return this.scopeProvider.getScope(refInfo).getAllElements(); } /** * Override this method to change how reference completion items are created. * * To change the `kind` of a completion item, override the `NodeKindProvider` service instead. * To change the `documentation`, override the `DocumentationProvider` service instead. * * @param nodeDescription The description of a reference candidate * @param _refInfo Information about the reference for which the candidate is proposed * @param _context The completion context * @returns A partial completion item */ createReferenceCompletionItem(nodeDescription, _refInfo, _context) { const kind = this.nodeKindProvider.getCompletionItemKind(nodeDescription); const documentation = this.getReferenceDocumentation(nodeDescription); return { nodeDescription, kind, documentation, detail: nodeDescription.type, sortText: '0' }; } getReferenceDocumentation(nodeDescription) { if (!nodeDescription.node) { return undefined; } const documentationText = this.documentationProvider.getDocumentation(nodeDescription.node); if (!documentationText) { return undefined; } return { kind: 'markdown', value: documentationText }; } completionForKeyword(context, keyword, acceptor) { if (!this.filterKeyword(context, keyword)) { return; } acceptor(context, { label: keyword.value, kind: this.getKeywordCompletionItemKind(keyword), detail: 'Keyword', sortText: '1' }); } getKeywordCompletionItemKind(_keyword) { return CompletionItemKind.Keyword; } filterKeyword(context, keyword) { // Filter out keywords that do not contain any word character return /\p{L}/u.test(keyword.value); } fillCompletionItem(context, item) { let label; if (typeof item.label === 'string') { label = item.label; } else if ('node' in item) { const name = this.nameProvider.getName(item.node); if (!name) { return undefined; } label = name; } else if ('nodeDescription' in item) { label = item.nodeDescription.name; } else { return undefined; } let insertText; if (typeof item.textEdit?.newText === 'string') { insertText = item.textEdit.newText; } else if (typeof item.insertText === 'string') { insertText = item.insertText; } else { insertText = label; } const textEdit = item.textEdit ?? this.buildCompletionTextEdit(context, label, insertText); if (!textEdit) { return undefined; } // Copy all valid properties of `CompletionItem` const completionItem = { additionalTextEdits: item.additionalTextEdits, command: item.command, commitCharacters: item.commitCharacters, data: item.data, detail: item.detail, documentation: item.documentation, filterText: item.filterText, insertText: item.insertText, insertTextFormat: item.insertTextFormat, insertTextMode: item.insertTextMode, kind: item.kind, labelDetails: item.labelDetails, preselect: item.preselect, sortText: item.sortText, tags: item.tags, textEditText: item.textEditText, textEdit, label }; return completionItem; } buildCompletionTextEdit(context, label, newText) { const content = context.textDocument.getText(); const identifier = content.substring(context.tokenOffset, context.offset); if (this.fuzzyMatcher.match(identifier, label)) { const start = context.textDocument.positionAt(context.tokenOffset); const end = context.position; return { newText, range: { start, end } }; } else { return undefined; } } } //# sourceMappingURL=completion-provider.js.map