UNPKG

type2docfx

Version:

A tool to convert json format output from TypeDoc to universal reference model for DocFx to consume.

60 lines (59 loc) 2.13 kB
/** * @module botbuilder-choices */ /** * Copyright (c) Microsoft Corporation. All rights reserved. * Licensed under the MIT License. */ /** * :package: **botbuilder-choices** * * Individual token returned by a `TokenizerFunction`. */ export interface Token { /** Start character position of the token within the outer string. */ start: number; /** End character position of the token within the outer string. */ end: number; /** Original text of the token. */ text: string; /** Normalized form of the token. This can include things like lower casing or stemming. */ normalized: string; } /** * :package: **botbuilder-choices** * * Signature for an alternate word breaker that can be passed to `recognizeChoices()`, * `findChoices()`, or `findValues()`. The `defaultTokenizer()` is fairly simple and only breaks * on spaces and punctuation. * @param TokenizerFunction.text The text to be tokenized. * @param TokenizerFunction.locale (Optional) locale of the text if known. */ export declare type TokenizerFunction = (text: string, locale?: string) => Token[]; /** * :package: **botbuilder-choices** * * Simple tokenizer that breaks on spaces and punctuation. The only normalization done is to * lowercase the tokens. Developers can wrap this tokenizer with their own function to perform * additional normalization like [stemming](https://github.com/words/stemmer). * * **Usage Example** * * ```JavaScript * const { recognizeChoices, defaultTokenizer } = require('botbuilder-choices'); * const stemmer = require('stemmer'); * * function customTokenizer(text, locale) { * const tokens = defaultTokenizer(text, locale); * tokens.forEach((t) => { * t.normalized = stemmer(t.normalized); * }); * return tokens; * } * * const choices = ['red', 'green', 'blue']; * const utterance = context.activity.text; * const results = recognizeChoices(utterance, choices, { tokenizer: customTokenizer }); * ``` */ export declare function defaultTokenizer(text: string, locale?: string): Token[];