UNPKG

novel-segment

Version:

Chinese word segmentation 簡繁中文分词模块 以網路小說為樣本

17 lines (16 loc) 590 B
import { SubSModuleTokenizer, ISubTokenizerCreate } from '../mod'; import { IWord, IDICT, IDICT2 } from '../Segment'; /** * 注音 */ export declare class ZhuyinTokenizer extends SubSModuleTokenizer { name: string; protected _TABLE: IDICT<IWord>; protected _TABLE2: IDICT2<IWord>; protected _cache(...argv: any[]): void; split(words: IWord[]): IWord[]; splitZhuyin(text: string, cur?: number): IWord[]; } export declare const init: ISubTokenizerCreate<ZhuyinTokenizer, SubSModuleTokenizer>; export declare const type = "tokenizer"; export default ZhuyinTokenizer;