novel-segment
Version:
Chinese word segmentation 簡繁中文分词模块 以網路小說為樣本
17 lines (16 loc) • 590 B
TypeScript
import { SubSModuleTokenizer, ISubTokenizerCreate } from '../mod';
import { IWord, IDICT, IDICT2 } from '../Segment';
/**
* 注音
*/
export declare class ZhuyinTokenizer extends SubSModuleTokenizer {
name: string;
protected _TABLE: IDICT<IWord>;
protected _TABLE2: IDICT2<IWord>;
protected _cache(...argv: any[]): void;
split(words: IWord[]): IWord[];
splitZhuyin(text: string, cur?: number): IWord[];
}
export declare const init: ISubTokenizerCreate<ZhuyinTokenizer, SubSModuleTokenizer>;
export declare const type = "tokenizer";
export default ZhuyinTokenizer;