UNPKG

novel-segment

Version:

Chinese word segmentation 簡繁中文分词模块 以網路小說為樣本

19 lines (18 loc) 654 B
import { SubSModuleTokenizer, ISubTokenizerCreate } from '../mod'; import { IWord, IDICT, IDICT2 } from '../Segment'; /** * 此模組目前無任何用處與效果 * * @todo 部首 */ export declare class ZhRadicalTokenizer extends SubSModuleTokenizer { name: string; protected _TABLE: IDICT<IWord>; protected _TABLE2: IDICT2<IWord>; protected _cache(...argv: any[]): void; split(words: IWord[]): IWord[]; splitZhRadical(text: string, cur?: number): IWord[]; } export declare const init: ISubTokenizerCreate<ZhRadicalTokenizer, SubSModuleTokenizer>; export declare const type = "tokenizer"; export default ZhRadicalTokenizer;