@orama/tokenizers
Version:
Additional tokenizers for Orama
28 lines (22 loc) • 533 B
Markdown
# Orama Tokenizers
This package provides support for additional tokenizers for the Orama Search Engine.
Available tokenizers:
- Chinese (Mandarin, experimental)
- Japanese (experimental)
- Korean (work in progress)
Usage:
```js
import { create } from '@orama/orama'
import { createTokenizer } from '@orama/tokenizers/mandarin'
const db = await create({
schema: {
myProperty: 'string',
anotherProperty: 'number'
},
components: {
tokenizer: await createTokenizer()
}
})
```
# License
[Apache 2.0](/LICENSE.md)