lnx-search/lnx

How to use a customize tokenizer?

perkfly opened this issue · 1 comments

For east Asian languages such as Chinese, tokenizer can be very important. I notice that tantivy has some third-party tokenizers, how do I use it in lnx?

Currently, there's very minimal support for CJK language tokenizers within lnx, currently, we only really support Latin languages currently .(although future versions will have this)