tantivy-py icon indicating copy to clipboard operation
tantivy-py copied to clipboard

Python bindings for Tantivy

Results 19 tantivy-py issues
Sort by recently updated
recently updated
newest added

This change upgrades PyO3 to 0.21, which introduces a new API for python object references. The following references have more information: - Blog post by @davidhewitt : [Replacing PyO3's API...

Will there be support for other query method such as PhraseQuery, FuzzyTermQuery, RangeQuery?

enhancement
help wanted
feature-parity

Hi, Is there a way to score some documents higher than others? Something like `weight` in whoosh? For example: In an ecommerce site, we can have a `popularity` field for...

enhancement
help wanted
feature-parity

Incorporating the suggestion in #174. This is a little messy. Based on the [pdoc example for Mkdocs](https://github.com/mitmproxy/pdoc/tree/main/examples/mkdocs), it seems you have to run a script to execute pdoc before the...

documentation

I can see a contributor has already done work on exposing the field boost features. Is there a plan to add that to a release?

documentation
question

See #25 This is an early draft for discussion purposes. It's not ready to be used. The primary change is that we incorporate the Lindera tantivy tokenizer support under a...

enhancement
rust

Now we have `.pyi` file (#167), we can generate API docs by [pdoc](https://pdoc.dev/). For example, the following commands generate static HTML files in `apidoc` dir if you have `tantivy` module...

documentation

To reduce build effort and improve portability, PyO3 supports building in ["abi3" mode](https://pyo3.rs/v0.19.0/building_and_distribution#py_limited_apiabi3) where Python's stable ABI/API is used which makes the resulting binary wheels forward-compatible beginning at a specified...

maintainers

Hi there, I'm a user of [lancedb](https://lancedb.com/), which leverages `tantivy-py` for full text search indices (see https://lancedb.github.io/lancedb/fts). A current shortcoming of the lancedb FTS support is that it's only supported...

enhancement
help wanted

Currently, the tokenizer is hard-coded to default, it would be better to include some configurable tokenizer for Chinese (tantivy-jieba and cang-jie), Japanese (lindera and tantivy-tokenizer-tiny-segmente) and Korean (lindera + lindera-ko-dic-builder)...

enhancement