diff --git a/src/collector/filter_collector_wrapper.rs b/src/collector/filter_collector_wrapper.rs index c19253cbc2..a961c4edc1 100644 --- a/src/collector/filter_collector_wrapper.rs +++ b/src/collector/filter_collector_wrapper.rs @@ -64,9 +64,9 @@ use crate::{DocId, Score, SegmentReader, TantivyError}; /// # } /// ``` /// -/// Note that this is limited to fast fields which implement the [`FastValue`] trait, -/// e.g. `u64` but not `&[u8]`. To filter based on a bytes fast field, -/// use a [`BytesFilterCollector`] instead. +/// Note that this is limited to fast fields which implement the +/// [`FastValue`][crate::fastfield::FastValue] trait, e.g. `u64` but not `&[u8]`. +/// To filter based on a bytes fast field, use a [`BytesFilterCollector`] instead. pub struct FilterCollector where TPredicate: 'static + Clone { diff --git a/src/tokenizer/mod.rs b/src/tokenizer/mod.rs index dbc3dd867d..23f7893d29 100644 --- a/src/tokenizer/mod.rs +++ b/src/tokenizer/mod.rs @@ -154,7 +154,7 @@ pub use self::split_compound_words::SplitCompoundWords; pub use self::stemmer::{Language, Stemmer}; pub use self::stop_word_filter::StopWordFilter; pub use self::tokenized_string::{PreTokenizedStream, PreTokenizedString}; -pub use self::tokenizer::TextAnalyzer; +pub use self::tokenizer::{TextAnalyzer, TextAnalyzerBuilder}; pub use self::tokenizer_manager::TokenizerManager; pub use self::whitespace_tokenizer::WhitespaceTokenizer;