Пример #1
0
 public void CustomClass()
 {
     MyTokenizer tokenizer = new MyTokenizer();
     Token[] tokens = tokenizer.Tokenize("A.Bērziņš ").ToArray();
     Assert.AreEqual("A", tokens[0].Text);
     Assert.AreEqual(".", tokens[1].Text);
     Assert.AreEqual("Bērziņš", tokens[2].Text);
     Assert.AreEqual(" ", tokens[3].Text);
 }
Пример #2
0
        public void CustomClass()
        {
            MyTokenizer tokenizer = new MyTokenizer();

            Token[] tokens = tokenizer.Tokenize("A.Bērziņš ").ToArray();
            Assert.AreEqual("A", tokens[0].Text);
            Assert.AreEqual(".", tokens[1].Text);
            Assert.AreEqual("Bērziņš", tokens[2].Text);
            Assert.AreEqual(" ", tokens[3].Text);
        }
Пример #3
0
        protected override TokenStreamComponents CreateComponents(string fieldName, TextReader reader)
        {
            const LuceneVersion version = LuceneVersion.LUCENE_48;

            Tokenizer           baseTokenizer       = new MyTokenizer(version, reader);
            StandardFilter      standardFilter      = new StandardFilter(version, baseTokenizer);
            WordDelimiterFilter wordDelimiterFilter = new WordDelimiterFilter(version, standardFilter,
                                                                              WordDelimiterFlags.CATENATE_WORDS | WordDelimiterFlags.GENERATE_WORD_PARTS |
                                                                              WordDelimiterFlags.PRESERVE_ORIGINAL | WordDelimiterFlags.SPLIT_ON_CASE_CHANGE, CharArraySet.EMPTY_SET);
            LowerCaseFilter lcFilter = new LowerCaseFilter(version, wordDelimiterFilter);

            return(new TokenStreamComponents(baseTokenizer, lcFilter));
        }