예제 #1
0
        public void Spanish_BaseTest()
        {
            var stemmer = new SpanishStemmer();

            Assert.AreEqual("acerqu", stemmer.Stem("acerquen"));
            Assert.AreEqual("Vision", stemmer.Stem("Visionado"));
            Assert.AreEqual("algu", stemmer.Stem("algue"));
        }
예제 #2
0
        private TokenStream SpanishSteammer(TokenStream tokenStream)
        {
            //Obtener en una cadena cada token y aplicar el lematizador a cada término
            string      term    = string.Empty;
            IStemmer    stemmer = new SpanishStemmer();
            TokenStream tokenStreamtemp;
            var         termAttr = tokenStream.GetAttribute <ITermAttribute>();

            while (tokenStream.IncrementToken())
            {
                term = term + " " + stemmer.Stem(termAttr.Term);
            }
            tokenStreamtemp = new StandardTokenizer(Version.LUCENE_30, new System.IO.StringReader(term));
            return(tokenStreamtemp);
            //
        }
예제 #3
0
        public TokenStream SpanishSteammer(TokenStream tokenStream)
        {
            //Obtener en una cadena cada token y aplicar el lematizador a cada término
            string      term    = string.Empty;
            IStemmer    stemmer = new SpanishStemmer();
            TokenStream tokenStreamtemp;
            //var termAttr = tokenStream.GetAttribute<ITermAttribute>();
            string cadena = "";

            string[] token;
            Lucene.Net.Analysis.Token current;
            while ((current = tokenStream.Next()) != null)
            {
                cadena = current.ToString();
                token  = cadena.Split(',');
                cadena = cadena.Substring(1, token[0].Length - 1);
                term   = term + " " + stemmer.Stem(cadena);
            }
            tokenStreamtemp = new StandardTokenizer(new System.IO.StringReader(term));
            return(tokenStreamtemp);
            //
        }