コード例 #1
0
        public string ShallowParse(string input)
        {
            var output = new StringBuilder();

            string[] sentences = _sentenceDetector.SentenceDetect(input);
            foreach (string sentence in sentences)
            {
                string[] tokens = _tokenizer.Tokenize(sentence);
                string[] tags   = _posTagger.Tag(tokens);
                output.Append(string.Join(" ", _chunker.GetChunks(tokens, tags)));
                output.Append("\r\n\r\n");
            }
            return(output.ToString());
        }
コード例 #2
0
        private string ChunkSentence(string[] tokens, string[] tags)
        {
            if (_chunker == null)
            {
                _chunker = new EnglishTreebankChunker(_modelPath + "EnglishChunk.nbin");
            }

            return(string.Join(" ", _chunker.GetChunks(tokens, tags)));
        }
コード例 #3
0
ファイル: NlpUtilities.cs プロジェクト: ghdy/GHDY.All
        public static string ChunkSentence(string[] tokens, string[] tags)
        {
            if (mChunker == null)
            {
                mChunker = new OpenNLP.Tools.Chunker.EnglishTreebankChunker(mModelPath + "EnglishChunk.nbin");
            }

            return(mChunker.GetChunks(tokens, tags));
        }
コード例 #4
0
        internal string ChunkSentence(string[] tokens, string[] tags)
        {
            if (_mChunker == null)
            {
                _mChunker = new EnglishTreebankChunker(_modelPath + "EnglishChunk.nbin");
            }

            return(_mChunker.GetChunks(tokens, tags));
        }
コード例 #5
0
ファイル: OpenNLPForm.cs プロジェクト: MoawazAyub/CombinedNLP
        private List <SentenceChunk> MyChunker(string str)
        {
            var modelPath = Path.GetDirectoryName(Process.GetCurrentProcess().MainModule.FileName) + @"\Models\EnglishChunk.nbin";
            var chunker   = new EnglishTreebankChunker(modelPath);
            var tokens    = MyTokenizer(str);
            var pos       = MyPosTagger(str);

            var chunks = chunker.GetChunks(tokens, pos);

            return(chunks);
        }
コード例 #6
0
        public void Test()
        {
            var txt       = "1980 was certainly a year for bad backwoods slasher movies. \"Friday The 13th\" and \"The Burning\" may have been the best ones but there were like always a couple of stinkers not far behind like \"Don't Go Into The Woods Alone\" and this one. But in all fairness \"The Prey\" is nowhere near as bad as \"Don't Go Into The Woods\" but it's still not great either.";
            var sentences = sentenceDetector.SentenceDetect(txt);

            Assert.AreEqual(3, sentences.Length);
            var tokens = tokenizer.Tokenize(sentences[0]);

            Assert.AreEqual(11, tokens.Length);
            var tags = postTagger.Tag(tokens);

            Assert.AreEqual(11, tags.Length);
            var chunks = chunker.GetChunks(tokens, tags);

            Assert.AreEqual(7, chunks.Length);
        }
コード例 #7
0
 public string Chunk(IEnumerable <string> tokens, IEnumerable <string> tags)
 {
     return(chunker.GetChunks(tokens.ToArray(), tags.ToArray()));
 }