private static void Tokenize_IncompleteSentence_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings("This is an incomplete sentence ").ToArray(),
                             new[] { "This is an incomplete sentence " });
        }
예제 #2
0
        public void Tokenize_CompleteSentenceWithSpaceAtEnd_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("\"This is a complete sentence.\" \n"),
                        Is.EqualTo(new[] { "\"This is a complete sentence.\"" }));
        }
예제 #3
0
        public void Tokenize_Abbreviation_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer(new[] { "mr", "dr", "ms" });

            Assert.That(tokenizer.TokenizeToStrings("Mr. Smith went to Washington. This is the second sentence."),
                        Is.EqualTo(new[] { "Mr. Smith went to Washington.", "This is the second sentence." }));
        }
        private static void Tokenize_CompleteSentenceWithSpaceAtEnd_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings("\"This is a complete sentence.\" \n").ToArray(),
                             new[] { "\"This is a complete sentence.\"" });
        }
예제 #5
0
        public void Tokenize_IncompleteSentence_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("This is an incomplete sentence "),
                        Is.EqualTo(new[] { "This is an incomplete sentence " }));
        }
예제 #6
0
        public void Tokenize_QuotationInSentence_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("\"This is the first sentence!\" he said. This is the second sentence."),
                        Is.EqualTo(new[] { "\"This is the first sentence!\" he said.", "This is the second sentence." }));
        }
예제 #7
0
        public void Tokenize_Parens_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("This is the first sentence. (This is the second sentence.)"),
                        Is.EqualTo(new[] { "This is the first sentence.", "(This is the second sentence.)" }));
        }
        private static void Tokenize_Parens_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings("This is sentence one. (This is sentence two.)").ToArray(),
                             new[] { "This is sentence one.", "(This is sentence two.)" });
        }
예제 #9
0
        public void Tokenize_MultipleLines_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("This is the first sentence.\nThis is the second sentence."),
                        Is.EqualTo(new[] { "This is the first sentence.", "This is the second sentence." }));
        }
        private static void Tokenize_Abbreviation_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer(new[] { "mr", "dr", "ms" });

            assert.DeepEqual(tokenizer.TokenizeToStrings(
                                 "Mr. Smith went to Washington. This is sentence two.").ToArray(),
                             new[] { "Mr. Smith went to Washington.", "This is sentence two." });
        }
        private static void Tokenize_QuotationInSentence_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings(
                                 "\"This is sentence one!\" he said. This is sentence two.").ToArray(),
                             new[] { "\"This is sentence one!\" he said.", "This is sentence two." });
        }
        private static void Tokenize_Quotes_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings(
                                 "\"This is the first sentence.\" This is sentence two.").ToArray(),
                             new[] { "\"This is the first sentence.\"", "This is sentence two." });
        }
        private static void Tokenize_TwoSentences_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings(
                                 "This is the first sentence. This is the second sentence.").ToArray(),
                             new[] { "This is the first sentence.", "This is the second sentence." });
        }
예제 #14
0
        public void Tokenize_Empty_ReturnsEmpty()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings(""), Is.Empty);
        }
예제 #15
0
        public void Tokenize_SingleLine_ReturnsTokens()
        {
            var tokenizer = new LatinSentenceTokenizer();

            Assert.That(tokenizer.TokenizeToStrings("This is a test."), Is.EqualTo(new[] { "This is a test." }));
        }
        private static void Tokenize_Empty_ReturnsEmpty(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings("").ToArray(), new string[0]);
        }
        private static void Tokenize_SingleLine_ReturnsTokens(Assert assert)
        {
            var tokenizer = new LatinSentenceTokenizer();

            assert.DeepEqual(tokenizer.TokenizeToStrings("This is a test.").ToArray(), new[] { "This is a test." });
        }