Example #1
0
        public LexerResult <T> Tokenize(string source)
        {
            List <Token <T> > tokens = new List <Token <T> >();

            var currentIndex = 0;
            //List<Token<T>> tokens = new List<Token<T>>();
            var       currentLine           = 1;
            var       currentColumn         = 0;
            var       currentLineStartIndex = 0;
            Token <T> previousToken         = null;

            while (currentIndex < source.Length)
            {
                currentColumn = currentIndex - currentLineStartIndex + 1;
                TokenDefinition <T> matchedDefinition = null;
                var matchLength = 0;

                foreach (var rule in tokenDefinitions)
                {
                    var match = rule.Regex.Match(source.Substring(currentIndex));

                    if (match.Success && match.Index == 0)
                    {
                        matchedDefinition = rule;
                        matchLength       = match.Length;
                        break;
                    }
                }

                if (matchedDefinition == null)
                {
                    return(new LexerResult <T>(new LexicalError(currentLine, currentColumn, source[currentIndex])));
                }

                var value = source.Substring(currentIndex, matchLength);

                if (matchedDefinition.IsEndOfLine)
                {
                    currentLineStartIndex = currentIndex + matchLength;
                    currentLine++;
                }

                if (!matchedDefinition.IsIgnored)
                {
                    previousToken = new Token <T>(matchedDefinition.TokenID, value,
                                                  new LexerPosition(currentIndex, currentLine, currentColumn));
                    tokens.Add(previousToken);
                }

                currentIndex += matchLength;
            }

            var eos = new Token <T>();

            if (previousToken != null)
            {
                eos.Position = new LexerPosition(previousToken.Position.Index + 1, previousToken.Position.Line,
                                                 previousToken.Position.Column + previousToken.Value.Length);
            }
            else
            {
                eos.Position = new LexerPosition(0, 0, 0);
            }


            tokens.Add(eos);
            return(new LexerResult <T>(tokens));
        }
Example #2
0
 public void AddDefinition(TokenDefinition <T> tokenDefinition)
 {
     tokenDefinitions.Add(tokenDefinition);
 }
Example #3
0
        public IEnumerable <Token <T> > Tokenize(string source)
        {
            int currentIndex = 0;
            //List<Token<T>> tokens = new List<Token<T>>();
            int       currentLine           = 1;
            int       currentColumn         = 0;
            int       currentLineStartIndex = 0;
            Token <T> previousToken         = null;

            TokenDefinition <T> defEol = tokenDefinitions.ToList <TokenDefinition <T> >().Find(t => t.IsEndOfLine);
            T eol = defEol.TokenID;

            while (currentIndex < source.Length)
            {
                currentColumn = currentIndex - currentLineStartIndex + 1;
                TokenDefinition <T> matchedDefinition = null;
                int matchLength = 0;

                foreach (var rule in tokenDefinitions)
                {
                    var match = rule.Regex.Match(source.Substring(currentIndex));

                    if (match.Success && match.Index == 0)
                    {
                        matchedDefinition = rule;
                        matchLength       = match.Length;
                        break;
                    }
                }

                if (matchedDefinition == null)
                {
                    throw new LexerException(new LexicalError(currentLine, currentColumn, source[currentIndex]));
                }
                else
                {
                    var value = source.Substring(currentIndex, matchLength);

                    if (matchedDefinition.IsEndOfLine)
                    {
                        currentLineStartIndex = currentIndex + matchLength;
                        currentLine++;
                    }
                    if (!matchedDefinition.IsIgnored)
                    {
                        previousToken = new Token <T>(matchedDefinition.TokenID, value, new TokenPosition(currentIndex, currentLine, currentColumn));
                        yield return(previousToken);
                    }
                    currentIndex += matchLength;


                    ;
                }
            }

            var eos = new Token <T>();

            eos.Position = new TokenPosition(previousToken.Position.Index + 1, previousToken.Position.Line, previousToken.Position.Column + previousToken.Value.Length);


            yield return(eos);
        }
Example #4
0
 public void AddDefinition(TokenDefinition <IN> tokenDefinition)
 {
 }
Example #5
0
 public void AddDefinition(TokenDefinition <IN> tokenDefinition)
 {
     throw new NotImplementedException();
 }