Thanks to a lot of help, I've got the outer framework for my tokenizer
down to this:

    for line_number, line in enumerate(text):
        output = ''

        for char_number, char in enumerate(line):
            output += char

        print 'At ' + str(line_number) + ', '+ str(char_number) + ': '
+ output,

Or do I?

My other tokenizers all worked with array indices. When I came to,
say, an open quote character, I passed control to the get_a_string()
method. That would, locate the end of the string, push a string token
onto the stack and then forward the char array index to point to the
next character in the line.

Is there a pythonic design I'm overlooking?
-- 
http://mail.python.org/mailman/listinfo/python-list

Reply via email to