Hey,

thank you again for your explanation.

-snip-
> You could let the lexer simply create single tokens and create parser rules
> that match a certain range of tokens (like the `ab` rule below):
-snip-

Creating a different token for each character is not possible since I'm dealing 
with unicode.

Only separating out characters I care about syntactically might be an option.
But won't  this become awkward when I want to catch longer tokens like
'<ref name="'
then I would have to create lexer rules for '<', 'r' 'e' 'f' ... and use those 
to express this string as
LThan R E F Space N A M E Equals Quote

This doesn't seem like a good way to do this.

I still think that a scannerless parser might be a better alternative. Are 
there any good reasons against switching (apart from ANTLR being a great tool 
in general)?

Regards,

Patrick

List: http://www.antlr.org/mailman/listinfo/antlr-interest
Unsubscribe: 
http://www.antlr.org/mailman/options/antlr-interest/your-email-address

-- 
You received this message because you are subscribed to the Google Groups 
"il-antlr-interest" group.
To post to this group, send email to il-antlr-inter...@googlegroups.com.
To unsubscribe from this group, send email to 
il-antlr-interest+unsubscr...@googlegroups.com.
For more options, visit this group at 
http://groups.google.com/group/il-antlr-interest?hl=en.

Reply via email to