This issue tracker has been migrated to GitHub, and is currently read-only.
For more information, see the GitHub FAQs in the Python's Developer Guide.

Author mark.dickinson
Recipients barry, loewis, mark.dickinson, michael.foord, trent
Date 2008-03-18.18:01:21
SpamBayes Score 0.18736567
Marked as misclassified No
Message-id <1205863283.23.0.195193084381.issue719888@psf.upfronthosting.co.za>
In-reply-to
Content
Is it worth keeping generate_tokens as an alias for tokenize, just
to avoid gratuitous 2-to-3 breakage?  Maybe not---I guess they're
different beasts, in that one wants a string-valued iterator and the 
other wants a bytes-valued iterator.

So if I understand correctly, the readline argument to tokenize
would have to return bytes instances.  Would it be worth adding a check
for this, to catch possible misuse?  You could put the check in 
detect_encoding, so that just checks that the first one or two yields
from readline have the correct type, and assumes that the rest is okay.
History
Date User Action Args
2008-03-18 18:01:23mark.dickinsonsetspambayes_score: 0.187366 -> 0.18736567
recipients: + mark.dickinson, loewis, barry, trent, fuzzyman
2008-03-18 18:01:23mark.dickinsonsetspambayes_score: 0.187366 -> 0.187366
messageid: <1205863283.23.0.195193084381.issue719888@psf.upfronthosting.co.za>
2008-03-18 18:01:22mark.dickinsonlinkissue719888 messages
2008-03-18 18:01:21mark.dickinsoncreate