This issue tracker has been migrated to GitHub, and is currently read-only.
For more information, see the GitHub FAQs in the Python's Developer Guide.

Author ridgerat1611
Recipients eric.smith, ridgerat1611
Date 2021-03-17.17:14:44
SpamBayes Score -1.0
Marked as misclassified Yes
Message-id <1616001284.51.0.286703067814.issue43483@roundup.psfhosted.org>
In-reply-to
Content
Great minds think alike I guess... 

I was thinking of a much smaller carryover size... maybe 1K. With individual text blocks longer than that, the user will almost certainly be dealing with collecting and aggregating content text anyway, and in that case, the problem is solved before it happens. 

Here is a documentation change I was experimenting with...

-----------
ContentHandler.characters(content) -- The Parser will call this method to report chunks of character data.  In general, character data may be reported as a single chunk or as sequence of chunks; but character data sequences with fewer than  xml.sax.handler.ContiguousChunkLength characters, when uninterrupted any other xml.sax.handler.ContentHandler event, are guaranteed to be delivered as a single chunk...  
-----------

That puts users on notice, "...wait, are my chunks of text smaller than that?" and they are less likely to be caught unaware.  But of course, the implementation change would be helpful even without this extra warning.
History
Date User Action Args
2021-03-17 17:14:44ridgerat1611setrecipients: + ridgerat1611, eric.smith
2021-03-17 17:14:44ridgerat1611setmessageid: <1616001284.51.0.286703067814.issue43483@roundup.psfhosted.org>
2021-03-17 17:14:44ridgerat1611linkissue43483 messages
2021-03-17 17:14:44ridgerat1611create