What makes XML streams different from character streams is that the
context of each token matters; you can't just pump
out a stream of random tags and data and expect an XML processor to
make sense of it. For example, a stream of ten start tags followed by
no end tags is not very useful, and definitely not well-formed XML.
Any data that isn't well-formed will be rejected.
After all, the whole purpose of XML is to package data in a way that
guarantees the integrity of a document's structure
and labeling, right?
These contextual rules are helpful to the parser as well as the
front-end processor. XML was designed to be very easy to parse,
unlike other markup languages that can require look-ahead or
look-behind. For example, SGML does not have a rule requiring
nonempty elements to have an end tag. To know when an element ends
requires sophisticated reasoning by the parser. This requirement
leads to code complexity, slower processing speed, and increased
memory usage.