unicode encoding usablilty problem
aurora
aurora00 at gmail.com
Fri Feb 18 13:53:36 EST 2005
More information about the Python-list mailing list
Fri Feb 18 13:53:36 EST 2005
- Previous message (by thread): unicode encoding usablilty problem
- Next message (by thread): unicode encoding usablilty problem
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
On Fri, 18 Feb 2005 19:24:10 +0100, Fredrik Lundh <fredrik at pythonware.com> wrote: > that's how you should do things in Python too, of course. a unicode > string > uses unicode internally. decode on the way in, encode on the way out, and > things just work. > > the fact that you can mess things up by mixing unicode strings with > binary > strings doesn't mean that you have to mix unicode strings with binary > strings > in your program. I don't want to mix them. But how could I find them? How do I know this statement can be potential problem if a==b: where a and b can be instantiated individually far away from this line of code that put them together? In Java they are distinct data type and the compiler would catch all incorrect usage. In Python, the interpreter seems to 'help' us to promote binary string to unicode. Things works fine, unit tests pass, all until the first non-ASCII characters come in and then the program breaks. Is there a scheme for Python developer to use so that they are safe from incorrect mixing?
- Previous message (by thread): unicode encoding usablilty problem
- Next message (by thread): unicode encoding usablilty problem
- Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
More information about the Python-list mailing list