No, no, that's wrong. MySQL and the Python interface to it understand
Unicode. You don't want to convert data to UTF-8 before putting it in a
database; the database indexing won't work.
I doubt that indexing has anything to do with it whatsoever.
Here's how to do it right.
First, tell MySQL, before you create your MySQL tables, that the
tables are
to be stored in Unicode:
ALTER database yourdatabasename DEFAULT CHARACTER SET utf8;
You can also do this on a table by table basis, or even for single fields,
but you'll probably get confused if you do.
Then, when you connect to the database in Python, use something like
this:
db = MySQLdb.connect(host="localhost",
use_unicode = True, charset = "utf8",
user=username, passwd=password, db=database)
That tells MySQLdb to talk to the database in Unicode, and it tells the
database
(via "charset") that you're talking Unicode.
You confuse unicode with utf-8 here. And while this appears to be
nitpicking, it is important to write this small program and meditate the
better part of an hour in front of it running:
while True:
print "utf-8 is not unicode"
You continue to make that error below, so I snip that.
The important part is this: unicode is a standard that aims to provide a
codepoint for each and every character that humankind has invented. And
python unicode objects can also represent all characters one can imagine.
However, unicode as such is an abstraction. Harddisks, network sockets,
databases and the like don't deal with abstractions though - the eat
bytes. Which makes it necessary to encode unicode objects to
byte-strings when serializing them. Thus there are the thingies called
encodings: latin1 for most characters used in westen europe for example.
But it is limited to 256 characters (actually, even less), chinese or
russian customers won't get too happy with them.
So some encodings are defined that are capable of encoding _ALL_ unicode
codepoints. Either by being larger than one byte for each character. Or
by providing escape-mechanisms. The former are e.g. UCS4 (4 bytes per
character), the most important member of the latter is utf-8. Which uses
ascii + escapes to encode all codepoints.
Now what does that mean in python?
First of all, the coding:-declaration: it tells python which encoding to
use when dealing with unicode-literals, which are the
u"something"
thingies. If you use a coding of latin1, that means that the text
u"ö"
is expected to be one byte long, with the proper value that depicts the
german umlaut o in latin1. Which is 0xf6.
If coding: is set to utf-8, the same string has to consist not of one,
but of two bytes: 0xc3 0xb6.
So, when editing files that are supposed to contain "funny" characters,
you have to
- set your editor to save the file in an appropriate encoding
- specify the same encoding in the coding:-declaration
Regarding databases: they store bytes. Mostly. Some allow to store
unicode by means of one of the fixed-size-encodings, but you pay a
storage-size penalty for that.
So - you we're right when you said that one can change the encoding a db
uses, on several levels even.
But that's not all that is to it. Another thing is the encoding the
CONNECTION expects byte-strings to be passed, and will use to render
returned strings in. The conversion from and to the used storage
encoding is done automagically.
It is for example perfectly legal (and unfortunately happens
involuntarily) to have a database that internally uses utf-8 as storage,
potentially being able to store all possible codepoints.
But due to e.g. environmental settings, opened connections will deliver
the contents in e.g. latin1. Which of course will lead to problems if
you try to return data from the table with the topmost chines first names.
So you can alter the encoding the connection delivers and expects
byte-strings in. In mysql, this can be done explcit using
cursor.execute("set names <encoding>")
Or - as you said - as part of a connection-string.
db = MySQLdb.connect(host="localhost",
use_unicode = True, charset = "utf8",
user=username, passwd=password, db=database)
But there is more to it. If the DB-API supports it, then the API itself
will decode the returned strings, using the specified encoding, so that
the user will only deal with "real" unicode-objects, greatly reducing
the risk of mixing byte-strings with unicode-objects. That's what the
use_unicod-parameter is for: it makes the API accept and deliver
unicod-objects. But it would do so even if the charset-parameter was
"latin1". Which makes me repeat the lesson from the beginning:
while True:
print "utf-8 is not unicode"
Diez