MySQLdb

Daniel Bowett daniel at bowettsolutions.com
Tue Jan 25 15:55:26 EST 2005


Swaroop C H wrote:

>On Tue, 25 Jan 2005 20:43:54 +0000, Daniel Bowett
><daniel at bowettsolutions.com> wrote:
>  
>
>>I have just started playing around with MySQLdb for a project I am planning.
>>
>>As a test I have written a script that executes 3000 insert statements
>>on a table. The table contains 10 fields with a mix of text and numbers
>>- its a product table for a website eg UPC, ProductName, Price etc.
>>
>>The problem I have is that it takes just over two minuted to execute the
>>3000 insert statements which seems really slow! I am running it on a
>>machine with a 1.5 Ghz Pentium M Processor and Gig Of Ram. I dont think
>>the machine is to blame for the speed because during execution the
>>processor sits at about 10% and there is loads of free RAM.
>>    
>>
>
>I think a better option is to have the rows in a text file and then
>use `mysqlimport` or 'load data infile' SQL[1]
>
>[1]: http://dev.mysql.com/doc/mysql/en/load-data.html
>
>  
>
I remember using that a while ago when I dabbled with mysql so I think 
thats a potential solution. With my current solution I get full error 
logging with each row i try to insert. Will mysqlimport fail completeley 
if one row fails?



More information about the Python-list mailing list