[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread SilentGhost
SilentGhost added the comment: so using quoting=csv.QUOTE_NONE should solve the immediate problem of "losing" lines then, I'm not sure csv module ever supported dealing with corrupted files. -- ___ Python tracker

[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread Matthew Barnett
Matthew Barnett added the comment: I split the file into sections, each containing no more 1000 lines, and tried reading each section. Attached is a zip file of those that didn't return the expected number of rows. The problem appears to be due to unclosed quotes, which cause following lines

[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread Serhiy Storchaka
Serhiy Storchaka added the comment: What is the average number of columns on the file? -- ___ Python tracker ___ ___ Python-bugs-list

[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread Serhiy Storchaka
Serhiy Storchaka added the comment: > If I create a new file with all the skipped files, and I read it again in the > same way, around 30% of the rows are skipped. Could you please provide this smaller file? Or better make yet few iterations of keeping only skipped lines until the file will de

[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread SilentGhost
SilentGhost added the comment: Could you perhaps make the smaller file make available somewhere? -- nosy: +SilentGhost ___ Python tracker ___

[issue28642] csv reader loosing rows with big files and tab delimiter

2016-11-08 Thread Marc Garcia
New submission from Marc Garcia: I'm using the csv module from Python standard library, to read a 1.4Gb file with 11,157,064 of rows. The file is the Geonames dataset for all countries, which can be freely downloaded [1]. I'm using this code to read it: import csv with open('allCount