Hi, 

I've got a fairly large (but not huge, 58mb) tab seperated text file, with
approximately 200 columns and 56k rows of numbers and strings. 

Here's a snippet of my code to create a numpy matrix from the data file... 

#### 

data = map(lambda x : x.strip().split('\t'), sys.stdin.readlines()) 
data = array(data) 

### 
data = array(data)
It causes the following error: 

> ValueError: setting an array element with a sequence 

If I take the 1st 40,000 lines of the file, it works fine. 
If I take the last 40,000 lines of the file, it also works fine, so it isn't
a problem with the file. 

I've found a few other posts complaining of the same problem, but none of
their fixes work. 

It seems like a memory problem to me. This was reinforced when I tried to
break the dataset into 3 chunks and stack the resulting arrays - I got an
error message saying "memory error". 
I don't really understand why reading in this 57mb txt file is taking up
~2gb's of RAM.

Any advice? Thanks in advance 

Dave
-- 
View this message in context: 
http://www.nabble.com/Numpy-2D-array-from-a-list-error-tp25531145p25531145.html
Sent from the Numpy-discussion mailing list archive at Nabble.com.

_______________________________________________
NumPy-Discussion mailing list
NumPy-Discussion@scipy.org
http://mail.scipy.org/mailman/listinfo/numpy-discussion

Reply via email to