The workaround is to add -xm sequential. A MR version of seqdirectory was 
introduced in 0.8 and hence the default execution mode is MR if none is 
specified.






On Thursday, March 13, 2014 4:12 PM, Steven Cullens <srcull...@gmail.com> wrote:
 
Hi,

I have a large number of files on the order of kilobytes on my local
machine that I want to convert to a sequence file on HDFS.  Whenever, I try
to copy the local files to HDFS, hadoop complains about bad blocks,
presumably because each block is 64mb and there are more files than blocks.
In mahout 0.7, I would tell it that the input files are local, like:

mahout seqdirectory -i file://<input directory> -o <HDFS directory>

But I can't use the same command on Mahout 0.9, where it expects the file
system to be HDFS.  Is there a workaround to generating the sequence file
using Mahout 0.9?  Thanks.

Steven

Reply via email to