Hi, I have a large number of files on the order of kilobytes on my local machine that I want to convert to a sequence file on HDFS. Whenever, I try to copy the local files to HDFS, hadoop complains about bad blocks, presumably because each block is 64mb and there are more files than blocks. In mahout 0.7, I would tell it that the input files are local, like:
mahout seqdirectory -i file://<input directory> -o <HDFS directory> But I can't use the same command on Mahout 0.9, where it expects the file system to be HDFS. Is there a workaround to generating the sequence file using Mahout 0.9? Thanks. Steven