In Spark you can use the normal globs supported by Hadoop's FileSystem,
which are documented here:
http://hadoop.apache.org/docs/r2.3.0/api/org/apache/hadoop/fs/FileSystem.html#globStatus(org.apache.hadoop.fs.Path)


On Wed, Jun 18, 2014 at 12:09 AM, MEETHU MATHEW <meethu2...@yahoo.co.in>
wrote:

> Hi Jianshi,
>
> I have used wild card characters (*) in my program and it worked..
> My code was like this
> b = sc.textFile("hdfs:///path to file/data_file_2013SEP01*")
>
> Thanks & Regards,
> Meethu M
>
>
>   On Wednesday, 18 June 2014 9:29 AM, Jianshi Huang <
> jianshi.hu...@gmail.com> wrote:
>
>
>  It would be convenient if Spark's textFile, parquetFile, etc. can
> support path with wildcard, such as:
>
>   hdfs://domain/user/jianshuang/data/parquet/table/month=2014*
>
>  Or is there already a way to do it now?
>
> Jianshi
>
> --
> Jianshi Huang
>
> LinkedIn: jianshi
> Twitter: @jshuang
> Github & Blog: http://huangjs.github.com/
>
>
>

Reply via email to