In Spark you can use the normal globs supported by Hadoop's FileSystem, which are documented here: http://hadoop.apache.org/docs/r2.3.0/api/org/apache/hadoop/fs/FileSystem.html#globStatus(org.apache.hadoop.fs.Path)
On Wed, Jun 18, 2014 at 12:09 AM, MEETHU MATHEW <meethu2...@yahoo.co.in> wrote: > Hi Jianshi, > > I have used wild card characters (*) in my program and it worked.. > My code was like this > b = sc.textFile("hdfs:///path to file/data_file_2013SEP01*") > > Thanks & Regards, > Meethu M > > > On Wednesday, 18 June 2014 9:29 AM, Jianshi Huang < > jianshi.hu...@gmail.com> wrote: > > > It would be convenient if Spark's textFile, parquetFile, etc. can > support path with wildcard, such as: > > hdfs://domain/user/jianshuang/data/parquet/table/month=2014* > > Or is there already a way to do it now? > > Jianshi > > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ > > >