Thanks Stefan, my colleague Shashank has filed a bug for the same in jira

https://issues.apache.org/jira/browse/FLINK-6993

Regards,
Adarsh

On Fri, Jun 23, 2017 at 8:19 PM, Stefan Richter <s.rich...@data-artisans.com
> wrote:

> Hi,
>
> I suggest that you simply open an issue for this in our jira, describing
> the improvement idea. That should be the fastest way to get this changed.
>
> Best,
> Stefan
>
> Am 23.06.2017 um 15:08 schrieb Adarsh Jain <eradarshj...@gmail.com>:
>
> Hi Stefan,
>
> I think I found the problem, try it with a file which starts with
> underscore in the name like "_part-1-0.csv".
>
> While saving Flink appends a "_" to the file name however while reading at
> folder level it does not pick those files.
>
> Can you suggest if we can do a setting so that it does not pre appends
> underscore while saving a file.
>
> Regards,
> Adarsh
>
> On Fri, Jun 23, 2017 at 3:24 PM, Stefan Richter <
> s.rich...@data-artisans.com> wrote:
>
>> No, that doesn’t make a difference and also works.
>>
>> Am 23.06.2017 um 11:40 schrieb Adarsh Jain <eradarshj...@gmail.com>:
>>
>> I am using "val env = ExecutionEnvironment.getExecutionEnvironment", can
>> this be the problem?
>>
>> With "import org.apache.flink.api.scala.ExecutionEnvironment"
>>
>> Using scala in my program.
>>
>> Regards,
>> Adarsh
>>
>> On Fri, Jun 23, 2017 at 3:01 PM, Stefan Richter <
>> s.rich...@data-artisans.com> wrote:
>>
>>> I just copy pasted your code, adding the missing "val env
>>> = LocalEnvironment.createLocalEnvironment()" and exchanged the string
>>> with a local directory for some test files that I created. No other changes.
>>>
>>> Am 23.06.2017 um 11:25 schrieb Adarsh Jain <eradarshj...@gmail.com>:
>>>
>>> Hi Stefan,
>>>
>>> Thanks for your efforts in checking the same, still doesn't work for me.
>>>
>>> Can you copy paste the code you used maybe I am doing some silly mistake
>>> and am not able to figure out the same.
>>>
>>> Thanks again.
>>>
>>> Regards,
>>> Adarsh
>>>
>>>
>>> On Fri, Jun 23, 2017 at 2:32 PM, Stefan Richter <
>>> s.rich...@data-artisans.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> I tried this out on the current master and the 1.3 release and both
>>>> work for me everything works exactly as expected, for file names, a
>>>> directory, and even nested directories.
>>>>
>>>> Best,
>>>> Stefan
>>>>
>>>> Am 22.06.2017 um 21:13 schrieb Adarsh Jain <eradarshj...@gmail.com>:
>>>>
>>>> Hi Stefan,
>>>>
>>>> Yes your understood right, when I give full path till the filename it
>>>> works fine however when I give path till
>>>> directory it does not read the data, doesn't print any exceptions too
>>>> ... I am also not sure why it is behaving like this.
>>>>
>>>> Should be easily replicable, in case you can try. Will be really
>>>> helpful.
>>>>
>>>> Regards,
>>>> Adarsh
>>>>
>>>> On Thu, Jun 22, 2017 at 9:00 PM, Stefan Richter <
>>>> s.rich...@data-artisans.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I am not sure I am getting the problem right: the code works if you
>>>>> use a file name, but it does not work for directories? What exactly is not
>>>>> working? Do you get any exceptions?
>>>>>
>>>>> Best,
>>>>> Stefan
>>>>>
>>>>> Am 22.06.2017 um 17:01 schrieb Adarsh Jain <eradarshj...@gmail.com>:
>>>>>
>>>>> Hi,
>>>>>
>>>>> I am trying to use "Recursive Traversal of the Input Path Directory"
>>>>> in Flink 1.3 using scala. Snippet of my code below. If I give exact file
>>>>> name it is working fine. Ref https://ci.apache.org/proj
>>>>> ects/flink/flink-docs-release-1.3/dev/batch/index.html
>>>>>
>>>>> import org.apache.flink.api.java.utils.ParameterTool
>>>>> import org.apache.flink.api.java.{DataSet, ExecutionEnvironment}
>>>>> import org.apache.flink.configuration.Configuration
>>>>>
>>>>> val config = new Configuration
>>>>>     config.setBoolean("recursive.file.enumeration",true)
>>>>>
>>>>> val featuresSource: String = "file:///Users/adarsh/Document
>>>>> s/testData/featurecsv/31c710ac40/2017/06/22"
>>>>>
>>>>> val testInput = env.readTextFile(featuresSourc
>>>>> e).withParameters(config)
>>>>> testInput.print()
>>>>>
>>>>> Please guide how to fix this.
>>>>>
>>>>> Regards,
>>>>> Adarsh
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>>
>>
>>
>
>

Reply via email to