[jira] [Created] (FLINK-6886) Fix Timestamp field can not be selected in event time case when toDataStream[T], `T` not a `Row` Type.

2017-06-11 Thread sunjincheng (JIRA)
sunjincheng created FLINK-6886:
--

 Summary: Fix Timestamp field can not be selected in event time 
case when  toDataStream[T], `T` not a `Row` Type.
 Key: FLINK-6886
 URL: https://issues.apache.org/jira/browse/FLINK-6886
 Project: Flink
  Issue Type: Bug
  Components: Table API & SQL
Affects Versions: 1.4.0
Reporter: sunjincheng
Assignee: sunjincheng


Currently for event-time window(group/over), When contain `Timestamp` type 
field in `SELECT Clause`, And toDataStream[T], `T` not a `Row` Type, Such 
`PojoType`, will throw a exception. In this JIRA. will fix this bug.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (FLINK-6887) Split up CodeGenerator into several specific CodeGenerator

2017-06-11 Thread Jark Wu (JIRA)
Jark Wu created FLINK-6887:
--

 Summary: Split up CodeGenerator into several specific CodeGenerator
 Key: FLINK-6887
 URL: https://issues.apache.org/jira/browse/FLINK-6887
 Project: Flink
  Issue Type: Bug
  Components: Table API & SQL
Reporter: Jark Wu
 Fix For: 1.4.0


Currently, the {{CodeGenerator}} is very huge and a bit difficult to maintain. 
I suggest to split it up into several specific {{XXXCodeGenerator}}.

For example, create {{AggregationFunctionCodeGenerator}}  class and make it 
extend to {{CodeGenerator}} and move the {{def generateAggregations(...)}} 
method to it. The same as {{TableFunctionCollectorCodeGenerator}} and 
{{InputFormatCodeGenerator}}.

What do you think? [~fhueske], [~twalthr], [~sunjincheng121]



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


[jira] [Created] (FLINK-6888) Can not determine TypeInformation of ACC type of AggregateFunction when ACC is a Scala case/tuple class

2017-06-11 Thread Jark Wu (JIRA)
Jark Wu created FLINK-6888:
--

 Summary: Can not determine TypeInformation of ACC type of 
AggregateFunction when ACC is a Scala case/tuple class
 Key: FLINK-6888
 URL: https://issues.apache.org/jira/browse/FLINK-6888
 Project: Flink
  Issue Type: Bug
  Components: Table API & SQL
Reporter: Jark Wu
Assignee: Jark Wu
 Fix For: 1.4.0


Currently the {{ACC}} TypeInformation of 
{{org.apache.flink.table.functions.AggregateFunction[T, ACC]}} is extracted 
using {{TypeInformation.of(Class)}}. When {{ACC}} is a Scala case class or 
tuple class, the TypeInformation will fall back to {{GenericType}} which result 
in bad performance when state de/serialization. 

I suggest to extract the ACC TypeInformation when called 
{{TableEnvironment.registerFunction()}}.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)


Re: FlinkML on slack

2017-06-11 Thread Ziyad Muhammed
Hi Stavros

Could you please send me an invite to the slack channel?

Best
Ziyad


On Sun, Jun 11, 2017 at 1:53 AM, Stavros Kontopoulos <
st.kontopou...@gmail.com> wrote:

> @Henry @Tao @Martin invitations sent... Thnx @Theo for handling the Apache
> compliance issues.
>
> Best,
> Stavros
>
> On Sat, Jun 10, 2017 at 10:27 PM, Henry Saputra 
> wrote:
>
> > Hi Stavros,
> >
> > Could you also send me invite to the Slack?
> >
> > My email is hsapu...@apache.org
> >
> > Thanks,
> >
> > Henry
> >
> >
> > On Thu, Jun 8, 2017 at 2:21 AM, Stavros Kontopoulos <
> > st.kontopou...@gmail.com> wrote:
> >
> > > Hi Aljoscha,
> > >
> > > Slack is invite only to the best of my knowledge, I just sent you an
> > > invitation.
> > >
> > > Best,
> > > Stavros
> > >
> > >
> > > On Thu, Jun 8, 2017 at 11:31 AM, Aljoscha Krettek  >
> > > wrote:
> > >
> > > > Hi,
> > > >
> > > > Is the slack invite based? If yes, could you please send me one?
> > > >
> > > > Best,
> > > > Aljoscha
> > > >
> > > > > On 7. Jun 2017, at 21:56, Stavros Kontopoulos <
> > > st.kontopou...@gmail.com>
> > > > wrote:
> > > > >
> > > > > Hi all,
> > > > >
> > > > > We took the initiative to create the organization for FlinkML on
> > slack
> > > > > (thnx Eron).
> > > > > There is now a channel for model-serving
> > > > >  > > > fdEXPsPYPEywsE/edit#>.
> > > > > Another is coming for flink-jpmml.
> > > > > You are invited to join the channels and the efforts. @Gabor @Theo
> > > please
> > > > > consider adding channels for the other efforts there as well.
> > > > >
> > > > > FlinkMS on Slack  (
> > > > https://flinkml.slack.com/)
> > > > >
> > > > > Details for the efforts here: Flink Roadmap doc
> > > > >  > > > d06MIRhahtJ6dw/edit#>
> > > > >
> > > > > Github  (https://github.com/FlinkML)
> > > > >
> > > > >
> > > > > Stavros
> > > >
> > > >
> > >
> >
>


Re: [DISCUSS] GitBox

2017-06-11 Thread Ufuk Celebi
I would also like to see this happening for both flink-web and flink
if it allows committers to have control over the respective repos.

On Sat, Jun 10, 2017 at 4:05 PM, Chesnay Schepler  wrote:
> What are the downsides of this? Actually, is there any ASF resource that
> outlines what this would enable?
>
> In one of the threads i saw said that this would also allow committers to
> close PR's, assign labels and such.
> This sounds very interesting to me for the main repo actually.
>
>
> On 09.06.2017 17:41, Greg Hogan wrote:
>>
>> Robert has an open PR from March. I’ve found, for example, PRs adding
>> links to talks or slides left open for months.
>>
>> I’d suggest Fluo is to Accumulo as flink-web is to the flink repo, and
>> that migration looks to be satisfactory.
>>
>>
>>> On Jun 9, 2017, at 11:15 AM, Ted Yu  wrote:
>>>
>>> bq. better track the oft-neglected contributions
>>>
>>> Do you have estimate on how many contributions were not paid attention in
>>> the current infrastructure.
>>>
>>> Looking at #2, it seems Accumulo community hasn't reached consensus yet.
>>>
>>> Cheers
>>>
>>> On Fri, Jun 9, 2017 at 7:54 AM, Greg Hogan  wrote:
>>>
 All,

 ASF now has available (and maybe mandatory for new projects or repos)
 GitBox [0] which enables bi-directional sync to GitHub and links
 committers' accounts, allowing for greater use of GitHub functionality
 by
 contributors and for committers to perform many tasks otherwise
 requiring
 INFRA tickets.

 I'd like to propose moving flink-web [1] to GitBox, using GitHub issues,
 and enabling notifications to the mailing lists. Apache Accumulo has
 recently discussed [2] this topic with a list of benefits after
 migrating
 Fluo. By migrating flink-web we can better track the oft-neglected
 contributions and also test the waters for future migrations (perhaps
 for
 the future sub-projects).

 [0] https://gitbox.apache.org/
 [1] https://github.com/apache/flink-web/pulls
 [2]
 http://apache-accumulo.1065345.n5.nabble.com/DISCUSS-
 GitBox-tp21160p21497.html

 Greg
>
>
>


[QUESTION] OutOfMemoryError when writing into HDFS

2017-06-11 Thread Mu Kong
Hi all,

Thanks a lot for your work for the community!

This question is more like a discuss.
Currently, I'm experiencing an *OutOfMemoryError* when writing files
from Kafka into HDFS using *BucketingSink*.

The log looks like this:

2017-06-10 08:58:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  -
Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33283, GC
COUNT: 977], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
2017-06-10 08:59:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  - Memory
usage stats: [HEAP: 14080/20480/20480 MB, NON HEAP: 81/83/-1 MB
(used/committed/max)]
2017-06-10 08:59:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  - Direct
memory stats: Count: 16846, Total Capacity: 443738663, Used Memory:
443738664
2017-06-10 08:59:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  -
Off-heap pool stats: [Code Cache: 24/25/240 MB (used/committed/max)],
[Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
Space: 6/6/1024 MB (used/committed/max)]
2017-06-10 08:59:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  -
Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33311, GC
COUNT: 978], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
2017-06-10 09:00:20,485 INFO
org.apache.flink.shaded.org.apache.curator.framework.imps.CuratorFrameworkImpl
 - backgroundOperationsLoop exiting
2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ZooKeeper
 - Session: 0x55591b10666ea92 closed
2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ClientCnxn
 - EventThread shut down
2017-06-10 09:00:34,999 INFO
org.apache.flink.shaded.org.apache.curator.framework.imps.CuratorFrameworkImpl
 - backgroundOperationsLoop exiting
2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ZooKeeper
 - Session: 0x55591b10666ea94 closed
2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ClientCnxn
 - EventThread shut down
2017-06-10 09:00:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  - Memory
usage stats: [HEAP: 12820/20480/20480 MB, NON HEAP: 81/83/-1 MB
(used/committed/max)]
2017-06-10 09:00:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  - Direct
memory stats: Count: 17438, Total Capacity: 458405794, Used Memory:
458405795
2017-06-10 09:00:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  -
Off-heap pool stats: [Code Cache: 25/25/240 MB (used/committed/max)],
[Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
Space: 6/6/1024 MB (used/committed/max)]
2017-06-10 09:00:42,248 INFO
org.apache.flink.runtime.taskmanager.TaskManager  -
Garbage collector stats: [G1 Young Generation, GC TIME (ms): 9, GC
COUNT: 979], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
2017-06-10 09:01:04,962 INFO
org.apache.flink.runtime.taskmanager.Task -
Source: Custom Source -> Sink: Unnamed (7/22)
(57d3c79ae13fd06de79ca6cb8f1431b4) switched from RUNNING to FAILED.
java.lang.OutOfMemoryError: unable to create new native thread
at java.lang.Thread.start0(Native Method)
at java.lang.Thread.start(Thread.java:714)
at org.apache.hadoop.hdfs.DFSOutputStream.start(DFSOutputStream.java:2170)
at 
org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1685)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689)
at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624)
at 
org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:448)
at 
org.apache.hadoop.hdfs.DistributedFileSystem$7.doCall(DistributedFileSystem.java:444)
at 
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
at 
org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:459)
at 
org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:387)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
at 
org.apache.flink.streaming.connectors.fs.StreamWriterBase.open(StreamWriterBase.java:126)
at 
org.apache.flink.streaming.connectors.fs.StringWriter.open(StringWriter.java:62)
at 
org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.openNewPartFile(BucketingSink.java:546)
at 
org.apache.flink.streaming.connectors.fs.bucketing.BucketingSink.invoke(BucketingSink.java:441)
at 
org.apache.flink.streaming.api.operators.StreamSink.processElement(StreamSink.java:41)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.pushToOperator(OperatorChain.java:528)
at 
org.apache.flink.streaming.runtime.tasks.OperatorChain$CopyingChainingOutput.collect(OperatorChain.java:503

Re: [QUESTION] OutOfMemoryError when writing into HDFS

2017-06-11 Thread Ted Yu
Can you see if the following post helps in troubleshooting ?

https://blog.fastthread.io/2016/07/06/troubleshoot-outofmemoryerror-unable-to-create-new-native-thread/

Thanks

On Sun, Jun 11, 2017 at 6:01 PM, Mu Kong  wrote:

> Hi all,
>
> Thanks a lot for your work for the community!
>
> This question is more like a discuss.
> Currently, I'm experiencing an *OutOfMemoryError* when writing files
> from Kafka into HDFS using *BucketingSink*.
>
> The log looks like this:
>
> 2017-06-10 08:58:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  -
> Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33283, GC
> COUNT: 977], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> 2017-06-10 08:59:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  - Memory
> usage stats: [HEAP: 14080/20480/20480 MB, NON HEAP: 81/83/-1 MB
> (used/committed/max)]
> 2017-06-10 08:59:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  - Direct
> memory stats: Count: 16846, Total Capacity: 443738663, Used Memory:
> 443738664
> 2017-06-10 08:59:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  -
> Off-heap pool stats: [Code Cache: 24/25/240 MB (used/committed/max)],
> [Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
> Space: 6/6/1024 MB (used/committed/max)]
> 2017-06-10 08:59:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  -
> Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33311, GC
> COUNT: 978], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> 2017-06-10 09:00:20,485 INFO
> org.apache.flink.shaded.org.apache.curator.framework.imps.
> CuratorFrameworkImpl
>  - backgroundOperationsLoop exiting
> 2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ZooKeeper
>  - Session: 0x55591b10666ea92 closed
> 2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ClientCnxn
>  - EventThread shut down
> 2017-06-10 09:00:34,999 INFO
> org.apache.flink.shaded.org.apache.curator.framework.imps.
> CuratorFrameworkImpl
>  - backgroundOperationsLoop exiting
> 2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ZooKeeper
>  - Session: 0x55591b10666ea94 closed
> 2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ClientCnxn
>  - EventThread shut down
> 2017-06-10 09:00:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  - Memory
> usage stats: [HEAP: 12820/20480/20480 MB, NON HEAP: 81/83/-1 MB
> (used/committed/max)]
> 2017-06-10 09:00:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  - Direct
> memory stats: Count: 17438, Total Capacity: 458405794, Used Memory:
> 458405795
> 2017-06-10 09:00:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  -
> Off-heap pool stats: [Code Cache: 25/25/240 MB (used/committed/max)],
> [Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
> Space: 6/6/1024 MB (used/committed/max)]
> 2017-06-10 09:00:42,248 INFO
> org.apache.flink.runtime.taskmanager.TaskManager  -
> Garbage collector stats: [G1 Young Generation, GC TIME (ms): 9, GC
> COUNT: 979], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> 2017-06-10 09:01:04,962 INFO
> org.apache.flink.runtime.taskmanager.Task -
> Source: Custom Source -> Sink: Unnamed (7/22)
> (57d3c79ae13fd06de79ca6cb8f1431b4) switched from RUNNING to FAILED.
> java.lang.OutOfMemoryError: unable to create new native thread
> at java.lang.Thread.start0(Native Method)
> at java.lang.Thread.start(Thread.java:714)
> at org.apache.hadoop.hdfs.DFSOutputStream.start(
> DFSOutputStream.java:2170)
> at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(
> DFSOutputStream.java:1685)
> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689)
> at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624)
> at org.apache.hadoop.hdfs.DistributedFileSystem$7.
> doCall(DistributedFileSystem.java:448)
> at org.apache.hadoop.hdfs.DistributedFileSystem$7.
> doCall(DistributedFileSystem.java:444)
> at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(
> FileSystemLinkResolver.java:81)
> at org.apache.hadoop.hdfs.DistributedFileSystem.create(
> DistributedFileSystem.java:459)
> at org.apache.hadoop.hdfs.DistributedFileSystem.create(
> DistributedFileSystem.java:387)
> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:909)
> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:890)
> at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:787)
> at org.apache.flink.streaming.connectors.fs.StreamWriterBase.open(
> StreamWriterBase.java:126)
> at org.apache.flink.streaming.connectors.fs.StringWriter.
> open(StringWriter.java:62)
> at org.apache.flink.streaming.connectors.fs.bucketing.
> BucketingSink.openNewPartFile(BucketingSink.java:546)
> at org.apache.fl

Re: [QUESTION] OutOfMemoryError when writing into HDFS

2017-06-11 Thread Mu Kong
Hi Ted,

Thanks for the useful link there.

Actually, I just found out that the error might be triggered by a kerberos
issue in our system.
I'll use the ticket cache and execute kinit in crontab to see if it is the
case.

But thanks for the link you provided. I'll definitely check it :)

Best regards,
Mu

On Mon, Jun 12, 2017 at 10:45 AM, Ted Yu  wrote:

> Can you see if the following post helps in troubleshooting ?
>
> https://blog.fastthread.io/2016/07/06/troubleshoot-
> outofmemoryerror-unable-to-create-new-native-thread/
>
> Thanks
>
> On Sun, Jun 11, 2017 at 6:01 PM, Mu Kong  wrote:
>
> > Hi all,
> >
> > Thanks a lot for your work for the community!
> >
> > This question is more like a discuss.
> > Currently, I'm experiencing an *OutOfMemoryError* when writing files
> > from Kafka into HDFS using *BucketingSink*.
> >
> > The log looks like this:
> >
> > 2017-06-10 08:58:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  -
> > Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33283, GC
> > COUNT: 977], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> > 2017-06-10 08:59:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  - Memory
> > usage stats: [HEAP: 14080/20480/20480 MB, NON HEAP: 81/83/-1 MB
> > (used/committed/max)]
> > 2017-06-10 08:59:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  - Direct
> > memory stats: Count: 16846, Total Capacity: 443738663, Used Memory:
> > 443738664
> > 2017-06-10 08:59:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  -
> > Off-heap pool stats: [Code Cache: 24/25/240 MB (used/committed/max)],
> > [Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
> > Space: 6/6/1024 MB (used/committed/max)]
> > 2017-06-10 08:59:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  -
> > Garbage collector stats: [G1 Young Generation, GC TIME (ms): 33311, GC
> > COUNT: 978], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> > 2017-06-10 09:00:20,485 INFO
> > org.apache.flink.shaded.org.apache.curator.framework.imps.
> > CuratorFrameworkImpl
> >  - backgroundOperationsLoop exiting
> > 2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ZooKeeper
> >  - Session: 0x55591b10666ea92 closed
> > 2017-06-10 09:00:20,488 INFO  org.apache.zookeeper.ClientCnxn
> >  - EventThread shut down
> > 2017-06-10 09:00:34,999 INFO
> > org.apache.flink.shaded.org.apache.curator.framework.imps.
> > CuratorFrameworkImpl
> >  - backgroundOperationsLoop exiting
> > 2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ZooKeeper
> >  - Session: 0x55591b10666ea94 closed
> > 2017-06-10 09:00:35,001 INFO  org.apache.zookeeper.ClientCnxn
> >  - EventThread shut down
> > 2017-06-10 09:00:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  - Memory
> > usage stats: [HEAP: 12820/20480/20480 MB, NON HEAP: 81/83/-1 MB
> > (used/committed/max)]
> > 2017-06-10 09:00:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  - Direct
> > memory stats: Count: 17438, Total Capacity: 458405794, Used Memory:
> > 458405795
> > 2017-06-10 09:00:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  -
> > Off-heap pool stats: [Code Cache: 25/25/240 MB (used/committed/max)],
> > [Metaspace: 50/51/-1 MB (used/committed/max)], [Compressed Class
> > Space: 6/6/1024 MB (used/committed/max)]
> > 2017-06-10 09:00:42,248 INFO
> > org.apache.flink.runtime.taskmanager.TaskManager  -
> > Garbage collector stats: [G1 Young Generation, GC TIME (ms): 9, GC
> > COUNT: 979], [G1 Old Generation, GC TIME (ms): 325, GC COUNT: 2]
> > 2017-06-10 09:01:04,962 INFO
> > org.apache.flink.runtime.taskmanager.Task -
> > Source: Custom Source -> Sink: Unnamed (7/22)
> > (57d3c79ae13fd06de79ca6cb8f1431b4) switched from RUNNING to FAILED.
> > java.lang.OutOfMemoryError: unable to create new native thread
> > at java.lang.Thread.start0(Native Method)
> > at java.lang.Thread.start(Thread.java:714)
> > at org.apache.hadoop.hdfs.DFSOutputStream.start(
> > DFSOutputStream.java:2170)
> > at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(
> > DFSOutputStream.java:1685)
> > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1689)
> > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1624)
> > at org.apache.hadoop.hdfs.DistributedFileSystem$7.
> > doCall(DistributedFileSystem.java:448)
> > at org.apache.hadoop.hdfs.DistributedFileSystem$7.
> > doCall(DistributedFileSystem.java:444)
> > at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(
> > FileSystemLinkResolver.java:81)
> > at org.apache.hadoop.hdfs.DistributedFileSystem.create(
> > DistributedFileSystem.java:459)
> > at org.apache.hadoop.hdfs.DistributedFileSystem.create(
> > DistributedFileSystem.java: