It works with Twill-0.9.0. So far I have been able to narrow it down to one commit
5986553 (TWILL-63) Speed up application launch time Let me see if can nail down to a particular change. Sam > On Mar 25, 2017, at 13:34, Sam William <[email protected]> wrote: > > HI Terence, > Our cloudera installation is CDH-5.7 and I use hadoop 2.3.0 packages for my > fat jars. > > SAm >> On Mar 25, 2017, at 12:31, Terence Yim <[email protected]> wrote: >> >> Hi, >> >> Haven't seen this error before. What is the version of Hadoop that the >> cluster is running with? Also, seems like the $HADOOP_CONF is not in the >> classpath as the FileContext is trying to use local file system instead of >> the distributed one. >> >> Terence >> >> Sent from my iPhone >> >>> On Mar 25, 2017, at 12:25 PM, Sam William <[email protected]> wrote: >>> >>> Hi, >>> I have been using Twill for sometime now and I just tried to upgrade our >>> application from Twill-0.8.0 to 0.10.0. I havent made any kind of code >>> changes besides changing the Twill version string in the build script. The >>> application fails immediately and I see this on the RM UI. Any idea why >>> this could be happening? >>> >>> Diagnostics: >>> Application application_1484158548936_11154 failed 2 times due to AM >>> Container for appattempt_1484158548936_11154_000002 exited with exitCode: >>> -1000 >>> For more detailed output, check application tracking page<> Then, click on >>> links to logs of each attempt. >>> Diagnostics: No such file or directory >>> ENOENT: No such file or directory >>> at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmodImpl(Native Method) >>> at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmod(NativeIO.java:230) >>> at >>> org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:660) >>> at >>> org.apache.hadoop.fs.DelegateToFileSystem.setPermission(DelegateToFileSystem.java:206) >>> at org.apache.hadoop.fs.FilterFs.setPermission(FilterFs.java:251) >>> at org.apache.hadoop.fs.FileContext$10.next(FileContext.java:955) >>> at org.apache.hadoop.fs.FileContext$10.next(FileContext.java:951) >>> at org.apache.hadoop.fs.FSLinkResolver.resolve(FSLinkResolver.java:90) >>> at org.apache.hadoop.fs.FileContext.setPermission(FileContext.java:951) >>> at org.apache.hadoop.yarn.util.FSDownload$3.run(FSDownload.java:419) >>> at org.apache.hadoop.yarn.util.FSDownload$3.run(FSDownload.java:417) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:422) >>> at >>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) >>> at >>> org.apache.hadoop.yarn.util.FSDownload.changePermissions(FSDownload.java:417) >>> at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:363) >>> at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>> at java.lang.Thread.run(Thread.java:745) >>> Failing this attempt. Failing the application. >>> >>> >>> Sam >
