forward to mail list for discussion
On Mon, May 4, 2015 at 3:39 PM, hongbin ma <[email protected]> wrote: > WARNING: Failed to process JAR > [jar:file:/data1/app/hadoop-2.6.0/contrib/capacity-scheduler/.jar!/] for > TLD files > java.io.FileNotFoundException: > /data1/app/hadoop-2.6.0/contrib/capacity-scheduler/.jar (No such file or > directory) > at java.util.zip.ZipFile.open(Native Method) > at java.util.zip.ZipFile.(ZipFile.java:215) > at java.util.zip.ZipFile.(ZipFile.java:145) > at java.util.jar.JarFile.(JarFile.java:153) > at java.util.jar.JarFile.(JarFile.java:90) > at sun.net.www.protocol.jar.URLJarFile.(URLJarFile.java:93) > at sun.net.www.protocol.jar.URLJarFile.getJarFile(URLJarFile.java:69) > at sun.net.www.protocol.jar.JarFileFactory.get(JarFileFactory.java:99) > at > sun.net.www.protocol.jar.JarURLConnection.connect(JarURLConnection.java:122) > at > sun.net.www.protocol.jar.JarURLConnection.getJarFile(JarURLConnection.java:89) > at org.apache.tomcat.util.scan.FileUrlJar.(FileUrlJar.java:41) > at org.apache.tomcat.util.scan.JarFactory.newInstance(JarFactory.java:34) > at org.apache.catalina.startup.TldConfig.tldScanJar(TldConfig.java:485) > at org.apache.catalina.startup.TldConfig.access$100(TldConfig.java:61) > at > org.apache.catalina.startup.TldConfig$TldJarScannerCallback.scan(TldConfig.java:296) > at > org.apache.tomcat.util.scan.StandardJarScanner.process(StandardJarScanner.java:258) > at > org.apache.tomcat.util.scan.StandardJarScanner.scan(StandardJarScanner.java:220) > at org.apache.catalina.startup.TldConfig.execute(TldConfig.java:269) > at org.apache.catalina.startup.TldConfig.lifecycleEvent(TldConfig.java:565) > at > org.apache.catalina.util.LifecycleSupport.fireLifecycleEvent(LifecycleSupport.java:117) > at > org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:90) > at > org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5412) > at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:150) > at > org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:901) > at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:877) > at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:649) > at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1081) > at > org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1877) > at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) > at java.lang.Thread.run(Thread.java:744) > hadoop2.x contrib文件夹应该是没有的,这个为什么还会用这个文件夹? > > -- Regards, *Bin Mahone | 马洪宾* Apache Kylin: http://kylin.io Github: https://github.com/binmahone
