How much effort we need to put in to create 2.12 module? Is that just a change in POM files or code change is required?
We have release planned for July to upgrade Jupyter and Livy to utilize spark 2.4.2. This is blocking us from upgrade. From: Saisai Shao <sai.sai.s...@gmail.com<mailto:sai.sai.s...@gmail.com>> Date: Monday, Jun 03, 2019, 9:02 PM To: user@livy.incubator.apache.org <user@livy.incubator.apache.org<mailto:user@livy.incubator.apache.org>> Subject: [External] Re: Support for Livy with Scala 2.12 Like what we did before to support both Scala 2.10 and 2.11 in Livy, I think we should also have a new module to support 2.12. <santosh.dan...@ubs.com<mailto:santosh.dan...@ubs.com>> 于2019年6月4日周二 上午7:40写道: Yes, the spark binary we downloaded is built with default Scala 2.12. We want to use databricks delta which I think only support Scala 2.12. So, I'm stuck with Scala 2.12. Moreover, Spark community is going to decommission Scala 2.11 completely from Spark 3.0 release. We might need to prepare Livy to support Scala 2.12 by default. From: Kevin Risden [mailto:kris...@apache.org<mailto:kris...@apache.org>] Sent: Monday, June 03, 2019 6:35 PM To: user@livy.incubator.apache.org<mailto:user@livy.incubator.apache.org> Subject: [External] Re: Support for Livy with Scala 2.12 Looks like the issue might be Spark 2.4.2 only? From https://spark.apache.org/downloads.html, "Note that, Spark is pre-built with Scala 2.11 except version 2.4.2, which is pre-built with Scala 2.12." So maybe you just got unlucky with using Spark 2.4.2? Kevin Risden On Mon, Jun 3, 2019 at 6:19 PM <santosh.dan...@ubs.com<mailto:santosh.dan...@ubs.com>> wrote: Kevin, I'm using Livy 0.6.0. The issues is related to not finding repl jars that support scala 2.12. The error "requirement failed: Cannot find Livy REPL jars." is thrown because it couldn't find folder repl_2.12-jars under LIVY directory. I performed a test to make sure this issue is related to scala 2.12 compatibility , I copied contents of repl_2.11-jars under Livy directory into new directory LIVY/repl_2.12-jars and this time I didn't get REPL jars exception it went ahead and created session but failed to start session due to rsc jars version incompatibility. LIVY Folder structure for error " requirement failed: Cannot find Livy REPL jars."" [/app/risk/ha02/livy]$ ls -ltr total 116 -rwxr-xr-x 1 agriddev agriddev 160 Mar 19 14:39 NOTICE -rwxr-xr-x 1 agriddev agriddev 18665 Mar 19 14:39 LICENSE -rwxr-xr-x 1 agriddev agriddev 537 Mar 19 14:39 DISCLAIMER -rwxr-xr-x 1 agriddev agriddev 46355 Mar 19 14:42 THIRD-PARTY drwxr-xr-x 2 agriddev agriddev 4096 Mar 19 14:43 bin drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 repl_2.11-jars drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 rsc-jars drwxr-xr-x 2 agriddev agriddev 12288 Apr 14 22:37 jars drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 apache-livy-0.6.0-incubating-bin drwxr-xr-x 2 agriddev agriddev 4096 Jun 3 17:37 conf drwxr-xr-x 2 agriddev agriddev 4096 Jun 3 21:51 logs LIVY FOLDER STRUCTURE TO BYPASS "REQUIREMENT FAILED:CANNOT FIND LIVY REPL JARS" [/app/risk/ha02/livy]$ ls -ltr total 116 -rwxr-xr-x 1 agriddev agriddev 160 Mar 19 14:39 NOTICE -rwxr-xr-x 1 agriddev agriddev 18665 Mar 19 14:39 LICENSE -rwxr-xr-x 1 agriddev agriddev 537 Mar 19 14:39 DISCLAIMER -rwxr-xr-x 1 agriddev agriddev 46355 Mar 19 14:42 THIRD-PARTY drwxr-xr-x 2 agriddev agriddev 4096 Mar 19 14:43 bin drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 repl_2.11-jars drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 rsc-jars drwxr-xr-x 2 agriddev agriddev 12288 Apr 14 22:37 jars drwxr-xr-x 2 agriddev agriddev 4096 Apr 14 22:37 apache-livy-0.6.0-incubating-bin drwxr-xr-x 2 agriddev agriddev 4096 Jun 3 17:37 conf drwxr-xr-x 2 agriddev agriddev 4096 Jun 3 21:50 repl_2.12-jars drwxr-xr-x 2 agriddev agriddev 4096 Jun 3 21:51 logs Error Information zip 19/06/03 21:52:00 INFO LineBufferedStream: 19/06/03 21:52:00 INFO SecurityManager: Changing view acls to: agriddev 19/06/03 21:52:00 INFO LineBufferedStream: 19/06/03 21:52:00 INFO SecurityManager: Changing modify acls to: agriddev 19/06/03 21:52:00 INFO LineBufferedStream: 19/06/03 21:52:00 INFO SecurityManager: Changing view acls groups to: 19/06/03 21:52:00 INFO LineBufferedStream: 19/06/03 21:52:00 INFO SecurityManager: Changing modify acls groups to: 19/06/03 21:52:00 INFO LineBufferedStream: 19/06/03 21:52:00 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(agriddev); groups with view permissions: Set(); users with modify permissions: Set(agriddev); groups with modify permissions: Set() 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO Client: Submitting application application_1559316432251_0172 to ResourceManager 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO YarnClientImpl: Submitted application application_1559316432251_0172 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO Client: Application report for application_1559316432251_0172 (state: ACCEPTED) 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO Client: 19/06/03 21:52:01 INFO LineBufferedStream: client token: N/A 19/06/03 21:52:01 INFO LineBufferedStream: diagnostics: [Mon Jun 03 21:52:01 +0000 2019] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:2150400, vCores:180> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; 19/06/03 21:52:01 INFO LineBufferedStream: ApplicationMaster host: N/A 19/06/03 21:52:01 INFO LineBufferedStream: ApplicationMaster RPC port: -1 19/06/03 21:52:01 INFO LineBufferedStream: queue: default 19/06/03 21:52:01 INFO LineBufferedStream: start time: 1559598721629 19/06/03 21:52:01 INFO LineBufferedStream: final status: UNDEFINED 19/06/03 21:52:01 INFO LineBufferedStream: tracking URL: http://xzur1315dap.zur.swissbank.com:8088/proxy/application_1559316432251_0172/ 19/06/03 21:52:01 INFO LineBufferedStream: user: agriddev 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO ShutdownHookManager: Shutdown hook called 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO ShutdownHookManager: Deleting directory /tmp/spark-74eee398-fced-4173-8682-b512a95adea6 19/06/03 21:52:01 INFO LineBufferedStream: 19/06/03 21:52:01 INFO ShutdownHookManager: Deleting directory /app/risk/ds2/nvme0n1/ds2_spark_cluster_integration/tmp/spark_tmp/spark-91e38128-8064-409b-b5de-3d012b6ad81d 19/06/03 21:52:08 WARN RSCClient: Client RPC channel closed unexpectedly. 19/06/03 21:52:08 WARN RSCClient: Error stopping RPC. io.netty.util.concurrent.BlockingOperationException: DefaultChannelPromise@21867c24(uncancellable) at io.netty.util.concurrent.DefaultPromise.checkDeadLock(DefaultPromise.java:394) at io.netty.channel.DefaultChannelPromise.checkDeadLock(DefaultChannelPromise.java:157) at io.netty.util.concurrent.DefaultPromise.await(DefaultPromise.java:230) at io.netty.channel.DefaultChannelPromise.await(DefaultChannelPromise.java:129) at io.netty.channel.DefaultChannelPromise.await(DefaultChannelPromise.java:28) at io.netty.util.concurrent.DefaultPromise.sync(DefaultPromise.java:336) at io.netty.channel.DefaultChannelPromise.sync(DefaultChannelPromise.java:117) at io.netty.channel.DefaultChannelPromise.sync(DefaultChannelPromise.java:28) at org.apache.livy.rsc.rpc.Rpc.close(Rpc.java:310) at org.apache.livy.rsc.RSCClient.stop(RSCClient.java:232) at org.apache.livy.rsc.RSCClient$2$1.onSuccess(RSCClient.java:129) at org.apache.livy.rsc.RSCClient$2$1.onSuccess(RSCClient.java:123) at org.apache.livy.rsc.Utils$2.operationComplete(Utils.java:108) at io.netty.util.concurrent.DefaultPromise.notifyListener0(DefaultPromise.java:518) at io.netty.util.concurrent.DefaultPromise.notifyListenersNow(DefaultPromise.java:492) at io.netty.util.concurrent.DefaultPromise.notifyListenersWithStackOverFlowProtection(DefaultPromise.java:431) at io.netty.util.concurrent.DefaultPromise.notifyListeners(DefaultPromise.java:420) at io.netty.util.concurrent.DefaultPromise.trySuccess(DefaultPromise.java:108) at io.netty.channel.DefaultChannelPromise.trySuccess(DefaultChannelPromise.java:82) at io.netty.channel.AbstractChannel$CloseFuture.setClosed(AbstractChannel.java:995) at io.netty.channel.AbstractChannel$AbstractUnsafe.doClose0(AbstractChannel.java:621) at io.netty.channel.AbstractChannel$AbstractUnsafe.close(AbstractChannel.java:599) at io.netty.channel.AbstractChannel$AbstractUnsafe.close(AbstractChannel.java:543) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.closeOnRead(AbstractNioByteChannel.java:71) at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:158) at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:564) at io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:505) at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:419) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:391) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:112) at java.lang.Thread.run(Thread.java:748) 19/06/03 21:52:08 INFO RSCClient: Failing pending job 4d7789d1-a460-48c0-85fe-63fa24bb95a8 due to shutdown. 19/06/03 21:52:08 INFO InteractiveSession: Stopping InteractiveSession 0... 19/06/03 21:52:08 INFO InteractiveSession: Failed to ping RSC driver for session 0. Killing application. 19/06/03 21:52:09 INFO YarnClientImpl: Killed application application_1559316432251_0172 19/06/03 21:52:09 INFO InteractiveSession: Stopped InteractiveSession 0. ^C From: Kevin Risden [mailto:kris...@apache.org<mailto:kris...@apache.org>] Sent: Monday, June 03, 2019 4:46 PM To: user@livy.incubator.apache.org<mailto:user@livy.incubator.apache.org> Cc: Dandey, Santosh Subject: [External] Re: Support for Livy with Scala 2.12 "requirement failed: Cannot find Livy REPL jars." I didn't look where that error comes from, but my guess is that it looks like you don't have Livy pointing to the right location anymore where it can find Spark. Hopefully not sending you on a wild goose chase, but would check there first. Also you need to make sure you are Livy 0.6.0+ since Spark 2.4 support was only added to 0.6.0 and above. Kevin Risden On Mon, Jun 3, 2019 at 4:04 PM Pat Ferrel <p...@occamsmachete.com<mailto:p...@occamsmachete.com>> wrote: Spark 2.4.x does not require scala 2.12, in fact is is marked as “experimental” here: https://spark.apache.org/releases/spark-release-2-4-0.html <https://spark.apache.org/releases/spark-release-2-4-0.html> Moving to a new scala version is often a pain, because the libs you use may not be upgraded and version matter *unlike typical Java updates). Scala creates JVM objects and names them as it pleases. Sometimes naming changes from version to version of Scala and this causes big problem in using mixed libs from different versions of Scala. I’m no expert in Livy, but imagine you may need to build against a newer Spark. But avoid Scala 2.12 for now. From: santosh.dan...@ubs.com<mailto:santosh.dan...@ubs.com> <santosh.dan...@ubs.com><mailto:santosh.dan...@ubs.com> Reply: user@livy.incubator.apache.org<mailto:user@livy.incubator.apache.org> <user@livy.incubator.apache.org><mailto:user@livy.incubator.apache.org> Date: June 3, 2019 at 12:51:20 PM To: user@livy.incubator.apache.org<mailto:user@livy.incubator.apache.org> <user@livy.incubator.apache.org><mailto:user@livy.incubator.apache.org> Subject: Support for Livy with Scala 2.12 Hi, We have just upgraded our spark cluster version 2.3 to 2.4.2 and it broke Livy. It's throwing exception "Cannot Find Livy REPL Jars". Looks like I have to build Livy using Scala 2.12 version. Can anyone advise how to build Livy with Scala 2.12 with Maven? Will changing the scala version from 2.11 to 2.12 would build livy? Please advise. The code failed because of a fatal error: Invalid status code '400' from http://localhost:8998/sessions with error payload: {"msg":"requirement failed: Cannot find Livy REPL jars."}. Thanks Santosh Please visit our website at http://financialservicesinc.ubs.com/wealth/E-maildisclaimer.html for important disclosures and information about our e-mail policies. For your protection, please do not transmit orders or instructions by e-mail or include account numbers, Social Security numbers, credit card numbers, passwords, or other personal information. Please visit our website at http://financialservicesinc.ubs.com/wealth/E-maildisclaimer.html for important disclosures and information about our e-mail policies. For your protection, please do not transmit orders or instructions by e-mail or include account numbers, Social Security numbers, credit card numbers, passwords, or other personal information.