The manual download is not required on latest trunk code anymore. On Monday, May 9, 2016, Andrew Lee <alee...@hotmail.com> wrote:
> In fact, it does require ojdbc from Oracle which also requires a username > and password. This was added as part of the testing scope for > Oracle's docker. > > > I notice this PR and commit in branch-2.0 according to > https://issues.apache.org/jira/browse/SPARK-12941. > > In the comment, I'm not sure what does it mean by installing the JAR > locally while Spark QA test run. IF this is the case, > > it means someone downloaded the JAR from Oracle and manually added to the > local build machine that is building Spark branch-2.0 or internal maven > repository that will serve this ojdbc JAR. > > > ======== > > commit 8afe49141d9b6a603eb3907f32dce802a3d05172 > > Author: thomastechs <thomas.sebast...@tcs.com > <javascript:_e(%7B%7D,'cvml','thomas.sebast...@tcs.com');>> > > Date: Thu Feb 25 22:52:25 2016 -0800 > > > [SPARK-12941][SQL][MASTER] Spark-SQL JDBC Oracle dialect fails to map > string datatypes to Oracle VARCHAR datatype > > > > ## What changes were proposed in this pull request? > > > > This Pull request is used for the fix SPARK-12941, creating a data > type mapping to Oracle for the corresponding data type"Stringtype" from > > dataframe. This PR is for the master branch fix, where as another PR is > already tested with the branch 1.4 > > > > ## How was the this patch tested? > > > > (Please explain how this patch was tested. E.g. unit tests, > integration tests, manual tests) > > This patch was tested using the Oracle docker .Created a new > integration suite for the same.The oracle.jdbc jar was to be downloaded > from the maven repository.Since there was no jdbc jar available in the > maven repository, the jar was downloaded from oracle site manually and > installed in the local; thus tested. So, for SparkQA test case run, the > ojdbc jar might be manually placed in the local maven > repository(com/oracle/ojdbc6/11.2.0.2.0) while Spark QA test run. > > > > Author: thomastechs <thomas.sebast...@tcs.com > <javascript:_e(%7B%7D,'cvml','thomas.sebast...@tcs.com');>> > > > > Closes #11306 from thomastechs/master. > ======== > > > Meanwhile, I also notice that the ojdbc groupID provided by Oracle > (official website > https://blogs.oracle.com/dev2dev/entry/how_to_get_oracle_jdbc) is > different. > > > <dependency> > > <groupId>com.oracle.jdbc</groupId> > > <artifactId>ojdbc6</artifactId> > > <version>11.2.0.4</version> > > <scope>test</scope> > > </dependency> > > as oppose to the one in Spark branch-2.0 > > external/docker-integration-tests/pom.xml > > > <dependency> > > <groupId>com.oracle</groupId> > > <artifactId>ojdbc6</artifactId> > > <version>11.2.0.1.0</version> > > <scope>test</scope> > > </dependency> > > > The version is out of date and not available from the Oracle Maven repo. > The PR was created awhile back, so the solution may just cross Oracle's > maven release blog. > > > Just my inference based on what I see form git and JIRA, however, I do see > a fix required to patch pom.xml to apply the correct groupId and version # > for ojdbc6 driver. > > > Thoughts? > > > > Get Oracle JDBC drivers and UCP from Oracle Maven ... > <https://blogs.oracle.com/dev2dev/entry/how_to_get_oracle_jdbc> > blogs.oracle.com > Get Oracle JDBC drivers and UCP from Oracle Maven Repository (without > IDEs) By Nirmala Sundarappa-Oracle on Feb 15, 2016 > > > > > > > > ------------------------------ > *From:* Mich Talebzadeh <mich.talebza...@gmail.com > <javascript:_e(%7B%7D,'cvml','mich.talebza...@gmail.com');>> > *Sent:* Tuesday, May 3, 2016 1:04 AM > *To:* Luciano Resende > *Cc:* Hien Luu; ☼ R Nair (रविशंकर नायर); user > *Subject:* Re: Spark build failure with com.oracle:ojdbc6:jar:11.2.0.1.0 > > which version of Spark are using? > > Dr Mich Talebzadeh > > > > LinkedIn * > https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw > <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* > > > > http://talebzadehmich.wordpress.com > > > > On 3 May 2016 at 02:13, Luciano Resende <luckbr1...@gmail.com > <javascript:_e(%7B%7D,'cvml','luckbr1...@gmail.com');>> wrote: > >> You might have a settings.xml that is forcing your internal Maven >> repository to be the mirror of external repositories and thus not finding >> the dependency. >> >> On Mon, May 2, 2016 at 6:11 PM, Hien Luu <hien...@gmail.com >> <javascript:_e(%7B%7D,'cvml','hien...@gmail.com');>> wrote: >> >>> Not I am not. I am considering downloading it manually and place it in >>> my local repository. >>> >>> On Mon, May 2, 2016 at 5:54 PM, ☼ R Nair (रविशंकर नायर) < >>> ravishankar.n...@gmail.com >>> <javascript:_e(%7B%7D,'cvml','ravishankar.n...@gmail.com');>> wrote: >>> >>>> Oracle jdbc is not part of Maven repository, are you keeping a >>>> downloaded file in your local repo? >>>> >>>> Best, RS >>>> On May 2, 2016 8:51 PM, "Hien Luu" <hien...@gmail.com >>>> <javascript:_e(%7B%7D,'cvml','hien...@gmail.com');>> wrote: >>>> >>>>> Hi all, >>>>> >>>>> I am running into a build problem with >>>>> com.oracle:ojdbc6:jar:11.2.0.1.0. It kept getting "Operation timed out" >>>>> while building Spark Project Docker Integration Tests module (see the >>>>> error >>>>> below). >>>>> >>>>> Has anyone run this problem before? If so, how did you resolve around >>>>> this problem? >>>>> >>>>> [INFO] Reactor Summary: >>>>> >>>>> [INFO] >>>>> >>>>> [INFO] Spark Project Parent POM ........................... SUCCESS [ >>>>> 2.423 s] >>>>> >>>>> [INFO] Spark Project Test Tags ............................ SUCCESS [ >>>>> 0.712 s] >>>>> >>>>> [INFO] Spark Project Sketch ............................... SUCCESS [ >>>>> 0.498 s] >>>>> >>>>> [INFO] Spark Project Networking ........................... SUCCESS [ >>>>> 1.743 s] >>>>> >>>>> [INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [ >>>>> 0.587 s] >>>>> >>>>> [INFO] Spark Project Unsafe ............................... SUCCESS [ >>>>> 0.503 s] >>>>> >>>>> [INFO] Spark Project Launcher ............................. SUCCESS [ >>>>> 4.894 s] >>>>> >>>>> [INFO] Spark Project Core ................................. SUCCESS [ >>>>> 17.953 s] >>>>> >>>>> [INFO] Spark Project GraphX ............................... SUCCESS [ >>>>> 3.480 s] >>>>> >>>>> [INFO] Spark Project Streaming ............................ SUCCESS [ >>>>> 6.022 s] >>>>> >>>>> [INFO] Spark Project Catalyst ............................. SUCCESS [ >>>>> 8.664 s] >>>>> >>>>> [INFO] Spark Project SQL .................................. SUCCESS [ >>>>> 12.440 s] >>>>> >>>>> [INFO] Spark Project ML Local Library ..................... SUCCESS [ >>>>> 0.498 s] >>>>> >>>>> [INFO] Spark Project ML Library ........................... SUCCESS [ >>>>> 8.594 s] >>>>> >>>>> [INFO] Spark Project Tools ................................ SUCCESS [ >>>>> 0.162 s] >>>>> >>>>> [INFO] Spark Project Hive ................................. SUCCESS [ >>>>> 9.834 s] >>>>> >>>>> [INFO] Spark Project HiveContext Compatibility ............ SUCCESS [ >>>>> 1.428 s] >>>>> >>>>> [INFO] Spark Project Docker Integration Tests ............. FAILURE >>>>> [02:32 min] >>>>> >>>>> [INFO] Spark Project REPL ................................. SKIPPED >>>>> >>>>> [INFO] Spark Project Assembly ............................. SKIPPED >>>>> >>>>> [INFO] Spark Project External Flume Sink .................. SKIPPED >>>>> >>>>> [INFO] Spark Project External Flume ....................... SKIPPED >>>>> >>>>> [INFO] Spark Project External Flume Assembly .............. SKIPPED >>>>> >>>>> [INFO] Spark Project External Kafka ....................... SKIPPED >>>>> >>>>> [INFO] Spark Project Examples ............................. SKIPPED >>>>> >>>>> [INFO] Spark Project External Kafka Assembly .............. SKIPPED >>>>> >>>>> [INFO] Spark Project Java 8 Tests ......................... SKIPPED >>>>> >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> >>>>> [INFO] BUILD FAILURE >>>>> >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> >>>>> [INFO] Total time: 03:53 min >>>>> >>>>> [INFO] Finished at: 2016-05-02T17:44:57-07:00 >>>>> >>>>> [INFO] Final Memory: 80M/1525M >>>>> >>>>> [INFO] >>>>> ------------------------------------------------------------------------ >>>>> >>>>> [ERROR] Failed to execute goal on project >>>>> spark-docker-integration-tests_2.11: Could not resolve dependencies for >>>>> project >>>>> org.apache.spark:spark-docker-integration-tests_2.11:jar:2.0.0-SNAPSHOT: >>>>> Failed to collect dependencies at com.oracle:ojdbc6:jar:11.2.0.1.0: Failed >>>>> to read artifact descriptor for com.oracle:ojdbc6:jar:11.2.0.1.0: Could >>>>> not >>>>> transfer artifact com.oracle:ojdbc6:pom:11.2.0.1.0 from/to >>>>> uber-artifactory >>>>> (http://artifactory.uber.internal:4587/artifactory/repo/): Connect to >>>>> artifactory.uber.internal:4587 [artifactory.uber.internal/10.162.11.61] >>>>> failed: Operation timed out -> [Help 1] >>>>> >>>>> >>> >>> >>> -- >>> Regards, >>> >> >> >> >> -- >> Luciano Resende >> http://twitter.com/lresende1975 >> http://lresende.blogspot.com/ >> > > -- Sent from my Mobile device