In fact, it does require ojdbc from Oracle which also requires a username and 
password. This was added as part of the testing scope for Oracle's docker.


I notice this PR and commit in branch-2.0 according to 
https://issues.apache.org/jira/browse/SPARK-12941.

In the comment, I'm not sure what does it mean by installing the JAR locally 
while Spark QA test run. IF this is the case,

it means someone downloaded the JAR from Oracle and manually added to the local 
build machine that is building Spark branch-2.0 or internal maven repository 
that will serve this ojdbc JAR.


========

commit 8afe49141d9b6a603eb3907f32dce802a3d05172

Author: thomastechs <thomas.sebast...@tcs.com>

Date:   Thu Feb 25 22:52:25 2016 -0800


    [SPARK-12941][SQL][MASTER] Spark-SQL JDBC Oracle dialect fails to map 
string datatypes to Oracle VARCHAR datatype



    ## What changes were proposed in this pull request?



    This Pull request is used for the fix SPARK-12941, creating a data type 
mapping to Oracle for the corresponding data type"Stringtype" from

dataframe. This PR is for the master branch fix, where as another PR is already 
tested with the branch 1.4



    ## How was the this patch tested?



    (Please explain how this patch was tested. E.g. unit tests, integration 
tests, manual tests)

    This patch was tested using the Oracle docker .Created a new integration 
suite for the same.The oracle.jdbc jar was to be downloaded from the maven 
repository.Since there was no jdbc jar available in the maven repository, the 
jar was downloaded from oracle site manually and installed in the local; thus 
tested. So, for SparkQA test case run, the ojdbc jar might be manually placed 
in the local maven repository(com/oracle/ojdbc6/11.2.0.2.0) while Spark QA test 
run.



    Author: thomastechs <thomas.sebast...@tcs.com>



    Closes #11306 from thomastechs/master.

========


Meanwhile, I also notice that the ojdbc groupID provided by Oracle (official 
website https://blogs.oracle.com/dev2dev/entry/how_to_get_oracle_jdbc)  is 
different.


    <dependency>

      <groupId>com.oracle.jdbc</groupId>

      <artifactId>ojdbc6</artifactId>

      <version>11.2.0.4</version>

      <scope>test</scope>

    </dependency>


as oppose to the one in Spark branch-2.0

external/docker-integration-tests/pom.xml


    <dependency>

      <groupId>com.oracle</groupId>

      <artifactId>ojdbc6</artifactId>

      <version>11.2.0.1.0</version>

      <scope>test</scope>

    </dependency>



The version is out of date and not available from the Oracle Maven repo. The PR 
was created awhile back, so the solution may just cross Oracle's maven release 
blog.


Just my inference based on what I see form git and JIRA, however, I do see a 
fix required to patch pom.xml to apply the correct groupId and version # for 
ojdbc6 driver.


Thoughts?



Get Oracle JDBC drivers and UCP from Oracle Maven 
...<https://blogs.oracle.com/dev2dev/entry/how_to_get_oracle_jdbc>
blogs.oracle.com
Get Oracle JDBC drivers and UCP from Oracle Maven Repository (without IDEs) By 
Nirmala Sundarappa-Oracle on Feb 15, 2016








________________________________
From: Mich Talebzadeh <mich.talebza...@gmail.com>
Sent: Tuesday, May 3, 2016 1:04 AM
To: Luciano Resende
Cc: Hien Luu; ☼ R Nair (रविशंकर नायर); user
Subject: Re: Spark build failure with com.oracle:ojdbc6:jar:11.2.0.1.0

which version of Spark are using?


Dr Mich Talebzadeh



LinkedIn  
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw



http://talebzadehmich.wordpress.com<http://talebzadehmich.wordpress.com/>



On 3 May 2016 at 02:13, Luciano Resende 
<luckbr1...@gmail.com<mailto:luckbr1...@gmail.com>> wrote:
You might have a settings.xml that is forcing your internal Maven repository to 
be the mirror of external repositories and thus not finding the dependency.

On Mon, May 2, 2016 at 6:11 PM, Hien Luu 
<hien...@gmail.com<mailto:hien...@gmail.com>> wrote:
Not I am not.  I am considering downloading it manually and place it in my 
local repository.

On Mon, May 2, 2016 at 5:54 PM, ☼ R Nair (रविशंकर नायर) 
<ravishankar.n...@gmail.com<mailto:ravishankar.n...@gmail.com>> wrote:

Oracle jdbc is not part of Maven repository,  are you keeping a downloaded file 
in your local repo?

Best, RS

On May 2, 2016 8:51 PM, "Hien Luu" 
<hien...@gmail.com<mailto:hien...@gmail.com>> wrote:
Hi all,

I am running into a build problem with com.oracle:ojdbc6:jar:11.2.0.1.0.  It 
kept getting "Operation timed out" while building Spark Project Docker 
Integration Tests module (see the error below).

Has anyone run this problem before? If so, how did you resolve around this 
problem?

[INFO] Reactor Summary:

[INFO]

[INFO] Spark Project Parent POM ........................... SUCCESS [  2.423 s]

[INFO] Spark Project Test Tags ............................ SUCCESS [  0.712 s]

[INFO] Spark Project Sketch ............................... SUCCESS [  0.498 s]

[INFO] Spark Project Networking ........................... SUCCESS [  1.743 s]

[INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [  0.587 s]

[INFO] Spark Project Unsafe ............................... SUCCESS [  0.503 s]

[INFO] Spark Project Launcher ............................. SUCCESS [  4.894 s]

[INFO] Spark Project Core ................................. SUCCESS [ 17.953 s]

[INFO] Spark Project GraphX ............................... SUCCESS [  3.480 s]

[INFO] Spark Project Streaming ............................ SUCCESS [  6.022 s]

[INFO] Spark Project Catalyst ............................. SUCCESS [  8.664 s]

[INFO] Spark Project SQL .................................. SUCCESS [ 12.440 s]

[INFO] Spark Project ML Local Library ..................... SUCCESS [  0.498 s]

[INFO] Spark Project ML Library ........................... SUCCESS [  8.594 s]

[INFO] Spark Project Tools ................................ SUCCESS [  0.162 s]

[INFO] Spark Project Hive ................................. SUCCESS [  9.834 s]

[INFO] Spark Project HiveContext Compatibility ............ SUCCESS [  1.428 s]

[INFO] Spark Project Docker Integration Tests ............. FAILURE [02:32 min]

[INFO] Spark Project REPL ................................. SKIPPED

[INFO] Spark Project Assembly ............................. SKIPPED

[INFO] Spark Project External Flume Sink .................. SKIPPED

[INFO] Spark Project External Flume ....................... SKIPPED

[INFO] Spark Project External Flume Assembly .............. SKIPPED

[INFO] Spark Project External Kafka ....................... SKIPPED

[INFO] Spark Project Examples ............................. SKIPPED

[INFO] Spark Project External Kafka Assembly .............. SKIPPED

[INFO] Spark Project Java 8 Tests ......................... SKIPPED

[INFO] ------------------------------------------------------------------------

[INFO] BUILD FAILURE

[INFO] ------------------------------------------------------------------------

[INFO] Total time: 03:53 min

[INFO] Finished at: 2016-05-02T17:44:57-07:00

[INFO] Final Memory: 80M/1525M

[INFO] ------------------------------------------------------------------------

[ERROR] Failed to execute goal on project spark-docker-integration-tests_2.11: 
Could not resolve dependencies for project 
org.apache.spark:spark-docker-integration-tests_2.11:jar:2.0.0-SNAPSHOT: Failed 
to collect dependencies at com.oracle:ojdbc6:jar:11.2.0.1.0: Failed to read 
artifact descriptor for com.oracle:ojdbc6:jar:11.2.0.1.0: Could not transfer 
artifact com.oracle:ojdbc6:pom:11.2.0.1.0 from/to uber-artifactory 
(http://artifactory.uber.internal:4587/artifactory/repo/): Connect to 
artifactory.uber.internal:4587 
[artifactory.uber.internal/10.162.11.61<http://10.162.11.61>] failed: Operation 
timed out -> [Help 1]




--
Regards,



--
Luciano Resende
http://twitter.com/lresende1975
http://lresende.blogspot.com/

Reply via email to