>From cassandra.yaml :

native_transport_port: 9042

FYI

On Tue, Mar 8, 2016 at 9:13 PM, Saurabh Bajaj <bajaj.onl...@gmail.com>
wrote:

> Hi Andy,
>
> I believe you need to set the host and port settings separately
> spark.cassandra.connection.host
> spark.cassandra.connection.port
>
> https://github.com/datastax/spark-cassandra-connector/blob/master/doc/reference.md#cassandra-connection-parameters
>
> Looking at the logs, it seems your port config is not being set and it's
> falling back to default.
> Let me know if that helps.
>
> Saurabh Bajaj
>
> On Tue, Mar 8, 2016 at 6:25 PM, Andy Davidson <
> a...@santacruzintegration.com> wrote:
>
>> Hi Ted
>>
>> I believe by default cassandra listens on 9042
>>
>> From: Ted Yu <yuzhih...@gmail.com>
>> Date: Tuesday, March 8, 2016 at 6:11 PM
>> To: Andrew Davidson <a...@santacruzintegration.com>
>> Cc: "user @spark" <user@spark.apache.org>
>> Subject: Re: pyspark spark-cassandra-connector java.io.IOException:
>> Failed to open native connection to Cassandra at {192.168.1.126}:9042
>>
>> Have you contacted spark-cassandra-connector related mailing list ?
>>
>> I wonder where the port 9042 came from.
>>
>> Cheers
>>
>> On Tue, Mar 8, 2016 at 6:02 PM, Andy Davidson <
>> a...@santacruzintegration.com> wrote:
>>
>>>
>>> I am using spark-1.6.0-bin-hadoop2.6. I am trying to write a python
>>> notebook that reads a data frame from Cassandra.
>>>
>>> *I connect to cassadra using an ssh tunnel running on port 9043.* CQLSH
>>> works how ever I can not figure out how to configure my notebook. I have
>>> tried various hacks any idea what I am doing wrong
>>>
>>> : java.io.IOException: Failed to open native connection to Cassandra at 
>>> {192.168.1.126}:9042
>>>
>>>
>>>
>>>
>>> Thanks in advance
>>>
>>> Andy
>>>
>>>
>>>
>>> $ extraPkgs="--packages com.databricks:spark-csv_2.11:1.3.0 \
>>>             --packages
>>> datastax:spark-cassandra-connector:1.6.0-M1-s_2.11"
>>>
>>> $ export PYSPARK_PYTHON=python3
>>> $ export PYSPARK_DRIVER_PYTHON=python3
>>> $ IPYTHON_OPTS=notebook $SPARK_ROOT/bin/pyspark $extraPkgs $*
>>>
>>>
>>>
>>> In [15]:
>>> 1
>>>
>>> sqlContext.setConf("spark.cassandra.connection.host”,”127.0.0.1:9043")
>>>
>>> 2
>>>
>>> df = sqlContext.read\
>>>
>>> 3
>>>
>>>     .format("org.apache.spark.sql.cassandra")\
>>>
>>> 4
>>>
>>>     .options(table=“time_series", keyspace="notification")\
>>>
>>> 5
>>>
>>>     .load()
>>>
>>> 6
>>>
>>> ​
>>>
>>> 7
>>>
>>> df.printSchema()
>>>
>>> 8
>>>
>>> df.show()
>>>
>>> ---------------------------------------------------------------------------Py4JJavaError
>>>                              Traceback (most recent call 
>>> last)<ipython-input-15-9d8f6dcf210f> in <module>()      1 
>>> sqlContext.setConf("spark.cassandra.connection.host","localhost:9043")----> 
>>> 2 df = sqlContext.read    .format("org.apache.spark.sql.cassandra")    
>>> .options(table="kv", keyspace="notification")    .load()      3       4 
>>> df.printSchema()      5 
>>> df.show()/Users/andrewdavidson/workSpace/spark/spark-1.6.0-bin-hadoop2.6/python/pyspark/sql/readwriter.py
>>>  in load(self, path, format, schema, **options)    137                 
>>> return self._df(self._jreader.load(path))    138         else:--> 139       
>>>       return self._df(self._jreader.load())    140     141     
>>> @since(1.4)/Users/andrewdavidson/workSpace/spark/spark-1.6.0-bin-hadoop2.6/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py
>>>  in __call__(self, *args)    811         answer = 
>>> self.gateway_client.send_command(command)    812         return_value = 
>>> get_return_value(--> 813             answer, self.gateway_client, 
>>> self.target_id, self.name)    814     815         for temp_arg in 
>>> temp_args:/Users/andrewdavidson/workSpace/spark/spark-1.6.0-bin-hadoop2.6/python/pyspark/sql/utils.py
>>>  in deco(*a, **kw)     43     def deco(*a, **kw):     44         try:---> 
>>> 45             return f(*a, **kw)     46         except 
>>> py4j.protocol.Py4JJavaError as e:     47             s = 
>>> e.java_exception.toString()/Users/andrewdavidson/workSpace/spark/spark-1.6.0-bin-hadoop2.6/python/lib/py4j-0.9-src.zip/py4j/protocol.py
>>>  in get_return_value(answer, gateway_client, target_id, name)    306        
>>>          raise Py4JJavaError(    307                     "An error occurred 
>>> while calling {0}{1}{2}.\n".--> 308                     format(target_id, 
>>> ".", name), value)
>>>     309             else:    310                 raise Py4JError(
>>> Py4JJavaError: An error occurred while calling o280.load.
>>> : java.io.IOException: Failed to open native connection to Cassandra at 
>>> {192.168.1.126}:9042
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector$.com$datastax$spark$connector$cql$CassandraConnector$$createSession(CassandraConnector.scala:162)
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$2.apply(CassandraConnector.scala:148)
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector$$anonfun$2.apply(CassandraConnector.scala:148)
>>>     at 
>>> com.datastax.spark.connector.cql.RefCountedCache.createNewValueAndKeys(RefCountedCache.scala:31)
>>>     at 
>>> com.datastax.spark.connector.cql.RefCountedCache.acquire(RefCountedCache.scala:56)
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector.openSession(CassandraConnector.scala:81)
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector.withSessionDo(CassandraConnector.scala:109)
>>>     at 
>>> com.datastax.spark.connector.rdd.partitioner.CassandraRDDPartitioner$.getTokenFactory(CassandraRDDPartitioner.scala:184)
>>>     at 
>>> org.apache.spark.sql.cassandra.CassandraSourceRelation$.apply(CassandraSourceRelation.scala:267)
>>>     at 
>>> org.apache.spark.sql.cassandra.DefaultSource.createRelation(DefaultSource.scala:57)
>>>     at 
>>> org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:158)
>>>     at org.apache.spark.sql.DataFrameReader.load(DataFrameReader.scala:119)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at 
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>     at 
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>     at java.lang.reflect.Method.invoke(Method.java:497)
>>>     at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
>>>     at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
>>>     at py4j.Gateway.invoke(Gateway.java:259)
>>>     at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
>>>     at py4j.commands.CallCommand.execute(CallCommand.java:79)
>>>     at py4j.GatewayConnection.run(GatewayConnection.java:209)
>>>     at java.lang.Thread.run(Thread.java:745)
>>> Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: 
>>> All host(s) tried for query failed (tried: /192.168.1.126:9042 
>>> (com.datastax.driver.core.exceptions.TransportException: [/192.168.1.126] 
>>> Cannot connect))
>>>     at 
>>> com.datastax.driver.core.ControlConnection.reconnectInternal(ControlConnection.java:231)
>>>     at 
>>> com.datastax.driver.core.ControlConnection.connect(ControlConnection.java:77)
>>>     at com.datastax.driver.core.Cluster$Manager.init(Cluster.java:1414)
>>>     at com.datastax.driver.core.Cluster.getMetadata(Cluster.java:393)
>>>     at 
>>> com.datastax.spark.connector.cql.CassandraConnector$.com$datastax$spark$connector$cql$CassandraConnector$$createSession(CassandraConnector.scala:155)
>>>     ... 22 more
>>>
>>>
>>>
>>
>

Reply via email to