Actually it seems, when I have two requests, both jobs seems to crash
throwing the following errors:

phoenixdb.errors.InternalError: ("RuntimeException:
org.apache.phoenix.exception.PhoenixIOException:
org.apache.phoenix.exception.PhoenixIOException: Failed after attempts=36,
exceptions:\nThu Nov 23 12:02:31 UTC 2017, null,
java.net.SocketTimeoutException: callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
PhoenixIOException: org.apache.phoenix.exception.PhoenixIOException: Failed
after attempts=36, exceptions:\nThu Nov 23 12:02:31 UTC 2017, null,
java.net.SocketTimeoutException: callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
ExecutionException: org.apache.phoenix.exception.PhoenixIOException: Failed
after attempts=36, exceptions:\nThu Nov 23 12:02:31 UTC 2017, null,
java.net.SocketTimeoutException: callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
PhoenixIOException: Failed after attempts=36, exceptions:\nThu Nov 23
12:02:31 UTC 2017, null, java.net.SocketTimeoutException:
callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
PhoenixIOException: Failed after attempts=36, exceptions:\nThu Nov 23
12:02:31 UTC 2017, null, java.net.SocketTimeoutException:
callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
RetriesExhaustedException: Failed after attempts=36, exceptions:\nThu Nov
23 12:02:31 UTC 2017, null, java.net.SocketTimeoutException:
callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973\n ->
SocketTimeoutException: callTimeout=60000, callDuration=60108: row
'2017-07-01-BM8955-1498884512.0-386-1509295361.5' on table 'BGALLSALES' at
region=BGALLSALES,2017-06-24-BM8912-1498306109.0-58-1509294220.81,1510055260213.c2fbd3a07c560958f5944cda3c29c04f.,
hostname=bigdata-datanode,16020,1511427352525, seqNum=330973 ->
IOException: Call to bigdata-datanode/10.10.10.166:16020 failed on local
exception: org.apache.hadoop.hbase.ipc.CallTimeoutException: Call id=48582,
waitTime=60004, operationTimeout=60000 expired. -> CallTimeoutException:
Call id=48582, waitTime=60004, operationTimeout=60000 expired.", None,
None, None)


I've changed the hbase configuration like this:

      <property>
      <name>phoenix.query.timeoutMs</name>
      <value>1800000</value>
      </property>
      <property>
      <name>hbase.regionserver.lease.period</name>
      <value>1200000</value>
      </property>

      <property>
      <name>hbase.regionserver.lease.period</name>
      <value>1200000</value>
      </property>
      <property>
      <name>hbase.rpc.timeout</name>
      <value>1200000</value>
      </property>
      <property>
      <name>hbase.client.scanner.caching</name>
      <value>1000</value>
      </property>
      <property>
      <name>hbase.client.scanner.timeout.period</name>
      <value>1200000</value>
      </property>

but with no help. Any suggessation would be great, Mostly I use python
client, but this happens when I use sqlline as well.

Thanks


On Thu, Nov 23, 2017 at 10:30 AM, Vaghawan Ojha <vaghawan...@gmail.com>
wrote:

> Hi,
>
> Thank you very much for reaching out,
>
> - Phoenix version: 4.12.0-HBase-1.2
> - Hbase: 1.2.6
> - Hadoop: 2.8.1
>
>
> DEBUG from PQS:
>
> 2017-11-23 04:27:46,157 WARN org.apache.hadoop.hbase.client.ScannerCallable:
> Ignore, probably already closed
> org.apache.hadoop.hbase.UnknownScannerException: 
> org.apache.hadoop.hbase.UnknownScannerException:
> Unknown scanner '14111'. This can happen due$
> If the issue is due to reason (b), a possible fix would be increasing the
> value of'hbase.client.scanner.timeout.period' configuration.
>         at org.apache.hadoop.hbase.regionserver.RSRpcServices.
> scan(RSRpcServices.java:2394)
>         at org.apache.hadoop.hbase.protobuf.generated.
> ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
>         at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2196)
>         at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
>         at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(
> RpcExecutor.java:133)
>         at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.
> java:108)
>         at java.lang.Thread.run(Thread.java:748)
>
> The exception thrown in the python client is attached.
>
> The two query sent was something like this:
>
> (SELECT COL1, COL2, COL3, COL4, COL5 FROM SALESTABLE WHERE
> PURCHSHED_DATE>=TO_TIMESTAMP(d) AND PURCHSHED_DATE<=TO_TIMESTAMP(d))
> Having real data of 1 million records.
> (SELECT * FROM USERS) Having dummy data of 10 records.
>
> I've one namenode and one datanode.
>
> Thank you for the help.
>
> On Thu, Nov 23, 2017 at 4:42 AM, Josh Elser <els...@apache.org> wrote:
>
>> There is no such configuration which would preclude your ability to issue
>> two queries concurrently.
>>
>> Some relevant information you should share if you'd like more help:
>>
>> * Versions of HDFS, HBase, and Phoenix
>> * What your "request" is
>> * Thread-dump (e.g. jstack) of your client and the PQS
>> * DEBUG logs from PQS
>>
>>
>> On 11/21/17 12:30 AM, Vaghawan Ojha wrote:
>>
>>> Hi,
>>>
>>> I've a one Namenode server and another Datanode server, phoenix query
>>> server are running on both.
>>>
>>> For a single request, everything is fine, but when I do multiple
>>> requests simultaneously, the query server timeouts , the table consists
>>> only 2-3 gbs of data.
>>>
>>> I think, I'm missing some configuration to manage this, if somebody
>>> could help me out, it would be great.
>>>
>>> Thank you
>>> Vaghawan
>>>
>>
>

Reply via email to