[jira] [Updated] (TRAFODION-2664) Instance will be down when the zookeeper on name node has been down

2017-06-21 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek updated TRAFODION-2664:
-
Description: 
Description: Instance will be down when the zookeeper on name node has been down
Test Steps:
Step 1. Start OE and 4 long queries with trafci on the first node esggy-clu-n010
Step 2. Wait several minutes and stop zookeeper on name node of node 
esggy-clu-n010  in Cloudera Manager page.
Step 3. With trafci, run a basic query and 4 long queries again.

In the above Step 3, we will see the whole instance as down after a while. For 
this test scenario, I tried it several times, always found instance as down.

Timestamp:
Test Start Time: 20170616132939
Test End  Time: 20170616134350
Stop zookeeper on name node of node esggy-clu-n010: 20170616133344

Check logs:
1) Each node displays the following error:
2017-06-16 13:33:46,276, ERROR, MON, Node Number: 0,, PIN: 5017 , Process Name: 
$MONITOR,,, TID: 5429, Message ID: 101371801, [CZClient::IsZNodeExpired], 
zoo_exists() for /trafodion/instance/cluster/esggy-clu-n010.esgyn.cn failed 
with error ZCONNECTIONLOSS
2) Zookeeper displays:
ls /trafodion/instance/cluster
[]
So, It seems zclient has been lost on each node.

Location of logs:
esggy-clu-n010: 
/data4/jarek/ha.interactive/trafodion_and_cluster_logs/cluster_logs.20170616134816.tar.gz
 and trafodion_logs.20170616134816.tar.gz
By the way, because the size of the logs is out of the limited value, so i 
cannot upload it as the attachment in this JIRA ID.

How many zookeeper quorum servers in the cluster? total 3.
  
dcs.zookeeper.quorum

esggy-clu-n010.esgyn.cn,esggy-clu-n011.esgyn.cn,esggy-clu-n012.esgyn.cn
  

How to access the cluster?
1. Login 10.10.10.8 from US machine: trafodion/traf123
2. Login 10.10.23.19 from 10.10.10.8: trafodion/traf123




  was:
Description: Instance will be down when the zookeeper on name node has been down
Test Steps:
Step 1. Start OE and 4 long queries with trafci on the first node esggy-clu-n010
Step 2. Wait several minutes and stop zookeeper on name node of node 
esggy-clu-n010  in Cloudera Manager page.
Step 3. With trafci, run a basic query and 4 long queries again.

In the above Step 3, we will see the whole instance as down after a while. For 
this test scenario, I tried it several times, always found instance as down.

Timestamp:
Test Start Time: 20170616132939
Test End  Time: 20170616134350
Stop zookeeper on name node of node esggy-clu-n010: 20170616133344

Check logs:
1) Each node displays the following error:
2017-06-16 13:33:46,276, ERROR, MON, Node Number: 0,, PIN: 5017 , Process Name: 
$MONITOR,,, TID: 5429, Message ID: 101371801, [CZClient::IsZNodeExpired], 
zoo_exists() for /trafodion/instance/cluster/esggy-clu-n010.esgyn.cn failed 
with error ZCONNECTIONLOSS
2) Zookeeper displays:
ls /trafodion/instance/cluster
[]
So, It seems zclient has been lost on each node.

Location of logs:
esggy-clu-n010: 
/data4/jarek/ha.interactive/trafodion_and_cluster_logs/cluster_logs.20170616134816.tar.gz
 and trafodion_logs.20170616134816.tar.gz
By the way, because the size of the logs is out of the limited size, so i 
cannot upload it as the attachment in this JIRA ID.

How many zookeeper quorum servers in the cluster? total 3.
  
dcs.zookeeper.quorum

esggy-clu-n010.esgyn.cn,esggy-clu-n011.esgyn.cn,esggy-clu-n012.esgyn.cn
  

How to access the cluster?
1. Login 10.10.10.8 from US machine: trafodion/traf123
2. Login 10.10.23.19 from 10.10.10.8: trafodion/traf123





> Instance will be down when the zookeeper on name node has been down
> ---
>
> Key: TRAFODION-2664
> URL: https://issues.apache.org/jira/browse/TRAFODION-2664
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: foundation
>Affects Versions: 2.2-incubating
> Environment: Test Environment:
> CDH5.4.8: 10.10.23.19:7180, total 6 nodes.
> HDFS-HA and DCS-HA: enabled
> OS: Centos6.8, physic machine.
> SW Build: R2.2.3 (EsgynDB_Enterprise Release 2.2.3 (Build release [sbroeder], 
> branch 1ce8d39-xdc_nari, date 11Jun17)
>Reporter: Jarek
>Priority: Critical
>  Labels: build
> Fix For: 2.2-incubating
>
>
> Description: Instance will be down when the zookeeper on name node has been 
> down
> Test Steps:
> Step 1. Start OE and 4 long queries with trafci on the first node 
> esggy-clu-n010
> Step 2. Wait several minutes and stop zookeeper on name node of node 
> esggy-clu-n010  in Cloudera Manager page.
> Step 3. With trafci, run a basic query and 4 long queries again.
> In the above Step 3, we will see the whole instance as down after a while. 
> For this test scenario, I tried it several times, always found instance as 
> down.
> Timestamp:
> Test Start Time: 20170616132939
> Test End  Time: 

[jira] [Updated] (TRAFODION-2664) Instance will be down when the zookeeper on name node has been down

2017-06-21 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2664?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek updated TRAFODION-2664:
-
Description: 
Description: Instance will be down when the zookeeper on name node has been down
Test Steps:
Step 1. Start OE and 4 long queries with trafci on the first node esggy-clu-n010
Step 2. Wait several minutes and stop zookeeper on name node of node 
esggy-clu-n010  in Cloudera Manager page.
Step 3. With trafci, run a basic query and 4 long queries again.

In the above Step 3, we will see the whole instance as down after a while. For 
this test scenario, I tried it several times, always found instance as down.

Timestamp:
Test Start Time: 20170616132939
Test End  Time: 20170616134350
Stop zookeeper on name node of node esggy-clu-n010: 20170616133344

Check logs:
1) Each node displays the following error:
2017-06-16 13:33:46,276, ERROR, MON, Node Number: 0,, PIN: 5017 , Process Name: 
$MONITOR,,, TID: 5429, Message ID: 101371801, [CZClient::IsZNodeExpired], 
zoo_exists() for /trafodion/instance/cluster/esggy-clu-n010.esgyn.cn failed 
with error ZCONNECTIONLOSS
2) Zookeeper displays:
ls /trafodion/instance/cluster
[]
So, It seems zclient has been lost on each node.

Location of logs:
esggy-clu-n010: 
/data4/jarek/ha.interactive/trafodion_and_cluster_logs/cluster_logs.20170616134816.tar.gz
 and trafodion_logs.20170616134816.tar.gz
By the way, because the size of the logs is out of the limited size, so i 
cannot upload it as the attachment in this JIRA ID.

How many zookeeper quorum servers in the cluster? total 3.
  
dcs.zookeeper.quorum

esggy-clu-n010.esgyn.cn,esggy-clu-n011.esgyn.cn,esggy-clu-n012.esgyn.cn
  

How to access the cluster?
1. Login 10.10.10.8 from US machine: trafodion/traf123
2. Login 10.10.23.19 from 10.10.10.8: trafodion/traf123




  was:
Description: Instance will be down when the zookeeper on name node has been down
Test Steps:
Step 1. Start OE and 4 long queries with trafci on the first node esggy-clu-n010
Step 2. Wait several minutes and stop zookeeper on name node of node 
esggy-clu-n010  in Cloudera Manager page.
Step 3. With trafci, run a basic query and 4 long queries again.

In the above Step 3, we will see the whole instance as down after a while. For 
this test scenario, I tried it several times, always found instance as down.

Timestamp:
Test Start Time: 20170616132939
Test End  Time: 20170616134350
Stop zookeeper on name node of node esggy-clu-n010: 20170616133344

Check logs:
1) Each node displays the following error:
2017-06-16 13:33:46,276, ERROR, MON, Node Number: 0,, PIN: 5017 , Process Name: 
$MONITOR,,, TID: 5429, Message ID: 101371801, [CZClient::IsZNodeExpired], 
zoo_exists() for /trafodion/instance/cluster/esggy-clu-n010.esgyn.cn failed 
with error ZCONNECTIONLOSS
2) Zookeeper displays:
ls /trafodion/instance/cluster
[]
So, It seems zclient has been lost on each node.

Location of logs:
esggy-clu-n010: 
/data4/jarek/ha.interactive/trafodion_and_cluster_logs/cluster_logs.20170616134816.tar.gz
 and trafodion_logs.20170616134816.tar.gz
By the way, because the size of the logs is out of the limited size, so i 
cannot upload it as the attachment in this JIRA ID.

How to access the cluster?
1. Login 10.10.10.8 from US machine: trafodion/traf123
2. Login 10.10.23.19 from 10.10.10.8: trafodion/traf123





> Instance will be down when the zookeeper on name node has been down
> ---
>
> Key: TRAFODION-2664
> URL: https://issues.apache.org/jira/browse/TRAFODION-2664
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: foundation
>Affects Versions: 2.2-incubating
> Environment: Test Environment:
> CDH5.4.8: 10.10.23.19:7180, total 6 nodes.
> HDFS-HA and DCS-HA: enabled
> OS: Centos6.8, physic machine.
> SW Build: R2.2.3 (EsgynDB_Enterprise Release 2.2.3 (Build release [sbroeder], 
> branch 1ce8d39-xdc_nari, date 11Jun17)
>Reporter: Jarek
>Priority: Critical
>  Labels: build
> Fix For: 2.2-incubating
>
>
> Description: Instance will be down when the zookeeper on name node has been 
> down
> Test Steps:
> Step 1. Start OE and 4 long queries with trafci on the first node 
> esggy-clu-n010
> Step 2. Wait several minutes and stop zookeeper on name node of node 
> esggy-clu-n010  in Cloudera Manager page.
> Step 3. With trafci, run a basic query and 4 long queries again.
> In the above Step 3, we will see the whole instance as down after a while. 
> For this test scenario, I tried it several times, always found instance as 
> down.
> Timestamp:
> Test Start Time: 20170616132939
> Test End  Time: 20170616134350
> Stop zookeeper on name node of node esggy-clu-n010: 20170616133344
> Check logs:
> 1) Each node displays the following error:
> 2017-06-16 13:33:46,276, ERROR, 

[jira] [Created] (TRAFODION-2664) Instance will be down when the zookeeper on name node has been down

2017-06-21 Thread Jarek (JIRA)
Jarek created TRAFODION-2664:


 Summary: Instance will be down when the zookeeper on name node has 
been down
 Key: TRAFODION-2664
 URL: https://issues.apache.org/jira/browse/TRAFODION-2664
 Project: Apache Trafodion
  Issue Type: Bug
  Components: foundation
Affects Versions: 2.2-incubating
 Environment: Test Environment:
CDH5.4.8: 10.10.23.19:7180, total 6 nodes.
HDFS-HA and DCS-HA: enabled
OS: Centos6.8, physic machine.
SW Build: R2.2.3 (EsgynDB_Enterprise Release 2.2.3 (Build release [sbroeder], 
branch 1ce8d39-xdc_nari, date 11Jun17)
Reporter: Jarek
Priority: Critical
 Fix For: 2.2-incubating


Description: Instance will be down when the zookeeper on name node has been down
Test Steps:
Step 1. Start OE and 4 long queries with trafci on the first node esggy-clu-n010
Step 2. Wait several minutes and stop zookeeper on name node of node 
esggy-clu-n010  in Cloudera Manager page.
Step 3. With trafci, run a basic query and 4 long queries again.

In the above Step 3, we will see the whole instance as down after a while. For 
this test scenario, I tried it several times, always found instance as down.

Timestamp:
Test Start Time: 20170616132939
Test End  Time: 20170616134350
Stop zookeeper on name node of node esggy-clu-n010: 20170616133344

Check logs:
1) Each node displays the following error:
2017-06-16 13:33:46,276, ERROR, MON, Node Number: 0,, PIN: 5017 , Process Name: 
$MONITOR,,, TID: 5429, Message ID: 101371801, [CZClient::IsZNodeExpired], 
zoo_exists() for /trafodion/instance/cluster/esggy-clu-n010.esgyn.cn failed 
with error ZCONNECTIONLOSS
2) Zookeeper displays:
ls /trafodion/instance/cluster
[]
So, It seems zclient has been lost on each node.

Location of logs:
esggy-clu-n010: 
/data4/jarek/ha.interactive/trafodion_and_cluster_logs/cluster_logs.20170616134816.tar.gz
 and trafodion_logs.20170616134816.tar.gz
By the way, because the size of the logs is out of the limited size, so i 
cannot upload it as the attachment in this JIRA ID.

How to access the cluster?
1. Login 10.10.10.8 from US machine: trafodion/traf123
2. Login 10.10.23.19 from 10.10.10.8: trafodion/traf123






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)


[jira] [Created] (TRAFODION-2070) Trafodion cannot adjust your working status in time when network broken.

2016-06-17 Thread Jarek (JIRA)
Jarek created TRAFODION-2070:


 Summary: Trafodion cannot adjust your working status in time when 
network broken.
 Key: TRAFODION-2070
 URL: https://issues.apache.org/jira/browse/TRAFODION-2070
 Project: Apache Trafodion
  Issue Type: Bug
  Components: dtm
Affects Versions: 2.0-incubating, 2.1-incubating
Reporter: Jarek


Issue Title: Trafodion cannot adjust your working status in time when network 
broken.

Test Steps (including part 1 and part 2):
Preconditoin: the testing environment is good, including HDFS, HBase and EsgnDB.
Part 1: Network broken occurred for a long time, here limit it as 15 minutes.
Step 0. We login a traci interface to run a SQL statement ‘STMT_A’, like 
insert/delete/update/select, at the same time, the sql statement has running 
for several minutes.
Step 1. Use command ‘iptables -I INPUT -s $NODE_HOST -j DROP’ to make nap104 
node’s network unreachable for 15 minutes.
Step 2. Start to do check Step 0, major SQL command and HDFS/HBase running 
status.
 Here, check Step 0 and major SQL command should be on nap101, nap102 and 
nap103 nodes.
T-1 : 
Step 0  Comments
Expect  1.  When TRAFCI is connected to nap104 node, the SQL statement 
‘STMT_A’ run failed and exit TRAFCI normally.
2.  When TRAFCI is connected to nap101/nap102/nap103 node, the SQL 
statement ‘STMT_A’ run success and exit TRAFCI normally.
Actual  For expect 1: 
ISSUE 1: By checking the QID status of the SQL statement ‘STMT_A’, it normally 
displays “SQL>get statistics for qid
MXID110030258942123314457118951450206U300_339_SQL_CUR_7;

*** ERROR[2024] Server Process $ZSM003 is not running or could not be created. 
Operating System Error 14 was returned. [2016-05-31 09:22:48]”, back to TRAFCI 
interface, the SQL statement ‘STMT_A’ cannot be return in time but hang for a 
long time.

ISSUE 2: At the same time, open a new TRAFCI that is connected to other node 
for example nap102, do SQL query statement ‘STMT_B’, using command ‘./offender 
-s active’ to check the QID status of the SQL statement ‘STMT_B’,  but print 
the following error and say 0 row(s) selected 
“*** ERROR[8921] The request to obtain runtime statistics for ACTIVE_QUERIES=30 
timed out. Timeout period specified is 4 seconds.

--- 0 row(s) selected.” 
ISSUE 3: back to the TRAFCI session of ‘STMT_A’ and ‘STMT_B’, we can see these 
TRAFCI sessions are interrupted because of the below error.
“SQL>select [last 1] * from TRAFODION.JAVABENCH.YCSB_TABLE_20;

*** ERROR[29157] There was a problem reading from the server
*** ERROR[29160] The message header was not long enough

SQL>insert into josh_test_after values (1);

*** ERROR[29443] Database connection does not exist. Please connect to the 
database by using the connect or the reconnect command.”.

For expect 2:
ISSUE 1:  By checking the QID status of the SQL statement ‘STMT_A’, it normally 
displays
“SQL>get statistics for qid
MXID110020366282123314488178785910206U300_325_SQL_CUR_5;+>

Qid  
MXID110020366282123314488178785910206U300_325_SQL_CUR_5
Compile Start Time   2016/05/31 10:02:56.968052
Compile End Time 2016/05/31 10:02:59.207231
Compile Elapsed Time 0:00:02.239179
Execute Start Time   2016/05/31 10:02:59.207468
Execute End Time -1   
Execute Elapsed Time 0:05:46.731948
StateCLOSE”, back to TRAFCI interface, the SQL statement 
‘STMT_A’ cannot be return in time but hang for a long time.
ISSUE 2: at the same time, open a new TRAFCI that is connected to other node 
for example nap102, do SQL query statement ‘STMT_B’, using command ‘./offender 
-s active’ to check the QID status of the SQL statement ‘STMT_B’,  but print 
the following error and say 0 row(s) selected
“*** ERROR[8921] The request to obtain runtime statistics for ACTIVE_QUERIES=30 
timed out. Timeout period specified is 4 seconds.

--- 0 row(s) selected.
>>”
BTW, All TRAFCI session of ‘STMT_A’ and ‘STMT_B’ are closed normally.

T-2
Command: sqcheckDTM DownRMS DownDCS Master Down DCS 
Server Down MxoSrvr DownComments
Expect  1   2   0   1   4   Return in 1 minute
Actual  1   2   0   1   4   Return in 1 minute

T-3
Command: dcscheck   DCS Master Down DCS Server Down MxoSrvr DownComments
Expect  0   1   4   Return in 1 minute
Actual  0   1   4   Return about 5 minutes

T-4
Command: shell -c node info Comments
Expect  Only nap104 node down, return in 30 seconds
Actual  Only nap104 node down, return in 10 seconds

T-5
Command: cstat  Comments
Expect  Return in 30 seconds. 
Actual  Return about 3 minutes.

T-6
Commands: trafciComments
Expect  1.  Login success, Return in 1 minute whatever login success or 
failed.
2.  Run new 

[jira] [Updated] (TRAFODION-2063) SQL Errror 8102 occurred when down the region server

2016-06-12 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek updated TRAFODION-2063:
-
Affects Version/s: 2.0-incubating

> SQL Errror 8102 occurred when down the region server
> 
>
> Key: TRAFODION-2063
> URL: https://issues.apache.org/jira/browse/TRAFODION-2063
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: sql-general
>Affects Versions: 2.0-incubating, 2.1-incubating
> Environment: OS Type and Version: Centos  release 6.7
> Trafodion SW Version: 2.1
> CDH Version: 5.5.4
> JDK Version: 1.7.0_67
>Reporter: Jarek
> Attachments: 
> coast_2016-06-08_16.05.46.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log
>
>
> Issue Description:
> The following SQL Error 8102 occurred when down an region server.
> "16:09:28  ***ERROR: SQLExecDirect: Expected: SQL_SUCCESS Actual: SQL_ERROR
>File: ../../../../src/coast/coast_unified/src/jmtload.c   Line: 187
>State: 23000
>Native Error: 4294959194
>Error: [Trafodion ODBC Driver][Trafodion Database] SQL ERROR:*** 
> ERROR[8102] The operation is prevented by a unique constraint. [2016-06-08 
> 16:09:28]"
>
> Steps:
> Steps (cluster administrator page http://10.10.10.161:7180, username: admin, 
> password: admin)
> Step 1. Enable High Availability in cluster testing environment.
> Step 2. Run multi threads program that have connected to 32 mxosrsvr(s).
> Step 3. Stop region server on centosha-5.novalocal node
> Step 4. Check output of the multi threads program.
> 1) output as below,
> DEBUG: plan to delete total rows is  180385
> DEBUG: remaining rows to delete is  180296
> DEBUG: Remaining rows 310886
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 89
> DEBUG: Missing insertion rows 25  # please note 
> the missing insertion rows because of the above error 8102.
> Total Tests=1  Failed=1
> 2) log file attached, 
> “/opt/share/jarek/testing/odbc/builds_unix/coast/unified/linux64/coast_2016-06-08_16.05.46.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log”.
> BTW, our table structure is below, and we can see the cache size is 25 that 
> is used to avoid concurrency conflict when insert data by multi threads 
> program, but it still occurred.
>  
> SQL>showddl trafodion.j_schema_2.j_table_1;
>  
>  
> CREATE TABLE TRAFODION.J_SCHEMA_2.J_TABLE_1
>   ( 
> C0   LARGEINT GENERATED BY DEFAULT AS IDENTITY
>   (  START WITH 1  INCREMENT BY 1  MAXVALUE 9223372036854775806  MINVALUE 
> 1
>CACHE 25  NO CYCLE  LARGEINT  ) NOT NULL NOT DROPPABLE SERIALIZED
>   , C1   CHAR(20) CHARACTER SET ISO88591 COLLATE
>   DEFAULT NO DEFAULT NOT NULL NOT DROPPABLE SERIALIZED
>   , C2   INT NO DEFAULT SERIALIZED
>   , C3   SMALLINT NO DEFAULT NOT NULL NOT 
> DROPPABLE
>   SERIALIZED
>   , C4   DOUBLE PRECISION DEFAULT
>   -1.727233711019E-76 NOT SERIALIZED
>   , C5   DOUBLE PRECISION DEFAULT
>   -2.2250738585072014E-308 NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C6   DATE DEFAULT CURRENT NOT SERIALIZED
>   , C7   TIME(0) DEFAULT CURRENT NOT SERIALIZED
>   , C8   DECIMAL(18, 10) DEFAULT
>   12345678.1234567890 NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C9   LARGEINT DEFAULT 9.223E18 SERIALIZED
>   , C10  NUMERIC(128, 0) DEFAULT
>   
> 12345678901234567890123456789012345678901234567890123456789012345678901234567890123456789012345678901234567890123456789012345678
>   NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C11  REAL DEFAULT -1.1579208E38 NOT SERIALIZED
>   , C12  INTERVAL YEAR(5) TO MONTH DEFAULT NULL 
> NOT
>   SERIALIZED
>   , C13  CHAR(12) CHARACTER SET ISO88591 COLLATE
>   DEFAULT UPSHIFT DEFAULT _ISO88591'defaULT' SERIALIZED
>   , C14  CHAR(8) CHARACTER SET ISO88591 COLLATE
>  DEFAULT DEFAULT _ISO88591'summer' SERIALIZED
>   , C15  VARCHAR(10) CHARACTER SET ISO88591 
> COLLATE
>   DEFAULT DEFAULT _ISO88591'china' SERIALIZED
>   , CTIMESTAMP(6) DEFAULT CURRENT NOT
>   SERIALIZED
>   , PRIMARY KEY (C0 ASC)
>   )
> ;
>  
> -- The following sequence is a system created sequence --
>  
> CREATE SEQUENCE TRAFODION.J_SCHEMA_2."_TRAFODION_J_SCHEMA_2_J_TABLE_1_C0_" /* 
> INTERNAL */ 
>   START WITH 1 /* NEXT AVAILABLE VALUE 311001 */
>   INCREMENT BY 1
>   MAXVALUE 9223372036854775806
>   MINVALUE 1
>   CACHE 25
>   NO CYCLE
>   LARGEINT 
> ;
>  
> --- SQL operation 

[jira] [Updated] (TRAFODION-2063) SQL Errror 8102 occurred when down the region server

2016-06-12 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek updated TRAFODION-2063:
-
Attachment: 
coast_2016-06-08_16.05.46.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log

> SQL Errror 8102 occurred when down the region server
> 
>
> Key: TRAFODION-2063
> URL: https://issues.apache.org/jira/browse/TRAFODION-2063
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: sql-general
>Affects Versions: 2.1-incubating
> Environment: OS Type and Version: Centos  release 6.7
> Trafodion SW Version: 2.1
> CDH Version: 5.5.4
> JDK Version: 1.7.0_67
>Reporter: Jarek
> Attachments: 
> coast_2016-06-08_16.05.46.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log
>
>
> Issue Description:
> The following SQL Error 8102 occurred when down an region server.
> "16:09:28  ***ERROR: SQLExecDirect: Expected: SQL_SUCCESS Actual: SQL_ERROR
>File: ../../../../src/coast/coast_unified/src/jmtload.c   Line: 187
>State: 23000
>Native Error: 4294959194
>Error: [Trafodion ODBC Driver][Trafodion Database] SQL ERROR:*** 
> ERROR[8102] The operation is prevented by a unique constraint. [2016-06-08 
> 16:09:28]"
>
> Steps:
> Steps (cluster administrator page http://10.10.10.161:7180, username: admin, 
> password: admin)
> Step 1. Enable High Availability in cluster testing environment.
> Step 2. Run multi threads program that have connected to 32 mxosrsvr(s).
> Step 3. Stop region server on centosha-5.novalocal node
> Step 4. Check output of the multi threads program.
> 1) output as below,
> DEBUG: plan to delete total rows is  180385
> DEBUG: remaining rows to delete is  180296
> DEBUG: Remaining rows 310886
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 89
> DEBUG: Missing insertion rows 25  # please note 
> the missing insertion rows because of the above error 8102.
> Total Tests=1  Failed=1
> 2) log file attached, 
> “/opt/share/jarek/testing/odbc/builds_unix/coast/unified/linux64/coast_2016-06-08_16.05.46.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log”.
> BTW, our table structure is below, and we can see the cache size is 25 that 
> is used to avoid concurrency conflict when insert data by multi threads 
> program, but it still occurred.
>  
> SQL>showddl trafodion.j_schema_2.j_table_1;
>  
>  
> CREATE TABLE TRAFODION.J_SCHEMA_2.J_TABLE_1
>   ( 
> C0   LARGEINT GENERATED BY DEFAULT AS IDENTITY
>   (  START WITH 1  INCREMENT BY 1  MAXVALUE 9223372036854775806  MINVALUE 
> 1
>CACHE 25  NO CYCLE  LARGEINT  ) NOT NULL NOT DROPPABLE SERIALIZED
>   , C1   CHAR(20) CHARACTER SET ISO88591 COLLATE
>   DEFAULT NO DEFAULT NOT NULL NOT DROPPABLE SERIALIZED
>   , C2   INT NO DEFAULT SERIALIZED
>   , C3   SMALLINT NO DEFAULT NOT NULL NOT 
> DROPPABLE
>   SERIALIZED
>   , C4   DOUBLE PRECISION DEFAULT
>   -1.727233711019E-76 NOT SERIALIZED
>   , C5   DOUBLE PRECISION DEFAULT
>   -2.2250738585072014E-308 NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C6   DATE DEFAULT CURRENT NOT SERIALIZED
>   , C7   TIME(0) DEFAULT CURRENT NOT SERIALIZED
>   , C8   DECIMAL(18, 10) DEFAULT
>   12345678.1234567890 NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C9   LARGEINT DEFAULT 9.223E18 SERIALIZED
>   , C10  NUMERIC(128, 0) DEFAULT
>   
> 12345678901234567890123456789012345678901234567890123456789012345678901234567890123456789012345678901234567890123456789012345678
>   NOT NULL NOT DROPPABLE NOT SERIALIZED
>   , C11  REAL DEFAULT -1.1579208E38 NOT SERIALIZED
>   , C12  INTERVAL YEAR(5) TO MONTH DEFAULT NULL 
> NOT
>   SERIALIZED
>   , C13  CHAR(12) CHARACTER SET ISO88591 COLLATE
>   DEFAULT UPSHIFT DEFAULT _ISO88591'defaULT' SERIALIZED
>   , C14  CHAR(8) CHARACTER SET ISO88591 COLLATE
>  DEFAULT DEFAULT _ISO88591'summer' SERIALIZED
>   , C15  VARCHAR(10) CHARACTER SET ISO88591 
> COLLATE
>   DEFAULT DEFAULT _ISO88591'china' SERIALIZED
>   , CTIMESTAMP(6) DEFAULT CURRENT NOT
>   SERIALIZED
>   , PRIMARY KEY (C0 ASC)
>   )
> ;
>  
> -- The following sequence is a system created sequence --
>  
> CREATE SEQUENCE TRAFODION.J_SCHEMA_2."_TRAFODION_J_SCHEMA_2_J_TABLE_1_C0_" /* 
> INTERNAL */ 
>   START WITH 1 /* NEXT AVAILABLE VALUE 311001 */
>   INCREMENT BY 1
>   MAXVALUE 9223372036854775806
>   MINVALUE 1
>   CACHE 25
>   NO CYCLE
>   LARGEINT 
> ;

[jira] [Closed] (TRAFODION-2061) SQL Error 1583 "Sequence metadata could not be updated"

2016-06-11 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek closed TRAFODION-2061.

Resolution: Duplicate

> SQL Error 1583 "Sequence metadata could not be updated"
> ---
>
> Key: TRAFODION-2061
> URL: https://issues.apache.org/jira/browse/TRAFODION-2061
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: sql-general
>Affects Versions: 2.1-incubating
> Environment: OS Type and Version: Centos  release 6.7
> Trafodion SW Version: 2.1
> CDH Version: 5.5.4
> JDK Version: 1.7.0_67
>Reporter: Jarek
> Attachments: 
> coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log, 
> coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log
>
>
> Issue Description:
> When insert data using multi threads program, some insertion data is missing 
> because of SQL error 1583 "Sequence metadata could not be updated" occurred.
> Test Steps:
> Step 0. Hive Metastore Server Instance is running.
> Step 1. Start multi threads program.
> Step 2. Stop Hive Metastore Server Instance.
> Step 3. Back to multi threads program, it still running, waiting it completed.
> Step 4. Check multi threads program output and the log file.
> 1) output is,
> DEBUG: plan to delete total rows is  291488
> DEBUG: remaining rows to delete is  291400
> DEBUG: Remaining rows 310909
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 88
> DEBUG: Missing insertion rows 3# please note that here 3 is 
> not fixed, if you run several times, it will be changed other number.
> Total Tests=1  Failed=1
> 2) log file attached, 
> "coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"
> 13:46:52  ***ERROR: SQLExecDirect: Expected: SQL_SUCCESS Actual: SQL_ERROR
>File: ../../../../src/coast/coast_unified/src/jmtload.c   Line: 187
>   # please note that here 187 line is used to insert a row 
> into table
>State: X01G7
>Native Error: 4294965713
>Error: [Trafodion ODBC Driver][Trafodion Database] SQL ERROR:*** 
> ERROR[1583] Sequence metadata could not be updated. [2016-06-07 13:46:52]
> By the way, if skip step 3, no this kind error occurred, the output and log 
> file below,
> 1) output is,
> DEBUG: plan to delete total rows is  145895
> DEBUG: remaining rows to delete is  145809
> DEBUG: Remaining rows 310914
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 86
> DEBUG: Missing insertion rows 0
> Total Tests=1  Failed=0
> 2) log file attached, 
> "coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (TRAFODION-2061) SQL Error 1583 "Sequence metadata could not be updated"

2016-06-11 Thread Jarek (JIRA)

 [ 
https://issues.apache.org/jira/browse/TRAFODION-2061?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jarek updated TRAFODION-2061:
-
Attachment: 
coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log

coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log

> SQL Error 1583 "Sequence metadata could not be updated"
> ---
>
> Key: TRAFODION-2061
> URL: https://issues.apache.org/jira/browse/TRAFODION-2061
> Project: Apache Trafodion
>  Issue Type: Bug
>  Components: sql-general
>Affects Versions: 2.1-incubating
> Environment: OS Type and Version: Centos  release 6.7
> Trafodion SW Version: 2.1
> CDH Version: 5.5.4
> JDK Version: 1.7.0_67
>Reporter: Jarek
> Attachments: 
> coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log, 
> coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log
>
>
> Issue Description:
> When insert data using multi threads program, some insertion data is missing 
> because of SQL error 1583 "Sequence metadata could not be updated" occurred.
> Test Steps:
> Step 0. Hive Metastore Server Instance is running.
> Step 1. Start multi threads program.
> Step 2. Stop Hive Metastore Server Instance.
> Step 3. Back to multi threads program, it still running, waiting it completed.
> Step 4. Check multi threads program output and the log file.
> 1) output is,
> DEBUG: plan to delete total rows is  291488
> DEBUG: remaining rows to delete is  291400
> DEBUG: Remaining rows 310909
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 88
> DEBUG: Missing insertion rows 3# please note that here 3 is 
> not fixed, if you run several times, it will be changed other number.
> Total Tests=1  Failed=1
> 2) log file attached, 
> "coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"
> 13:46:52  ***ERROR: SQLExecDirect: Expected: SQL_SUCCESS Actual: SQL_ERROR
>File: ../../../../src/coast/coast_unified/src/jmtload.c   Line: 187
>   # please note that here 187 line is used to insert a row 
> into table
>State: X01G7
>Native Error: 4294965713
>Error: [Trafodion ODBC Driver][Trafodion Database] SQL ERROR:*** 
> ERROR[1583] Sequence metadata could not be updated. [2016-06-07 13:46:52]
> By the way, if skip step 3, no this kind error occurred, the output and log 
> file below,
> 1) output is,
> DEBUG: plan to delete total rows is  145895
> DEBUG: remaining rows to delete is  145809
> DEBUG: Remaining rows 310914
> DEBUG: Total rows 311000
> DEBUG: Deleted rows 86
> DEBUG: Missing insertion rows 0
> Total Tests=1  Failed=0
> 2) log file attached, 
> "coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (TRAFODION-2061) SQL Error 1583 "Sequence metadata could not be updated"

2016-06-11 Thread Jarek (JIRA)
Jarek created TRAFODION-2061:


 Summary: SQL Error 1583 "Sequence metadata could not be updated"
 Key: TRAFODION-2061
 URL: https://issues.apache.org/jira/browse/TRAFODION-2061
 Project: Apache Trafodion
  Issue Type: Bug
  Components: sql-general
Affects Versions: 2.1-incubating
 Environment: OS Type and Version: Centos  release 6.7
Trafodion SW Version: 2.1
CDH Version: 5.5.4
JDK Version: 1.7.0_67
Reporter: Jarek


Issue Description:
When insert data using multi threads program, some insertion data is missing 
because of SQL error 1583 "Sequence metadata could not be updated" occurred.

Test Steps:
Step 0. Hive Metastore Server Instance is running.
Step 1. Start multi threads program.
Step 2. Stop Hive Metastore Server Instance.
Step 3. Back to multi threads program, it still running, waiting it completed.
Step 4. Check multi threads program output and the log file.
1) output is,
DEBUG: plan to delete total rows is  291488
DEBUG: remaining rows to delete is  291400
DEBUG: Remaining rows 310909
DEBUG: Total rows 311000
DEBUG: Deleted rows 88
DEBUG: Missing insertion rows 3# please note that here 3 is not 
fixed, if you run several times, it will be changed other number.
Total Tests=1  Failed=1
2) log file attached, 
"coast_2016-06-07_13.43.58.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"
13:46:52  ***ERROR: SQLExecDirect: Expected: SQL_SUCCESS Actual: SQL_ERROR
   File: ../../../../src/coast/coast_unified/src/jmtload.c   Line: 187  
# please note that here 187 line is used to insert a row into 
table
   State: X01G7
   Native Error: 4294965713
   Error: [Trafodion ODBC Driver][Trafodion Database] SQL ERROR:*** ERROR[1583] 
Sequence metadata could not be updated. [2016-06-07 13:46:52]
By the way, if skip step 3, no this kind error occurred, the output and log 
file below,
1) output is,
DEBUG: plan to delete total rows is  145895
DEBUG: remaining rows to delete is  145809
DEBUG: Remaining rows 310914
DEBUG: Total rows 311000
DEBUG: Deleted rows 86
DEBUG: Missing insertion rows 0
Total Tests=1  Failed=0
2) log file attached, 
"coast_2016-06-07_13.53.49.ANSI.GBK.MultiThread.linux64.TRAF_GBK.log"






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)