Reka, Please include an integration test for BAM. This should have been detected much earlier.
On Tue, Oct 16, 2012 at 9:32 AM, Reka Thirunavukkarasu <r...@wso2.com>wrote: > Thanks for looking into this issue. > > Reka > > > On Tue, Oct 16, 2012 at 9:35 PM, Kasun Weranga <kas...@wso2.com> wrote: > >> This issue has been fixed. Thanks Buddhika for providing the fix. >> >> On Tue, Oct 16, 2012 at 2:31 PM, Reka Thirunavukkarasu <r...@wso2.com>wrote: >> >>> Hi >>> >>> The following hive query which works fine in BAM 2.0.0 is not working >>> now in BAM 2.0.1. Could you please have a look on this? >>> jira: https://wso2.org/jira/browse/CARBON-13943 >>> >>> set logs_column_family = "log_1_AS_2012_10_14"; >>> set file_path = >>> /media/Entertainment/stratos/BAM_LOGS/archive-logs/tmpLogs/1/AS/2012_10_14; >>> drop table LogStats; >>> set mapred.output.compress=true; >>> set hive.exec.compress.output=true; >>> set >>> mapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec; >>> set io.compression.codecs=org.apache.hadoop.io.compress.GzipCodec; >>> >>> CREATE EXTERNAL TABLE IF NOT EXISTS LogStats (key STRING, >>> payload_tenantID STRING,payload_serverName STRING, >>> payload_appName STRING,payload_message STRING, >>> payload_stacktrace STRING, >>> payload_logger STRING, >>> payload_priority STRING,payload_logTime BIGINT) >>> STORED BY >>> 'org.apache.hadoop.hive.cassandra.CassandraStorageHandler' >>> WITH SERDEPROPERTIES ( "cassandra.host" = "localhost", >>> "cassandra.port" ="9160","cassandra.ks.name" = "EVENT_KS", >>> "cassandra.ks.username" = "admin","cassandra.ks.password" >>> ="admin", >>> "cassandra.cf.name" = ${hiveconf:logs_column_family}, >>> "cassandra.columns.mapping" = >>> ":key,payload_tenantID, >>> payload_serverName,payload_appName,payload_message, >>> payload_stacktrace,payload_logger,payload_priority, >>> payload_logTime" ); >>> INSERT OVERWRITE LOCAL DIRECTORY 'file:///${hiveconf:file_path}' >>> select >>> concat('TID[',payload_tenantID, ']\t', >>> 'Server[',payload_serverName,']\t', >>> 'Application[',payload_appName,']\t', >>> 'Message[',payload_message,']\t', >>> 'Stacktrace ',payload_stacktrace,'\t', >>> 'Logger{',payload_logger,'}\t', >>> 'Priority[',payload_priority,']\t'), >>> concat('LogTime[', >>> (from_unixtime(cast(payload_logTime/1000 as BIGINT),'yyyy-MM-dd >>> HH:mm:ss.SSS' )),']\n') as LogTime from LogStats >>> ORDER BY LogTime; >>> >>> The exception trace is as following: >>> >>> Hive history file=/tmp/reka/hive_job_log_reka_201210161416_1379462536.txt >>> Hive history file=/tmp/reka/hive_job_log_reka_201210161416_844839482.txt >>> [2012-10-16 14:16:03,180] ERROR {hive.ql.metadata.Hive} - >>> NoSuchObjectException(message:default.LogStats table not found) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler$16.run(MultitenantMetaStoreHandler.java:1194) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler$16.run(MultitenantMetaStoreHandler.java:1189) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler.executeWithRetry(MultitenantMetaStoreHandler.java:324) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler.get_table(MultitenantMetaStoreHandler.java:1189) >>> at >>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.getTable(HiveMetaStoreClient.java:735) >>> at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:901) >>> at org.apache.hadoop.hive.ql.metadata.Hive.getTable(Hive.java:843) >>> at >>> org.apache.hadoop.hive.ql.exec.DDLTask.dropTable(DDLTask.java:3127) >>> at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:250) >>> at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:133) >>> at >>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:57) >>> at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1334) >>> at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1125) >>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:933) >>> at >>> org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:201) >>> at >>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:187) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:324) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:225) >>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>> at java.lang.Thread.run(Thread.java:662) >>> >>> OK >>> FAILED: Parse Error: line 1:536 mismatched input '$' expecting >>> StringLiteral near '=' in specifying key/value property >>> >>> [2012-10-16 14:16:03,186] ERROR {org.apache.hadoop.hive.ql.Driver} - >>> FAILED: Parse Error: line 1:536 mismatched input '$' expecting >>> StringLiteral near '=' in specifying key/value property >>> >>> org.apache.hadoop.hive.ql.parse.ParseException: line 1:536 mismatched >>> input '$' expecting StringLiteral near '=' in specifying key/value property >>> >>> at >>> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:438) >>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:419) >>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:339) >>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:891) >>> at >>> org.apache.hadoop.hive.service.HiveServer$HiveServerHandler.execute(HiveServer.java:201) >>> at >>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:187) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:324) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:225) >>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>> at java.lang.Thread.run(Thread.java:662) >>> >>> [2012-10-16 14:16:03,186] ERROR >>> {org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl} - Error >>> while executing Hive script. >>> Query returned non-zero code: 11, cause: FAILED: Parse Error: line 1:536 >>> mismatched input '$' expecting StringLiteral near '=' in specifying >>> key/value property >>> >>> java.sql.SQLException: Query returned non-zero code: 11, cause: FAILED: >>> Parse Error: line 1:536 mismatched input '$' expecting StringLiteral near >>> '=' in specifying key/value property >>> >>> at >>> org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:189) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:324) >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl$ScriptCallable.call(HiveExecutorServiceImpl.java:225) >>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>> at java.lang.Thread.run(Thread.java:662) >>> [2012-10-16 14:16:03,187] ERROR >>> {org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask} - Error while >>> executing script : test >>> org.wso2.carbon.analytics.hive.exception.HiveExecutionException: Error >>> while executing Hive script.Query returned non-zero code: 11, cause: >>> FAILED: Parse Error: line 1:536 mismatched input '$' expecting >>> StringLiteral near '=' in specifying key/value property >>> >>> at >>> org.wso2.carbon.analytics.hive.impl.HiveExecutorServiceImpl.execute(HiveExecutorServiceImpl.java:110) >>> at >>> org.wso2.carbon.analytics.hive.task.HiveScriptExecutorTask.execute(HiveScriptExecutorTask.java:60) >>> at >>> org.wso2.carbon.ntask.core.impl.TaskQuartzJobAdapter.execute(TaskQuartzJobAdapter.java:56) >>> at org.quartz.core.JobRunShell.run(JobRunShell.java:213) >>> at >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441) >>> at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:138) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >>> at java.lang.Thread.run(Thread.java:662) >>> [2012-10-16 14:16:03,203] ERROR {hive.ql.metadata.Hive} - >>> NoSuchObjectException(message:default.LogStats table not found) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler$16.run(MultitenantMetaStoreHandler.java:1194) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler$16.run(MultitenantMetaStoreHandler.java:1189) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler.executeWithRetry(MultitenantMetaStoreHandler.java:324) >>> at >>> org.apache.hadoop.hive.metastore.MultitenantMetaStoreHandler.get_table(MultitenantM >>> >>> Thanks, >>> Reka >>> >> >> >> >> -- >> *Kasun Weranga* >> Software Engineer >> ** >> *WSO2, Inc. >> *lean.enterprise.middleware. >> mobile : +94 772314602 >> <http://sanjeewamalalgoda.blogspot.com/>blog >> :<http://sanjeewamalalgoda.blogspot.com/> >> http://kasunweranga.blogspot.com/ >> > > > _______________________________________________ > Dev mailing list > Dev@wso2.org > http://wso2.org/cgi-bin/mailman/listinfo/dev > > -- Regards, Tharindu blog: http://mackiemathew.com/ M: +94777759908
_______________________________________________ Dev mailing list Dev@wso2.org http://wso2.org/cgi-bin/mailman/listinfo/dev