Re: Spark ODBC Driver Windows Desktop problem
If someone need, on tableau desktop 8.3 all is ok. >Вторник, 26 января 2016, 18:04 +03:00 от Я : > > >Hi, i'm trying to connect tableau desktop 9.2 to spark sql. >i'm using this guide http://samx18.io/blog/2015/09/05/tableau-spark-hive.html >but on last step, then i'm try to get the contents of the table i'm getting >only rows count and empty table. > >hadoop 2.6 hive 1.21 spark 1.4.1 tableau desktop 9.2 Simba Spark ODBC Driver >1.0 (64-bit) > >thriftserver log > >16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Running query 'SELECT 1 >AS `number_of_records` >FROM `default`.`event_96720107` `event_96720107` >LIMIT 10' >16/01/26 17:57:24 INFO ParseDriver: Parsing command: SELECT 1 AS >`number_of_records` >FROM `default`.`event_96720107` `event_96720107` >LIMIT 10 >16/01/26 17:57:24 INFO ParseDriver: Parse Completed >16/01/26 17:57:24 INFO HiveMetaStore: 2: get_table : db=default >tbl=event_96720107 >16/01/26 17:57:24 INFO audit: ugi=hduser ip=unknown-ip-addr >cmd=get_table : db=default tbl=event_96720107 >16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(400328) called with >curMem=1356967, maxMem=278019440 >16/01/26 17:57:24 INFO MemoryStore: Block broadcast_26 stored as values in >memory (estimated size 390.9 KB, free 263.5 MB) >16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(36040) called with >curMem=1757295, maxMem=278019440 >16/01/26 17:57:24 INFO MemoryStore: Block broadcast_26_piece0 stored as bytes >in memory (estimated size 35.2 KB, free 263.4 MB) >16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_26_piece0 in memory >on 192.168.245.128:43526 (size: 35.2 KB, free: 265.0 MB) >16/01/26 17:57:24 INFO SparkContext: Created broadcast 26 from Spark JDBC >Server Query >16/01/26 17:57:24 INFO FileInputFormat: Total input paths to process : 1 >16/01/26 17:57:24 INFO SparkContext: Starting job: Spark JDBC Server Query >16/01/26 17:57:24 INFO DAGScheduler: Got job 14 (Spark JDBC Server Query) with >1 output partitions (allowLocal=false) >16/01/26 17:57:24 INFO DAGScheduler: Final stage: ResultStage 14(Spark JDBC >Server Query) >16/01/26 17:57:24 INFO DAGScheduler: Parents of final stage: List() >16/01/26 17:57:24 INFO DAGScheduler: Missing parents: List() >16/01/26 17:57:24 INFO DAGScheduler: Submitting ResultStage 14 >(MapPartitionsRDD[84] at Spark JDBC Server Query), which has no missing parents >16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(16768) called with >curMem=1793335, maxMem=278019440 >16/01/26 17:57:24 INFO MemoryStore: Block broadcast_27 stored as values in >memory (estimated size 16.4 KB, free 263.4 MB) >16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(7945) called with >curMem=1810103, maxMem=278019440 >16/01/26 17:57:24 INFO MemoryStore: Block broadcast_27_piece0 stored as bytes >in memory (estimated size 7.8 KB, free 263.4 MB) >16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_27_piece0 in memory >on 192.168.245.128:43526 (size: 7.8 KB, free: 265.0 MB) >16/01/26 17:57:24 INFO SparkContext: Created broadcast 27 from broadcast at >DAGScheduler.scala:874 >16/01/26 17:57:24 INFO DAGScheduler: Submitting 1 missing tasks from >ResultStage 14 (MapPartitionsRDD[84] at Spark JDBC Server Query) >16/01/26 17:57:24 INFO TaskSchedulerImpl: Adding task set 14.0 with 1 tasks >16/01/26 17:57:24 INFO TaskSetManager: Starting task 0.0 in stage 14.0 (TID >14, 192.168.245.128, ANY, 1448 bytes) >16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_27_piece0 in memory >on 192.168.245.128:49810 (size: 7.8 KB, free: 265.0 MB) >16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_26_piece0 in memory >on 192.168.245.128:49810 (size: 35.2 KB, free: 265.0 MB) >16/01/26 17:57:24 INFO TaskSetManager: Finished task 0.0 in stage 14.0 (TID >14) in 89 ms on 192.168.245.128 (1/1) >16/01/26 17:57:24 INFO TaskSchedulerImpl: Removed TaskSet 14.0, whose tasks >have all completed, from pool >16/01/26 17:57:24 INFO DAGScheduler: ResultStage 14 (Spark JDBC Server Query) >finished in 0.088 s >16/01/26 17:57:24 INFO DAGScheduler: Job 14 finished: Spark JDBC Server Query, >took 0.105733 s >16/01/26 17:57:24 INFO StatsReportListener: Finished stage: >org.apache.spark.scheduler.StageInfo@108a7 >16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Result Schema: >ArrayBuffer(number_of_records#745) >16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Result Schema: >ArrayBuffer(number_of_records#745) >16/01/26 17:57:24 INFO StatsReportListener: task runtime:(count: 1, mean: >89.00, stdev: 0.00, max: 89.00, min: 89.00) >16/01/26 17:57:24 INFO StatsReportListener: 0% 5% 10% 25% 50% > 75% 90% 95% 100% >16/01/26 17:57:24 INFO StatsReportListener: 89.0 ms 89.0 ms 89.0 ms > 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms >16/01/26 17:57:24 INFO StatsReportListener: task result size:(count: 1, mean: >1925.00, stdev: 0.00, max: 1925.00, min: 1925.00) >16/01/26 17:57
Spark ODBC Driver Windows Desktop problem
Hi, i'm trying to connect tableau desktop 9.2 to spark sql. i'm using this guide http://samx18.io/blog/2015/09/05/tableau-spark-hive.html but on last step, then i'm try to get the contents of the table i'm getting only rows count and empty table. hadoop 2.6 hive 1.21 spark 1.4.1 tableau desktop 9.2 Simba Spark ODBC Driver 1.0 (64-bit) thriftserver log 16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Running query 'SELECT 1 AS `number_of_records` FROM `default`.`event_96720107` `event_96720107` LIMIT 10' 16/01/26 17:57:24 INFO ParseDriver: Parsing command: SELECT 1 AS `number_of_records` FROM `default`.`event_96720107` `event_96720107` LIMIT 10 16/01/26 17:57:24 INFO ParseDriver: Parse Completed 16/01/26 17:57:24 INFO HiveMetaStore: 2: get_table : db=default tbl=event_96720107 16/01/26 17:57:24 INFO audit: ugi=hduser ip=unknown-ip-addr cmd=get_table : db=default tbl=event_96720107 16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(400328) called with curMem=1356967, maxMem=278019440 16/01/26 17:57:24 INFO MemoryStore: Block broadcast_26 stored as values in memory (estimated size 390.9 KB, free 263.5 MB) 16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(36040) called with curMem=1757295, maxMem=278019440 16/01/26 17:57:24 INFO MemoryStore: Block broadcast_26_piece0 stored as bytes in memory (estimated size 35.2 KB, free 263.4 MB) 16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_26_piece0 in memory on 192.168.245.128:43526 (size: 35.2 KB, free: 265.0 MB) 16/01/26 17:57:24 INFO SparkContext: Created broadcast 26 from Spark JDBC Server Query 16/01/26 17:57:24 INFO FileInputFormat: Total input paths to process : 1 16/01/26 17:57:24 INFO SparkContext: Starting job: Spark JDBC Server Query 16/01/26 17:57:24 INFO DAGScheduler: Got job 14 (Spark JDBC Server Query) with 1 output partitions (allowLocal=false) 16/01/26 17:57:24 INFO DAGScheduler: Final stage: ResultStage 14(Spark JDBC Server Query) 16/01/26 17:57:24 INFO DAGScheduler: Parents of final stage: List() 16/01/26 17:57:24 INFO DAGScheduler: Missing parents: List() 16/01/26 17:57:24 INFO DAGScheduler: Submitting ResultStage 14 (MapPartitionsRDD[84] at Spark JDBC Server Query), which has no missing parents 16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(16768) called with curMem=1793335, maxMem=278019440 16/01/26 17:57:24 INFO MemoryStore: Block broadcast_27 stored as values in memory (estimated size 16.4 KB, free 263.4 MB) 16/01/26 17:57:24 INFO MemoryStore: ensureFreeSpace(7945) called with curMem=1810103, maxMem=278019440 16/01/26 17:57:24 INFO MemoryStore: Block broadcast_27_piece0 stored as bytes in memory (estimated size 7.8 KB, free 263.4 MB) 16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_27_piece0 in memory on 192.168.245.128:43526 (size: 7.8 KB, free: 265.0 MB) 16/01/26 17:57:24 INFO SparkContext: Created broadcast 27 from broadcast at DAGScheduler.scala:874 16/01/26 17:57:24 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 14 (MapPartitionsRDD[84] at Spark JDBC Server Query) 16/01/26 17:57:24 INFO TaskSchedulerImpl: Adding task set 14.0 with 1 tasks 16/01/26 17:57:24 INFO TaskSetManager: Starting task 0.0 in stage 14.0 (TID 14, 192.168.245.128, ANY, 1448 bytes) 16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_27_piece0 in memory on 192.168.245.128:49810 (size: 7.8 KB, free: 265.0 MB) 16/01/26 17:57:24 INFO BlockManagerInfo: Added broadcast_26_piece0 in memory on 192.168.245.128:49810 (size: 35.2 KB, free: 265.0 MB) 16/01/26 17:57:24 INFO TaskSetManager: Finished task 0.0 in stage 14.0 (TID 14) in 89 ms on 192.168.245.128 (1/1) 16/01/26 17:57:24 INFO TaskSchedulerImpl: Removed TaskSet 14.0, whose tasks have all completed, from pool 16/01/26 17:57:24 INFO DAGScheduler: ResultStage 14 (Spark JDBC Server Query) finished in 0.088 s 16/01/26 17:57:24 INFO DAGScheduler: Job 14 finished: Spark JDBC Server Query, took 0.105733 s 16/01/26 17:57:24 INFO StatsReportListener: Finished stage: org.apache.spark.scheduler.StageInfo@108a7 16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Result Schema: ArrayBuffer(number_of_records#745) 16/01/26 17:57:24 INFO SparkExecuteStatementOperation: Result Schema: ArrayBuffer(number_of_records#745) 16/01/26 17:57:24 INFO StatsReportListener: task runtime:(count: 1, mean: 89.00, stdev: 0.00, max: 89.00, min: 89.00) 16/01/26 17:57:24 INFO StatsReportListener: 0% 5% 10% 25% 50% 75% 90% 95% 100% 16/01/26 17:57:24 INFO StatsReportListener: 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 89.0 ms 16/01/26 17:57:24 INFO StatsReportListener: task result size:(count: 1, mean: 1925.00, stdev: 0.00, max: 1925.00, min: 1925.00) 16/01/26 17:57:24 INFO StatsReportListener: 0% 5% 10% 25% 50% 75% 90% 95% 100% 16/01/26 17:57:24 INFO StatsReportListener: 1925.0 B 1925.0 B 1925.0 B 1925.0 B