[jira] Created: (HIVE-661) Make PrimitiveJavaObjectInspectors and ReflectionObjectInspectors settable
Make PrimitiveJavaObjectInspectors and ReflectionObjectInspectors settable -- Key: HIVE-661 URL: https://issues.apache.org/jira/browse/HIVE-661 Project: Hadoop Hive Issue Type: New Feature Reporter: Zheng Shao Both primitive JavaObjectInspectors and ReflectionObjectInspectors should be settable, so that the Hive Operators can set the value of the fields if needed. Once ReflectionObjectInspectors are settable, we will be able to allow UDFs (not GenericUDFs) to accept complex parameters and return complex parameters. The efficiency won't be too high - because we probably need to do conversions, but it makes it much easier for people to write UDFs dealing with complex objects. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-662) Add a method to parse apache weblogs
[ https://issues.apache.org/jira/browse/HIVE-662?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733586#action_12733586 ] Johan Oskarsson commented on HIVE-662: -- What is the best route to take here? I would assume a custom serde is the way to go? Add a method to parse apache weblogs Key: HIVE-662 URL: https://issues.apache.org/jira/browse/HIVE-662 Project: Hadoop Hive Issue Type: New Feature Reporter: Johan Oskarsson Priority: Critical Fix For: 0.4.0 Apache weblogs is one of the more common formats for people to parse using Hadoop. Unfortunately the method provided to process the logs in Hive has some issues and seems to be on it's way out. See HIVE-519 and comments on HIVE-520. We should replace that method with something that works better and that can be supported in the future. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-662) Add a method to parse apache weblogs
Add a method to parse apache weblogs Key: HIVE-662 URL: https://issues.apache.org/jira/browse/HIVE-662 Project: Hadoop Hive Issue Type: New Feature Reporter: Johan Oskarsson Priority: Critical Fix For: 0.4.0 Apache weblogs is one of the more common formats for people to parse using Hadoop. Unfortunately the method provided to process the logs in Hive has some issues and seems to be on it's way out. See HIVE-519 and comments on HIVE-520. We should replace that method with something that works better and that can be supported in the future. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
Build failed in Hudson: Hive-trunk-h0.17 #160
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.17/160/ -- started Building remotely on minerva.apache.org (Ubuntu) Updating http://svn.apache.org/repos/asf/hadoop/hive/trunk ERROR: Failed to update http://svn.apache.org/repos/asf/hadoop/hive/trunk org.tmatesoft.svn.core.SVNException: svn: unknown host svn: OPTIONS request failed on '/repos/asf/hadoop/hive/trunk' at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:103) at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:87) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:601) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:257) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:245) at org.tmatesoft.svn.core.internal.io.dav.DAVConnection.exchangeCapabilities(DAVConnection.java:454) at org.tmatesoft.svn.core.internal.io.dav.DAVConnection.open(DAVConnection.java:97) at org.tmatesoft.svn.core.internal.io.dav.DAVRepository.openConnection(DAVRepository.java:664) at org.tmatesoft.svn.core.internal.io.dav.DAVRepository.getRepositoryUUID(DAVRepository.java:143) at org.tmatesoft.svn.core.wc.SVNBasicClient.createRepository(SVNBasicClient.java:339) at org.tmatesoft.svn.core.wc.SVNBasicClient.createRepository(SVNBasicClient.java:327) at org.tmatesoft.svn.core.wc.SVNUpdateClient.update(SVNUpdateClient.java:426) at org.tmatesoft.svn.core.wc.SVNUpdateClient.doUpdate(SVNUpdateClient.java:394) at org.tmatesoft.svn.core.wc.SVNUpdateClient.doUpdate(SVNUpdateClient.java:210) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:505) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:476) at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1283) at hudson.remoting.UserRequest.perform(UserRequest.java:69) at hudson.remoting.UserRequest.perform(UserRequest.java:23) at hudson.remoting.Request$2.run(Request.java:213) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) Caused by: java.net.UnknownHostException: svn.apache.org at java.net.Inet6AddressImpl.lookupAllHostAddr(Native Method) at java.net.InetAddress$1.lookupAllHostAddr(InetAddress.java:849) at java.net.InetAddress.getAddressFromNameService(InetAddress.java:1200) at java.net.InetAddress.getAllByName0(InetAddress.java:1153) at java.net.InetAddress.getAllByName(InetAddress.java:1083) at java.net.InetAddress.getAllByName(InetAddress.java:1019) at java.net.InetAddress.getByName(InetAddress.java:969) at org.tmatesoft.svn.core.internal.util.SVNSocketFactory.createAddres(SVNSocketFactory.java:108) at org.tmatesoft.svn.core.internal.util.SVNSocketFactory.createPlainSocket(SVNSocketFactory.java:51) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.connect(HTTPConnection.java:167) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:307) ... 23 more Publishing Javadoc Recording test results
[jira] Commented: (HIVE-642) udf equivalent to string split
[ https://issues.apache.org/jira/browse/HIVE-642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733630#action_12733630 ] Namit Jain commented on HIVE-642: - +1 looks good udf equivalent to string split -- Key: HIVE-642 URL: https://issues.apache.org/jira/browse/HIVE-642 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Assignee: Emil Ibrishimov Fix For: 0.4.0 Attachments: HIVE-642.1.patch, HIVE-642.2.patch It would be very useful to have a function equivalent to string split in java -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-639) Create a contrib directory for people to contribute file formats, serdes, and UDFs
[ https://issues.apache.org/jira/browse/HIVE-639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733639#action_12733639 ] Namit Jain commented on HIVE-639: - Don't you need a target to build the jar file: hive_contrib.jar ? The overall directory structure looks good. Create a contrib directory for people to contribute file formats, serdes, and UDFs -- Key: HIVE-639 URL: https://issues.apache.org/jira/browse/HIVE-639 Project: Hadoop Hive Issue Type: New Feature Components: Build Infrastructure Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-639.1.patch We want to add a contrib directory so people can contribute their own file formats, serdes, and UDFs easily. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-642) udf equivalent to string split
[ https://issues.apache.org/jira/browse/HIVE-642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733641#action_12733641 ] Min Zhou commented on HIVE-642: --- It's very useful for us . some comments: # Can you implement it directly with Text ? Avoiding string decoding and encoding would be faster. Of course that trick may lead to another problem, as String.split uses a regular expression for splitting. # getDisplayString() always return a string in lowercase. udf equivalent to string split -- Key: HIVE-642 URL: https://issues.apache.org/jira/browse/HIVE-642 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Assignee: Emil Ibrishimov Fix For: 0.4.0 Attachments: HIVE-642.1.patch, HIVE-642.2.patch It would be very useful to have a function equivalent to string split in java -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
Build failed in Hudson: Hive-trunk-h0.19 #162
See http://hudson.zones.apache.org/hudson/job/Hive-trunk-h0.19/162/ -- started Building remotely on minerva.apache.org (Ubuntu) Checking out http://svn.apache.org/repos/asf/hadoop/hive/trunk ERROR: Failed to check out http://svn.apache.org/repos/asf/hadoop/hive/trunk org.tmatesoft.svn.core.SVNException: svn: unknown host svn: OPTIONS request failed on '/repos/asf/hadoop/hive/trunk' at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:103) at org.tmatesoft.svn.core.internal.wc.SVNErrorManager.error(SVNErrorManager.java:87) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:601) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:257) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:245) at org.tmatesoft.svn.core.internal.io.dav.DAVConnection.exchangeCapabilities(DAVConnection.java:454) at org.tmatesoft.svn.core.internal.io.dav.DAVConnection.open(DAVConnection.java:97) at org.tmatesoft.svn.core.internal.io.dav.DAVRepository.openConnection(DAVRepository.java:664) at org.tmatesoft.svn.core.internal.io.dav.DAVRepository.getLatestRevision(DAVRepository.java:172) at org.tmatesoft.svn.core.wc.SVNBasicClient.getRevisionNumber(SVNBasicClient.java:479) at org.tmatesoft.svn.core.wc.SVNBasicClient.getLocations(SVNBasicClient.java:848) at org.tmatesoft.svn.core.wc.SVNBasicClient.createRepository(SVNBasicClient.java:531) at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:830) at org.tmatesoft.svn.core.wc.SVNUpdateClient.doCheckout(SVNUpdateClient.java:728) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:546) at hudson.scm.SubversionSCM$CheckOutTask.invoke(SubversionSCM.java:476) at hudson.FilePath$FileCallableWrapper.call(FilePath.java:1283) at hudson.remoting.UserRequest.perform(UserRequest.java:69) at hudson.remoting.UserRequest.perform(UserRequest.java:23) at hudson.remoting.Request$2.run(Request.java:213) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:441) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) at java.util.concurrent.FutureTask.run(FutureTask.java:138) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) at java.lang.Thread.run(Thread.java:619) Caused by: java.net.UnknownHostException: svn.apache.org at java.net.InetAddress.getAllByName0(InetAddress.java:1157) at java.net.InetAddress.getAllByName(InetAddress.java:1083) at java.net.InetAddress.getAllByName(InetAddress.java:1019) at java.net.InetAddress.getByName(InetAddress.java:969) at org.tmatesoft.svn.core.internal.util.SVNSocketFactory.createAddres(SVNSocketFactory.java:108) at org.tmatesoft.svn.core.internal.util.SVNSocketFactory.createPlainSocket(SVNSocketFactory.java:51) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.connect(HTTPConnection.java:167) at org.tmatesoft.svn.core.internal.io.dav.http.HTTPConnection.request(HTTPConnection.java:307) ... 23 more Recording test results
[jira] Created: (HIVE-663) column aliases should be supported
column aliases should be supported -- Key: HIVE-663 URL: https://issues.apache.org/jira/browse/HIVE-663 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain select key as x from src where x 10; should work -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-642) udf equivalent to string split
[ https://issues.apache.org/jira/browse/HIVE-642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733747#action_12733747 ] Emil Ibrishimov commented on HIVE-642: -- There are some easy (compromise) ways to optimize split: 1. Check if the regex argument actually contains some regex specific characters and if it doesn't, do a straightforward split without converting to strings. 2. Assume some default value for the second argument (for example - split(str) to be equivalent to split(str, ' ') and optimize for this value 3. Have two separate split functions - one that does regex and one that splits around plain text. I think that 1 is a good choice and can be done rather quickly. udf equivalent to string split -- Key: HIVE-642 URL: https://issues.apache.org/jira/browse/HIVE-642 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Assignee: Emil Ibrishimov Fix For: 0.4.0 Attachments: HIVE-642.1.patch, HIVE-642.2.patch It would be very useful to have a function equivalent to string split in java -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-642) udf equivalent to string split
[ https://issues.apache.org/jira/browse/HIVE-642?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733766#action_12733766 ] Namit Jain commented on HIVE-642: - filed https://issues.apache.org/jira/browse/HIVE-664 udf equivalent to string split -- Key: HIVE-642 URL: https://issues.apache.org/jira/browse/HIVE-642 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Assignee: Emil Ibrishimov Fix For: 0.4.0 Attachments: HIVE-642.1.patch, HIVE-642.2.patch It would be very useful to have a function equivalent to string split in java -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-665) Allow to query hadoop/mapreduce cluster status from hive server
Allow to query hadoop/mapreduce cluster status from hive server --- Key: HIVE-665 URL: https://issues.apache.org/jira/browse/HIVE-665 Project: Hadoop Hive Issue Type: New Feature Components: Server Infrastructure Reporter: Hao Liu Tools/infra around hadoop/hive need to check cluster status in many cases. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-666) We should not ignore exceptions in FetchTask
We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-615) new user functions needed for HIVE
[ https://issues.apache.org/jira/browse/HIVE-615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733793#action_12733793 ] Emil Ibrishimov commented on HIVE-615: -- Functions we still don't have: abs binary concat_ws field find_in_set make_set quote substring_index unhex new user functions needed for HIVE -- Key: HIVE-615 URL: https://issues.apache.org/jira/browse/HIVE-615 Project: Hadoop Hive Issue Type: New Feature Components: Query Processor Reporter: Namit Jain Assignee: Emil Ibrishimov http://dev.mysql.com/doc/refman/5.0/en/func-op-summary-ref.html Hive should support the functions above -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-657) .hiverc would be nice
[ https://issues.apache.org/jira/browse/HIVE-657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733796#action_12733796 ] Namit Jain commented on HIVE-657: - This might create a problem for unit tests. Do you want to add a configuration parameter for this - for unit tests, instead of user.home it might be better to overwrite it .hiverc would be nice - Key: HIVE-657 URL: https://issues.apache.org/jira/browse/HIVE-657 Project: Hadoop Hive Issue Type: Improvement Reporter: Adam Kramer Assignee: Raghotham Murthy Attachments: hive-657.1.patch It would be great if the hive CLI would execute ~/.hiverc on load, if it exists. There are some commands I would like to set every time I run hive. This is especially important in light of https://issues.apache.org/jira/browse/HIVE-243 as I often end up re-loading the CLI and then forgetting to re-set certain variables. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Assigned: (HIVE-666) We should not ignore exceptions in FetchTask
[ https://issues.apache.org/jira/browse/HIVE-666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao reassigned HIVE-666: --- Assignee: Zheng Shao We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-666.1.patch FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-666) We should not ignore exceptions in FetchTask
[ https://issues.apache.org/jira/browse/HIVE-666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-666: Attachment: HIVE-666.1.patch This patch propagates the exception. We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao Attachments: HIVE-666.1.patch FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-348) Provide type information to custom mappers and reducers.
[ https://issues.apache.org/jira/browse/HIVE-348?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733835#action_12733835 ] Zheng Shao commented on HIVE-348: - I think we should pass type information in the environment variable. Provide type information to custom mappers and reducers. Key: HIVE-348 URL: https://issues.apache.org/jira/browse/HIVE-348 Project: Hadoop Hive Issue Type: Improvement Reporter: Venky Iyer Either by setting an environment variable with the schema, or (ideally), configurably passing JSON data to custom mappers/reducers. This would be more data to pump into the pipe on each mapper/reducer, but it would be generated on the mapper/reducer, and it may mean much less processing for the custom script (using a fast json library compared to casting each column that needs casting based on the schema). -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-669) SELECT TRANSFORM / MAP / REDUCE to support optional ROW FORMAT
SELECT TRANSFORM / MAP / REDUCE to support optional ROW FORMAT -- Key: HIVE-669 URL: https://issues.apache.org/jira/browse/HIVE-669 Project: Hadoop Hive Issue Type: New Feature Components: Serializers/Deserializers Reporter: Zheng Shao We should allow TRANSFORM / MAP/ REDUCE to accept optional ROW FORMAT: {code} SELECT TRANSFORM(a, b, c) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\001' USING 'mypython.py' AS (d STRING, e INT, f DOUBLE) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t' {code} Also, we should allow optional type information in the AS clause. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-666) We should not ignore exceptions in FetchTask
[ https://issues.apache.org/jira/browse/HIVE-666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-666: Status: Patch Available (was: Open) We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-666.1.patch FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-666) We should not ignore exceptions in FetchTask
[ https://issues.apache.org/jira/browse/HIVE-666?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733852#action_12733852 ] Raghotham Murthy commented on HIVE-666: --- Can you add a test case where the IOException is thrown only with this patch? We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-666.1.patch FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-454) Support escaping of ; in strings in cli
[ https://issues.apache.org/jira/browse/HIVE-454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733858#action_12733858 ] Raghotham Murthy commented on HIVE-454: --- It seems cleaner to modify hive.g instead of adding more parsing/manipulation of query strings in CliDriver. Support escaping of ; in strings in cli --- Key: HIVE-454 URL: https://issues.apache.org/jira/browse/HIVE-454 Project: Hadoop Hive Issue Type: Bug Components: Clients Affects Versions: 0.3.0 Reporter: Ashish Thusoo Assignee: Ashish Thusoo Attachments: hive-454.patch If ; appears in string literals in a query the hive cli is not able to escape them properly. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-396) Hive performance benchmarks
[ https://issues.apache.org/jira/browse/HIVE-396?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Yuntao Jia updated HIVE-396: Attachment: hive_benchmark_2009-07-21.tar.gz Updated the benchmark script to make it more automatic. Now it outputs all the timings to a csv file which looks like: Timings, grep select, rankings select, uservisits aggregation, uservisits-rankings join Trial 1 Hive,126.3,25.0,546.1,447.9, PIG,240.5,31.0,672.3,658.3, Hadoop,135.4,21.6,394.9,486.1 Trial 2 Hive,126.3,25.0,546.1,447.9, PIG,240.5,31.0,672.3,658.3, Hadoop,135.4,21.6,394.9,486.1 Trial 3 Hive,126.3,25.0,546.1,447.9, PIG,240.5,31.0,672.3,658.3, Hadoop,135.4,21.6,394.9,486.1 The first line shows the queries, followed by query timings from different trials. Within each trial, there are three lines showing the query timings on Hive, PIG and Hadoop, respectively. The numbers here are for illustration purpose only. The file can be directly opened in excel. User can then easily generate a performance graph on top of it Hive performance benchmarks --- Key: HIVE-396 URL: https://issues.apache.org/jira/browse/HIVE-396 Project: Hadoop Hive Issue Type: New Feature Reporter: Zheng Shao Assignee: Yuntao Jia Attachments: hive_benchmark_2009-06-18.pdf, hive_benchmark_2009-06-18.tar.gz, hive_benchmark_2009-07-12.pdf, hive_benchmark_2009-07-21.tar.gz We need some performance benchmark to measure and track the performance improvements of Hive. Some references: PIG performance benchmarks PIG-200 PigMix: http://wiki.apache.org/pig/PigMix -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Created: (HIVE-670) Explain should show output column names
Explain should show output column names --- Key: HIVE-670 URL: https://issues.apache.org/jira/browse/HIVE-670 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Reporter: Zheng Shao Explain currently only shows the expressions (in which it references the output column names of the last operator). However, it does not show the output column names of the last operator, which makes it hard to debug. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-487) Hive does not compile with Hadoop 0.20.0
[ https://issues.apache.org/jira/browse/HIVE-487?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733948#action_12733948 ] Ashish Thusoo commented on HIVE-487: This sounds reasonable to me. Will go over the patch in more detail. Are you planning to upload another one soon or should I just review this one? Hive does not compile with Hadoop 0.20.0 Key: HIVE-487 URL: https://issues.apache.org/jira/browse/HIVE-487 Project: Hadoop Hive Issue Type: Bug Affects Versions: 0.3.0 Reporter: Aaron Kimball Assignee: Justin Lynn Fix For: 0.4.0 Attachments: HIVE-487-2.patch, hive-487.3.patch, hive-487.4.patch, HIVE-487.patch, jetty-patch.patch, junit-patch1.html Attempting to compile Hive with Hadoop 0.20.0 fails: aa...@jargon:~/src/ext/svn/hive-0.3.0$ ant -Dhadoop.version=0.20.0 package (several lines elided) compile: [echo] Compiling: hive [javac] Compiling 261 source files to /home/aaron/src/ext/svn/hive-0.3.0/build/ql/classes [javac] /home/aaron/src/ext/svn/hive-0.3.0/build/ql/java/org/apache/hadoop/hive/ql/exec/ExecDriver.java:94: cannot find symbol [javac] symbol : method getCommandLineConfig() [javac] location: class org.apache.hadoop.mapred.JobClient [javac] Configuration commandConf = JobClient.getCommandLineConfig(); [javac]^ [javac] /home/aaron/src/ext/svn/hive-0.3.0/build/ql/java/org/apache/hadoop/hive/ql/io/HiveInputFormat.java:241: cannot find symbol [javac] symbol : method validateInput(org.apache.hadoop.mapred.JobConf) [javac] location: interface org.apache.hadoop.mapred.InputFormat [javac] inputFormat.validateInput(newjob); [javac] ^ [javac] Note: Some input files use or override a deprecated API. [javac] Note: Recompile with -Xlint:deprecation for details. [javac] Note: Some input files use unchecked or unsafe operations. [javac] Note: Recompile with -Xlint:unchecked for details. [javac] 2 errors BUILD FAILED /home/aaron/src/ext/svn/hive-0.3.0/build.xml:145: The following error occurred while executing this line: /home/aaron/src/ext/svn/hive-0.3.0/ql/build.xml:135: Compile failed; see the compiler error output for details. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-666) We should not ignore exceptions in FetchTask
[ https://issues.apache.org/jira/browse/HIVE-666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-666: Attachment: HIVE-666.3.patch Modified QTestUtils.java to remove the temp table. Remove the line from the new .q file that drops the table. (It never gets executed because of the failure before it) We should not ignore exceptions in FetchTask Key: HIVE-666 URL: https://issues.apache.org/jira/browse/HIVE-666 Project: Hadoop Hive Issue Type: Bug Reporter: Zheng Shao Assignee: Zheng Shao Attachments: HIVE-666.1.patch, HIVE-666.2.patch, HIVE-666.3.patch FetchTask is ignoring some of the exceptions which makes it hard to debug the input file format code. We should propagate the exception back instead of ignoring it. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Commented: (HIVE-454) Support escaping of ; in strings in cli
[ https://issues.apache.org/jira/browse/HIVE-454?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=12733976#action_12733976 ] Edward Capriolo commented on HIVE-454: -- That is a good point. However it seems that the CLIDriver will be have to change regardless. Since the CLIDriver is specifically splitting on ';' a statement like set a=5; set b=8 gets passed as two separate qp.run()'s. Support escaping of ; in strings in cli --- Key: HIVE-454 URL: https://issues.apache.org/jira/browse/HIVE-454 Project: Hadoop Hive Issue Type: Bug Components: Clients Affects Versions: 0.3.0 Reporter: Ashish Thusoo Assignee: Ashish Thusoo Attachments: hive-454.patch If ; appears in string literals in a query the hive cli is not able to escape them properly. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-670) Explain should show output column names
[ https://issues.apache.org/jira/browse/HIVE-670?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zheng Shao updated HIVE-670: Attachment: HIVE-670.1.patch This patch adds outputColumnNames to select, group by, and join operators. Explain should show output column names --- Key: HIVE-670 URL: https://issues.apache.org/jira/browse/HIVE-670 Project: Hadoop Hive Issue Type: Improvement Components: Query Processor Reporter: Zheng Shao Attachments: HIVE-670.1.patch Explain currently only shows the expressions (in which it references the output column names of the last operator). However, it does not show the output column names of the last operator, which makes it hard to debug. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-665) Allow to query hadoop/mapreduce cluster status from hive server
[ https://issues.apache.org/jira/browse/HIVE-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Hwang updated HIVE-665: Status: Patch Available (was: Open) Allow to query hadoop/mapreduce cluster status from hive server --- Key: HIVE-665 URL: https://issues.apache.org/jira/browse/HIVE-665 Project: Hadoop Hive Issue Type: New Feature Components: Server Infrastructure Reporter: Hao Liu Attachments: HIVE-665.1.patch Tools/infra around hadoop/hive need to check cluster status in many cases. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.
[jira] Updated: (HIVE-665) Allow to query hadoop/mapreduce cluster status from hive server
[ https://issues.apache.org/jira/browse/HIVE-665?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Hwang updated HIVE-665: Attachment: HIVE-665.1.patch Allow to query hadoop/mapreduce cluster status from hive server --- Key: HIVE-665 URL: https://issues.apache.org/jira/browse/HIVE-665 Project: Hadoop Hive Issue Type: New Feature Components: Server Infrastructure Reporter: Hao Liu Attachments: HIVE-665.1.patch Tools/infra around hadoop/hive need to check cluster status in many cases. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.