Re: REST API for Hive queries?

2012-12-13 Thread Manish Malhotra
Ideally, push the aggregated data to some RDBMS like MySQL and have REST
API or some API to enable ui to build report or query out of it.

If the use case is ad-hoc query then once that qry is submitted, and result
is generated in batch mode, the REST API can be provided to get the results
from HDFS directly.
For this can use WebHDFS or build own which can internally using FileSystem
API.

Regards,
Manish


On Wed, Dec 12, 2012 at 11:30 PM, Nitin Pawar nitinpawar...@gmail.comwrote:

 Hive takes a longer time to respond to queries as the data gets larger.

 Best way to handle this is you process the data on hive and store in some
 rdbms like mysql etc.
 On top of that then you can write your own API or use pentaho like
 interface where they can write the queries or see predefined reports.

 Alternatively, pentaho does have hive connection as well. There are other
 platforms such as talend, datameer etc. You can have a look at them


 On Thu, Dec 13, 2012 at 1:15 AM, Leena Gupta gupta.le...@gmail.comwrote:

 Hi,

 We are using Hive as our data warehouse to run various queries on large
 amounts of data. There are some users who would like to get access to the
 output of these queries and display the data on an existing UI application.
 What is the best way to give them the output of these queries? Should we
 write REST APIs that the Front end can call to get the data? How can this
 be done?
  I'd like to know what have other people done to meet this requirement ?
 Any pointers would be very helpful.
 Thanks.




 --
 Nitin Pawar



Re: REST API for Hive queries?

2012-12-13 Thread Jagat Singh
If your requirement is that queries are not going to be run on fly then i
would suggest following.

1) Create Hive script
2) Combine it with Oozie workflow to run at scheduled time and push results
to some DB say MySQL
3) Use some application to talk to MySQL and generate those reports.

Thanks,

Jagat Singh



On Thu, Dec 13, 2012 at 7:15 PM, Manish Malhotra 
manish.hadoop.w...@gmail.com wrote:


 Ideally, push the aggregated data to some RDBMS like MySQL and have REST
 API or some API to enable ui to build report or query out of it.

 If the use case is ad-hoc query then once that qry is submitted, and
 result is generated in batch mode, the REST API can be provided to get the
 results from HDFS directly.
 For this can use WebHDFS or build own which can internally using
 FileSystem API.

 Regards,
 Manish


 On Wed, Dec 12, 2012 at 11:30 PM, Nitin Pawar nitinpawar...@gmail.comwrote:

 Hive takes a longer time to respond to queries as the data gets larger.

 Best way to handle this is you process the data on hive and store in some
 rdbms like mysql etc.
 On top of that then you can write your own API or use pentaho like
 interface where they can write the queries or see predefined reports.

 Alternatively, pentaho does have hive connection as well. There are other
 platforms such as talend, datameer etc. You can have a look at them


  On Thu, Dec 13, 2012 at 1:15 AM, Leena Gupta gupta.le...@gmail.comwrote:

 Hi,

 We are using Hive as our data warehouse to run various queries on large
 amounts of data. There are some users who would like to get access to the
 output of these queries and display the data on an existing UI application.
 What is the best way to give them the output of these queries? Should we
 write REST APIs that the Front end can call to get the data? How can this
 be done?
  I'd like to know what have other people done to meet this requirement ?
 Any pointers would be very helpful.
 Thanks.




 --
 Nitin Pawar