Re: Querying hive without providing schema during create table for JSON data

2015-09-10 Thread Muthu Pandi
You can use Apache Drill to query Json data without creating the schema but
you cant in Hive.

http://drill.apache.org/docs/querying-json-files/


*RegardsMuthupandi.K*

*muthupandi.strikingly.com *

 Think before you print.



On Thu, Sep 10, 2015 at 9:16 PM, Jeetendra G 
wrote:

> anybody here? can somebody guide me on this?
>
> On Thu, Sep 10, 2015 at 2:36 PM, Jeetendra G 
> wrote:
>
>> Hi All,
>>
>> I have JSON data Can I create table without schema? and query that data,
>> since all my data os JSON data so I was thinking there should be some way.
>>
>> Currently I am doing something like below, but I want to remove Schema
>> from this, because I have i ma having thousand of event type and all of
>> them i cant really create the schema, Highlighted part I want to remove and
>> let the hive decide the schema.
>>
>>
>> create external table if not exists impressions_data
>> (
>> *city string,*
>> *service string,*
>> *id int,*
>> *type string,*
>> *listings int,*
>> *visible_markers int,*
>> *visible_as_markers int,*
>> *city_id int,*
>> *location_type string,*
>> *polygon_id string,*
>> *polygon_name string,*
>> *BuyListViewCount string,*
>> *AutoSuggest string,*
>> *LocalityExpert string,*
>> *BuyerExperiment string,*
>> *SellerExperiment string,*
>> *PaidAdProducts string,*
>> *index_view_type string,*
>> *page_type string,*
>> *clid string,*
>> *device string,*
>> *results_by_profile string,*
>> *listing_index int,*
>> *huid string,*
>> *uid string,*
>> *sid int,*
>> *count int,*
>> *category string,*
>> *action string,*
>> *url string,*
>> *`timeStamp` bigint)*
>> partitioned by (`date` string)
>> ROW FORMAT SERDE 'org.openx.data.jsonserde.JsonSerDe'
>> LOCATION 'hdfs://
>> hadoop01.housing.com:8020/housing/events/validated/category=impressions/
>> ';
>>
>> MSCK REPAIR TABLE impressions_data;
>> SHOW PARTITIONS impressions_data;
>> select * from impressions_data limit 1;
>>
>
>


Re: java.io.FileNotFoundException: File does not exist (nexr-hive-udf-0.2-SNAPSHOT.jar)

2014-12-17 Thread Muthu Pandi
Hi Arthur

Have a look at here
https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-PermanentFunctions



*RegardsMuthupandi.K*

 Think before you print.



On Wed, Dec 17, 2014 at 2:53 PM, arthur.hk.c...@gmail.com 
arthur.hk.c...@gmail.com wrote:

 Hi,

 Please help!

 I am using *hiveserver2 *on HIVE 0.13 on Hadoop 2.4.1, also
 nexr-hive-udf-0.2-SNAPSHOT.jar

 I can run query from CLI, e.g.
 hive SELECT add_months(sysdate(), +12) FROM DUAL;
 Execution completed successfully
 MapredLocal task succeeded
 OK
 2015-12-17
 Time taken: 7.393 seconds, Fetched: 1 row(s)


 hive-site.xml (added)
  property
   namehive.aux.jars.path/name

 value$HIVE_HOME/nexr-hive-udf-0.2-SNAPSHOT.jar,$HIVE_HOME/csv-serde-1.1.2-0.11.0-all.jar/value
  /property

 hive-env.sh (added)
 export
 HIVE_AUX_JARS_PATH=$HIVE_HOME/lib/csv-serde-1.1.2-0.11.0-all.jar:$HIVE_HOME/lib/nexr-hive-udf-0.2-SNAPSHOT.jar


 However, if it is accessed via *hiveserver2, *I got the following error,
 please help.

 Regards
 Arthur




 14/12/17 16:47:52 WARN conf.Configuration:
 file:/tmp/hive_2014-12-17_16-47-51_096_5821374687950910377-1/-local-10003/jobconf.xml:an
 attempt to override final parameter:
 mapreduce.job.end-notification.max.attempts;  Ignoring.
 Execution log at:
 /tmp/hduser_20141217164747_80b15b85-7820-4e3a-88ea-afffa131ff5a.log
 java.io.FileNotFoundException: File does not exist:
 hdfs://mycluster/hadoop_data/hadoop_data/tmp/mapred/staging/hduser1962118853/.staging/job_local1962118853_0001/libjars/nexr-hive-udf-0.2-SNAPSHOT.jar
 at
 org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1128)
 at
 org.apache.hadoop.hdfs.DistributedFileSystem$17.doCall(DistributedFileSystem.java:1120)
 at
 org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
 at
 org.apache.hadoop.hdfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:1120)
 at
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:288)
 at
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.getFileStatus(ClientDistributedCacheManager.java:224)
 at
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestamps(ClientDistributedCacheManager.java:93)
 at
 org.apache.hadoop.mapreduce.filecache.ClientDistributedCacheManager.determineTimestampsAndCacheVisibilities(ClientDistributedCacheManager.java:57)
 at
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:265)
 at
 org.apache.hadoop.mapreduce.JobSubmitter.copyAndConfigureFiles(JobSubmitter.java:301)
 at
 org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:389)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
 at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1556)
 at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
 at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
 at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
 at java.security.AccessController.doPrivileged(Native Method)
 at javax.security.auth.Subject.doAs(Subject.java:415)
 at
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1556)
 at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
 at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
 at
 org.apache.hadoop.hive.ql.exec.mr.ExecDriver.execute(ExecDriver.java:420)
 at org.apache.hadoop.hive.ql.exec.mr.ExecDriver.main(ExecDriver.java:740)
 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)
 Job Submission failed with exception 'java.io.FileNotFoundException(File
 does not exist:
 hdfs://mycluster/hadoop_data/hadoop_data/tmp/mapred/staging/hduser1962118853/.staging/job_local1962118853_0001/libjars/nexr-hive-udf-0.2-SNAPSHOT.jar
 )'
 Execution failed with exit status: 1
 Obtaining error information




Re: Can't integrate Elasticsearch with Hive

2014-11-27 Thread Muthu Pandi
Hi Atul

 Check your typo tblproperties('es.resources' = 'employee/address')  need
to be es.resource as it rightly indicated in the Error.



*RegardsMuthupandi.K*

 Think before you print.



On Fri, Nov 28, 2014 at 3:43 AM, Edward Capriolo edlinuxg...@gmail.com
wrote:

 Because the elastic search storage handler is not part of hive you might
 be able to find more help asking the author of the storage handler. From
 the errors you are reporting it likely does not work and not tested against
 this version of hive.



Re: Can't integrate Elasticsearch with Hive

2014-11-27 Thread Muthu Pandi
How did you get elasticsearch jar??

Try to build the jar against your version and use that jar.



*RegardsMuthupandi.K*

 Think before you print.



On Fri, Nov 28, 2014 at 12:23 PM, Atul Paldhikar (apaldhik) 
apald...@cisco.com wrote:

  Thanks Muthu for pointing out the issue.



 However unfortunately I am back to the same old issue while accessing the
 external table. I am getting the following exception again, there wasn’t
 any change in the environment.



 I tried all the cleanup but still no luck :-( Here are the steps I tried



 1. Removed the other jar for es-hadoop 2.1.x completely from the server



 2. Updated the hive-site.xml as below, but didn't work



 property

   namehive.aux.jars.path/name


 value/apps/sas/elasticsearch-hadoop-2.0.2/dist/elasticsearch-hadoop-2.0.2.jar/value

   descriptionA comma separated list (with no spaces) of the jar
 files/description

 /property



 3. Added the jar file in the hiveconf as below, still the same issue



 hive --hiveconf
 hive.aux.jars.path=/apps/sas/elasticsearch-hadoop-2.0.2/dist/elasticsearch-hadoop-2.0.2.jar



 4. Tried adding the jar file in the hive session, still didn't work



 add jar
 /apps/sas/elasticsearch-hadoop-2.0.2/dist/elasticsearch-hadoop-2.0.2.jar;



 Thanks

 - Atul



 ERROR:



 2014-11-26 23:09:22,069 ERROR [main]: exec.DDLTask
 (DDLTask.java:execute(478)) - java.lang.IllegalAccessError: tried to access
 class org.elasticsearch.hadoop.hive.HiveUtils from class
 org.elasticsearch.hadoop.hive.EsSerDe

 at
 org.elasticsearch.hadoop.hive.EsSerDe.initialize(EsSerDe.java:81)

 at
 org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:339)

 at
 org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:288)

 at
 org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:281)

 at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:631)

 at
 org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:593)

 at
 org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4189)

 at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:281)

 at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:153)

 at
 org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)

 at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1503)

 at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1270)

 at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1088)

 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:911)

 at org.apache.hadoop.hive.ql.Driver.run(Driver.java:901)

 at
 org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:268)

 at
 org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:220)

 at
 org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:423)

 at
 org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:792)

 at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:686)

 at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)

 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)

 at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)

 at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)

 at java.lang.reflect.Method.invoke(Method.java:606)

 at org.apache.hadoop.util.RunJar.main(RunJar.java:212)



 2014-11-26 23:09:22,069 ERROR [main]: ql.Driver
 (SessionState.java:printError(545)) - FAILED: Execution Error, return code
 1 from org.apache.hadoop.hive.ql.exec.DDLTask. tried to access class
 org.elasticsearch.hadoop.hive.HiveUtils from class
 org.elasticsearch.hadoop.hive.EsSerDe



 *From:* Muthu Pandi [mailto:muthu1...@gmail.com]
 *Sent:* Thursday, November 27, 2014 9:08 PM
 *To:* user@hive.apache.org
 *Subject:* Re: Can't integrate Elasticsearch with Hive



 Hi Atul



  Check your typo tblproperties('es.resources' = 'employee/address')
  need to be es.resource as it rightly indicated in the Error.



 *Regards Muthupandi.K*

 *[image: Image removed by sender.]* Think before you print.





 On Fri, Nov 28, 2014 at 3:43 AM, Edward Capriolo edlinuxg...@gmail.com
 wrote:

 Because the elastic search storage handler is not part of hive you might
 be able to find more help asking the author of the storage handler. From
 the errors you are reporting it likely does not work and not tested against
 this version of hive.





Re: Unix script for identifying current active namenode in a HA cluster

2014-11-04 Thread Muthu Pandi
Good work Devopam Mittra.



*RegardsMuthupandi.K*

 Think before you print.



On Wed, Nov 5, 2014 at 12:31 PM, Nitin Pawar nitinpawar...@gmail.com
wrote:

 +1
 If you can optionally add hadoop home directory in the script and use that
 in path, it can be used out of the box.

 Also can you share this on github

 On Wed, Nov 5, 2014 at 10:02 AM, Devopam Mittra devo...@gmail.com wrote:

 hi All,
 Please find attached a simple shell script to dynamically determine the
 active namenode in the HA Cluster and subsequently run the Hive job / query
 via Talend OS generated workflows.

 It was tried successfully on a HDP2.1 cluster with 2 nn, 7 dn running on
 CentOS 6.5.
 Each ETL job invokes this script first in our framework to derive the NN
 FQDN and then run the hive jobs subsequently to avoid failures.
 Takes a max. of 2 secs to execute (small cost in our case, as compared to
 dealing with a failure and then recalculating the NN to resubmit the job).

 Sharing it with you in case you can leverage the same without spending
 effort to code it.

 Do share your feedback/ fixes if you spot any.

 --
 Devopam Mittra
 Life and Relations are not binary




 --
 Nitin Pawar



Re: oozie installation error

2014-09-25 Thread Muthu Pandi
Hi Rahul

Distro error may occur while using this command  bin/mkdistro.sh
-DskipTests instead of that use

mvn clean package assembly:single -P hadoop-2 -DskipTests

if you are building against hadoop 2

it resolved the distro error.

For more info follow http://gauravkohli.com/category/oozie/;



*RegardsMuthupandi.K*

 Think before you print.



On Wed, Sep 24, 2014 at 11:22 PM, Rahul Channe drah...@googlemail.com
wrote:

 hi All,

 I am trying to install oozie and getting following error. Any input is
 appreciated

 [INFO] BUILD FAILURE
 [INFO]
 
 [INFO] Total time: 18.326s
 [INFO] Finished at: Wed Sep 24 13:45:13 EDT 2014
 [INFO] Final Memory: 26M/64M
 [INFO]
 
 [ERROR] Failed to execute goal
 org.apache.maven.plugins:maven-jar-plugin:2.3.1:jar (default-jar) on
 project oozie-client: Error assembling JAR: Failed to read filesystem
 attributes for: /home/user/oozie/client/pom.xml: Failed to retrieve numeric
 file attributes using: '/bin/sh -c ls -1nlad
 /home/user/oozie/client/pom.xml': Error while executing process. Cannot run
 program /bin/sh: java.io.IOException: error=12, Cannot allocate memory -
 [Help 1]
 [ERROR]
 [ERROR] To see the full stack trace of the errors, re-run Maven with the
 -e switch.
 [ERROR] Re-run Maven using the -X switch to enable full debug logging.
 [ERROR]
 [ERROR] For more information about the errors and possible solutions,
 please read the following articles:
 [ERROR] [Help 1]
 http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
 [ERROR]
 [ERROR] After correcting the problems, you can resume the build with the
 command
 [ERROR]   mvn goals -rf :oozie-client

 ERROR, Oozie distro creation failed




Re: oozie installation error

2014-09-25 Thread Muthu Pandi
replace hadoop-2 with hadoop-1



*RegardsMuthupandi.K*

 Think before you print.



On Thu, Sep 25, 2014 at 4:48 PM, Rahul Channe drah...@googlemail.com
wrote:

 Hi Muthu,

 I am trying to build oozie against hadoop 1


 On Thursday, September 25, 2014, Muthu Pandi muthu1...@gmail.com wrote:

 Hi Rahul

 Distro error may occur while using this command  bin/mkdistro.sh
 -DskipTests instead of that use

 mvn clean package assembly:single -P hadoop-2 -DskipTests

 if you are building against hadoop 2

 it resolved the distro error.

 For more info follow http://gauravkohli.com/category/oozie/;



 *RegardsMuthupandi.K*

  Think before you print.



 On Wed, Sep 24, 2014 at 11:22 PM, Rahul Channe drah...@googlemail.com
 wrote:

 hi All,

 I am trying to install oozie and getting following error. Any input is
 appreciated

 [INFO] BUILD FAILURE
 [INFO]
 
 [INFO] Total time: 18.326s
 [INFO] Finished at: Wed Sep 24 13:45:13 EDT 2014
 [INFO] Final Memory: 26M/64M
 [INFO]
 
 [ERROR] Failed to execute goal
 org.apache.maven.plugins:maven-jar-plugin:2.3.1:jar (default-jar) on
 project oozie-client: Error assembling JAR: Failed to read filesystem
 attributes for: /home/user/oozie/client/pom.xml: Failed to retrieve numeric
 file attributes using: '/bin/sh -c ls -1nlad
 /home/user/oozie/client/pom.xml': Error while executing process. Cannot run
 program /bin/sh: java.io.IOException: error=12, Cannot allocate memory -
 [Help 1]
 [ERROR]
 [ERROR] To see the full stack trace of the errors, re-run Maven with the
 -e switch.
 [ERROR] Re-run Maven using the -X switch to enable full debug logging.
 [ERROR]
 [ERROR] For more information about the errors and possible solutions,
 please read the following articles:
 [ERROR] [Help 1]
 http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
 [ERROR]
 [ERROR] After correcting the problems, you can resume the build with the
 command
 [ERROR]   mvn goals -rf :oozie-client

 ERROR, Oozie distro creation failed





Re: Mysql - Hive Sync

2014-09-06 Thread Muthu Pandi
Yeah installing Mysql hadoop applier took lot of time when building and
installing GCC 4.6, and its working but its not serving the exact purpose.
So now am trying with my own python scripting.

Idea is reading insert query from binlog and save it under hive warehouse
as table and query from there.



*RegardsMuthupandi.K*

[image: Picture (Device Independent Bitmap)]



On Sat, Sep 6, 2014 at 4:47 AM, Stephen Sprague sprag...@gmail.com wrote:

 great find, Muthu.  I would be interested in hearing any about any success
 or failures using this adapter. almost sounds too good to be true.

 After reading the blog (
 http://innovating-technology.blogspot.com/2013/04/mysql-hadoop-applier-part-2.html)
 about it i see it comes with caveats and it looks a little rough around the
 edges for installing.  Not sure i'd bet the farm on this product but YMMV.

 Anyway, curious to know how it works out for you.



 On Tue, Sep 2, 2014 at 11:03 PM, Muthu Pandi muthu1...@gmail.com wrote:

 This cant be done since insert update delete are not supported in hive.

 Mysql Applier for Hadoop package servers the same purpose of the
 prototype tool which i intended to develop.

 link for Mysql Applier for Hadoop
 http://dev.mysql.com/tech-resources/articles/mysql-hadoop-applier.html



 *Regards Muthupandi.K*

  [image: Picture (Device Independent Bitmap)]



 On Wed, Sep 3, 2014 at 10:35 AM, Muthu Pandi muthu1...@gmail.com wrote:

 Yeah but we cant make it to work as near real time. Also my table doesnt
 have like 'ID' to use for --check-column that's why opted out of sqoop.



 *Regards Muthupandi.K*

  [image: Picture (Device Independent Bitmap)]



 On Wed, Sep 3, 2014 at 10:28 AM, Nitin Pawar nitinpawar...@gmail.com
 wrote:

 have you looked at sqoop?


 On Wed, Sep 3, 2014 at 10:15 AM, Muthu Pandi muthu1...@gmail.com
 wrote:

 Dear All

  Am developing a prototype of syncing tables from mysql to Hive
 using python and JDBC. Is it a good idea using the JDBC for this purpose.

 My usecase will be generating the sales report using the hive, data
 pulled from mysql using the prototype tool.My data will be around 2GB/day.



 *Regards Muthupandi.K*

  [image: Picture (Device Independent Bitmap)]




 --
 Nitin Pawar







Re: Mysql - Hive Sync

2014-09-03 Thread Muthu Pandi
This cant be done since insert update delete are not supported in hive.

Mysql Applier for Hadoop package servers the same purpose of the prototype
tool which i intended to develop.

link for Mysql Applier for Hadoop
http://dev.mysql.com/tech-resources/articles/mysql-hadoop-applier.html



*Regards Muthupandi.K*

 [image: Picture (Device Independent Bitmap)]



On Wed, Sep 3, 2014 at 10:35 AM, Muthu Pandi muthu1...@gmail.com wrote:

 Yeah but we cant make it to work as near real time. Also my table doesnt
 have like 'ID' to use for --check-column that's why opted out of sqoop.



 *Regards Muthupandi.K*

  [image: Picture (Device Independent Bitmap)]



 On Wed, Sep 3, 2014 at 10:28 AM, Nitin Pawar nitinpawar...@gmail.com
 wrote:

 have you looked at sqoop?


 On Wed, Sep 3, 2014 at 10:15 AM, Muthu Pandi muthu1...@gmail.com wrote:

 Dear All

  Am developing a prototype of syncing tables from mysql to Hive
 using python and JDBC. Is it a good idea using the JDBC for this purpose.

 My usecase will be generating the sales report using the hive, data
 pulled from mysql using the prototype tool.My data will be around 2GB/day.



 *Regards Muthupandi.K*

  [image: Picture (Device Independent Bitmap)]




 --
 Nitin Pawar





Mysql - Hive Sync

2014-09-02 Thread Muthu Pandi
Dear All

 Am developing a prototype of syncing tables from mysql to Hive using
python and JDBC. Is it a good idea using the JDBC for this purpose.

My usecase will be generating the sales report using the hive, data pulled
from mysql using the prototype tool.My data will be around 2GB/day.



*Regards Muthupandi.K*

 [image: Picture (Device Independent Bitmap)]


Re: Basic Hadoop 2.3 32-bit VM for general Hadoop Users

2014-08-22 Thread Muthu Pandi
Excellent initiative All the Best for the startups



*Regards Muthupandi.K*

 [image: Picture (Device Independent Bitmap)]



On Fri, Aug 22, 2014 at 11:38 PM, Support Team supp...@lighthadoop.com
wrote:

 We have released a very basic 32-bit VM (VirtualBox Image) for those users
 who want to get started with Hadoop, without worrying about configuration
 and dependencies.

 We have used CDH5.1 for this release which contains Hadoop 2.3 (YARN), Pig
 0.12, Hive 0.12, Sqoop 1.4.4 along with MySQL with 814MB of download size.

 We have also packaged a simple use case of Wiki Page Hits analysis, which
 is explained in our blog here http://www.lighthadoop.com/?p=1 at
 www.lighthadoop.com.

 This is a genuine effort to help adoption of Hadoop and its eco-system
 fast, especially for students, from our freelancing big data enthusiasts to
 get started with latest Hadoop, Pig and Hive. Thus reducing time and effort
 in installating and configuring the system, still keeping the Hardware
 requirement low.

 The motivation behind this VM is there are users who own 32-bit
 systems(can address 4GB RAM, thats enough for basic Hadoop setup) and still
 want to try latest stable Hadoop. Thus enabling them to solve a use case
 without needing to buy a latest PC/Laptop with high amount of RAM.

 Kindly send your feedback/suggestions to supp...@lighthadoop.com
 http://mailto:supp...@lighthadoop.com .

 All suggestions welcome! Suggestions make us grow, thus serving more
 opensource community!

 Thanks!
 LightHadoop Team