Can you share your create table ddl for table name docs?

Select statement does not need all those details. Those are part of create
table DDL only.


On Fri, Sep 13, 2013 at 4:24 PM, Sandeep Nemuri <nhsande...@gmail.com>wrote:

> Hi nithin
>
> Thanks for your help
> I have used this query in hive to retrieve the data from mongodb
>
> add jar /usr/lib/hadoop/lib/mongo-2.8.0.jar;
> add jar /usr/lib/hive/lib/hive-mongo-0.0.3-jar-with-dependencies.jar;
>
> select * from docs
> input format "org.yong3.hive.mongo.MongoStorageHandler"
> with serdeproperties ( "mongo.column.mapping" =
> "_id,dayOfWeek,bc3Year,bc5Year,bc10Year,bc20Year,bc1Month,bc2Year,bc3Year,bc30Year,bc1Year,bc7Year,bc6Year"
> )
> tblproperties ( "mongo.host" = "127.0.0.1", "mongo.port" = "27017",
> "mongo.db" = "sample", "mongo.user" = "sample", "mongo.passwd" =
> "password", "mongo.collection" = "docs" );
>
>
> I got an Error
>
> FAILED: Parse Error: line 2:6 mismatched input 'format' expecting EOF near
> 'input'
>
>
>
> On Thu, Sep 12, 2013 at 6:23 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>
>> try creating table with your existing mongo db and collection see the
>> data can be read by the user or not.
>> What you need to do is mongo collection column mapping exactly with same
>> names into hive column definition.
>>
>> if you can not see mongo data from hive query, do let me know what errors
>> do you see.
>>
>>
>> On Thu, Sep 12, 2013 at 5:28 PM, Sandeep Nemuri <nhsande...@gmail.com>wrote:
>>
>>> How we will get mongo data into mongo table ?
>>>
>>> By using this we can just create table
>>>
>>> create external table mongo_users(id int, name string, age int)
>>> stored by "org.yong3.hive.mongo.MongoStorageHandler"
>>> with serdeproperties ( "mongo.column.mapping" = "_id,name,age" )
>>> tblproperties ( "mongo.host" = "192.168.0.5", "mongo.port" = "11211",
>>> "mongo.db" = "test", "mongo.user" = "testUser", "mongo.passwd" = 
>>> "testPasswd", "mongo.collection" = "users" );
>>>
>>>
>>>
>>>
>>> On Thu, Sep 12, 2013 at 5:02 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>>>
>>>> If you are importing from hive to mongo, why can't you just select from
>>>> mongo table and insert into hive table?
>>>>
>>>>
>>>> On Thu, Sep 12, 2013 at 4:24 PM, Sandeep Nemuri 
>>>> <nhsande...@gmail.com>wrote:
>>>>
>>>>> Hi Nitin Pawar,
>>>>>                         I have checked That link .There data is
>>>>> importing from Hive-mongo.
>>>>> But my requirment is to import data from mongo-hive.
>>>>>
>>>>>
>>>>> On Thu, Sep 12, 2013 at 3:49 PM, Nitin Pawar 
>>>>> <nitinpawar...@gmail.com>wrote:
>>>>>
>>>>>> Sandip,
>>>>>>
>>>>>> Did you try using hive-mongo (https://github.com/yc-huang/Hive-mongo
>>>>>> ).
>>>>>>
>>>>>> Its pretty easy to use as well. If you want to start with analytics
>>>>>> directly.
>>>>>>
>>>>>>
>>>>>> On Thu, Sep 12, 2013 at 2:02 PM, Sandeep Nemuri <nhsande...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> Thanks all
>>>>>>> i am trying to import data with this program
>>>>>>> but when i compied this code i got errors
>>>>>>>
>>>>>>> Here is the code
>>>>>>>
>>>>>>> import java.io.*;
>>>>>>> import org.apache.commons.logging.*;
>>>>>>> import org.apache.hadoop.conf.*;
>>>>>>> import org.apache.hadoop.fs.Path;
>>>>>>> import org.apache.hadoop.io.*;
>>>>>>> import org.apache.hadoop.mapreduce.lib.output.*;
>>>>>>> import org.apache.hadoop.mapreduce.*;
>>>>>>> import org.bson.*;
>>>>>>> import com.mongodb.hadoop.*;
>>>>>>> import com.mongodb.hadoop.util.*;
>>>>>>>
>>>>>>> public class ImportWeblogsFromMongo {
>>>>>>>
>>>>>>> private static final Log log = LogFactory.
>>>>>>> getLog(ImportWeblogsFromMongo.class);
>>>>>>>
>>>>>>> public static class ReadWeblogsFromMongo extends Mapper<Object,
>>>>>>> BSONObject, Text, Text>{
>>>>>>>
>>>>>>> public void map(Object key, BSONObject value, Context context)
>>>>>>> throws IOException, InterruptedException{
>>>>>>>
>>>>>>> System.out.println("Key: " + key);
>>>>>>> System.out.println("Value: " + value);
>>>>>>>
>>>>>>> String md5 = value.get("md5").toString();
>>>>>>> String url = value.get("url").toString();
>>>>>>> String date = value.get("date").toString();
>>>>>>> String time = value.get("time").toString();
>>>>>>> String ip = value.get("ip").toString();
>>>>>>> String output = "\t" + url + "\t" + date + "\t" + time + "\t" + ip;
>>>>>>>
>>>>>>> context.write( new Text(md5), new Text(output));
>>>>>>> }
>>>>>>> }
>>>>>>>
>>>>>>> public static void main(String[] args) throws Exception{
>>>>>>>
>>>>>>> final Configuration conf = new Configuration();
>>>>>>>
>>>>>>> MongoConfigUtil.setInputURI(conf,"mongodb://localhost:27017/mongo_hadoop.example");
>>>>>>>
>>>>>>> MongoConfigUtil.setCreateInputSplits(conf, false);
>>>>>>> System.out.println("Configuration: " + conf);
>>>>>>>
>>>>>>> final Job job = new Job(conf, "Mongo Import");
>>>>>>> Path out = new Path("/user/mongo_data");
>>>>>>> FileOutputFormat.setOutputPath(job, out);
>>>>>>> job.setJarByClass(ImportWeblogsFromMongo.class);
>>>>>>> job.setMapperClass(ReadWeblogsFromMongo.class);
>>>>>>> job.setOutputKeyClass(Text.class);
>>>>>>> job.setOutputValueClass(Text.class);
>>>>>>> job.setInputFormatClass(MongoInputFormat.class);
>>>>>>> job.setOutputFormatClass(TextOutputFormat.class);
>>>>>>> job.setNumReduceTasks(0);
>>>>>>> System.exit(job.waitForCompletion(true) ? 0 : 1 );
>>>>>>> }
>>>>>>> }
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Sep 11, 2013 at 11:50 PM, Russell Jurney <
>>>>>>> russell.jur...@gmail.com> wrote:
>>>>>>>
>>>>>>>> The docs are at
>>>>>>>> https://github.com/mongodb/mongo-hadoop/tree/master/hive
>>>>>>>>
>>>>>>>> You need to build mongo-hadoop, and then use the documented syntax
>>>>>>>> to create BSON tables in Hive.
>>>>>>>>
>>>>>>>>
>>>>>>>> On Wed, Sep 11, 2013 at 11:11 AM, Jitendra Yadav <
>>>>>>>> jeetuyadav200...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> 1. you may use Hadoop-mongodb connector, create a map reduce
>>>>>>>>> program
>>>>>>>>> to process your data from mongodb to hive.
>>>>>>>>>
>>>>>>>>> https://github.com/mongodb/mongo-hadoop
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2. As an alternative you can also use pig mongodb combination to
>>>>>>>>> get
>>>>>>>>> the data from mongodb through pig, then after you can create a
>>>>>>>>> table
>>>>>>>>> in hive that will points to the pig output file on hdfs.
>>>>>>>>>
>>>>>>>>> https://github.com/mongodb/mongo-hadoop/blob/master/pig/README.md
>>>>>>>>>
>>>>>>>>> Regards
>>>>>>>>> Jitendra
>>>>>>>>> On 9/11/13, Jérôme Verdier <verdier.jerom...@gmail.com> wrote:
>>>>>>>>> > Hi,
>>>>>>>>> >
>>>>>>>>> > You can use Talend to import data from mongodb to hive
>>>>>>>>> >
>>>>>>>>> > More informations here : http://www.talend.com/products/big-data
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> > 2013/9/11 Sandeep Nemuri <nhsande...@gmail.com>
>>>>>>>>> >
>>>>>>>>> >> Hi every one ,
>>>>>>>>> >>                        I am trying to import data from mongodb
>>>>>>>>> to hive .
>>>>>>>>> >> i
>>>>>>>>> >> got some jar files to connect mongo and hive .
>>>>>>>>> >> now how to import the data from mongodb to hive ?
>>>>>>>>> >>
>>>>>>>>> >> Thanks in advance.
>>>>>>>>> >>
>>>>>>>>> >> --
>>>>>>>>> >> --Regards
>>>>>>>>> >>   Sandeep Nemuri
>>>>>>>>> >>
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> >
>>>>>>>>> > --
>>>>>>>>> > *Jérôme VERDIER*
>>>>>>>>> > 06.72.19.17.31
>>>>>>>>> > verdier.jerom...@gmail.com
>>>>>>>>> >
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>>>>> datasyndrome.com
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> --Regards
>>>>>>>   Sandeep Nemuri
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Nitin Pawar
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> --Regards
>>>>>   Sandeep Nemuri
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Nitin Pawar
>>>>
>>>
>>>
>>>
>>> --
>>> --Regards
>>>   Sandeep Nemuri
>>>
>>
>>
>>
>> --
>> Nitin Pawar
>>
>
>
>
> --
> --Regards
>   Sandeep Nemuri
>



-- 
Nitin Pawar

Reply via email to