How we will get mongo data into mongo table ?

By using this we can just create table

create external table mongo_users(id int, name string, age int)
stored by "org.yong3.hive.mongo.MongoStorageHandler"
with serdeproperties ( "mongo.column.mapping" = "_id,name,age" )
tblproperties ( "mongo.host" = "192.168.0.5", "mongo.port" = "11211",
"mongo.db" = "test", "mongo.user" = "testUser", "mongo.passwd" =
"testPasswd", "mongo.collection" = "users" );




On Thu, Sep 12, 2013 at 5:02 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:

> If you are importing from hive to mongo, why can't you just select from
> mongo table and insert into hive table?
>
>
> On Thu, Sep 12, 2013 at 4:24 PM, Sandeep Nemuri <nhsande...@gmail.com>wrote:
>
>> Hi Nitin Pawar,
>>                         I have checked That link .There data is
>> importing from Hive-mongo.
>> But my requirment is to import data from mongo-hive.
>>
>>
>> On Thu, Sep 12, 2013 at 3:49 PM, Nitin Pawar <nitinpawar...@gmail.com>wrote:
>>
>>> Sandip,
>>>
>>> Did you try using hive-mongo (https://github.com/yc-huang/Hive-mongo).
>>>
>>> Its pretty easy to use as well. If you want to start with analytics
>>> directly.
>>>
>>>
>>> On Thu, Sep 12, 2013 at 2:02 PM, Sandeep Nemuri <nhsande...@gmail.com>wrote:
>>>
>>>> Thanks all
>>>> i am trying to import data with this program
>>>> but when i compied this code i got errors
>>>>
>>>> Here is the code
>>>>
>>>> import java.io.*;
>>>> import org.apache.commons.logging.*;
>>>> import org.apache.hadoop.conf.*;
>>>> import org.apache.hadoop.fs.Path;
>>>> import org.apache.hadoop.io.*;
>>>> import org.apache.hadoop.mapreduce.lib.output.*;
>>>> import org.apache.hadoop.mapreduce.*;
>>>> import org.bson.*;
>>>> import com.mongodb.hadoop.*;
>>>> import com.mongodb.hadoop.util.*;
>>>>
>>>> public class ImportWeblogsFromMongo {
>>>>
>>>> private static final Log log = LogFactory.
>>>> getLog(ImportWeblogsFromMongo.class);
>>>>
>>>> public static class ReadWeblogsFromMongo extends Mapper<Object,
>>>> BSONObject, Text, Text>{
>>>>
>>>> public void map(Object key, BSONObject value, Context context) throws
>>>> IOException, InterruptedException{
>>>>
>>>> System.out.println("Key: " + key);
>>>> System.out.println("Value: " + value);
>>>>
>>>> String md5 = value.get("md5").toString();
>>>> String url = value.get("url").toString();
>>>> String date = value.get("date").toString();
>>>> String time = value.get("time").toString();
>>>> String ip = value.get("ip").toString();
>>>> String output = "\t" + url + "\t" + date + "\t" + time + "\t" + ip;
>>>>
>>>> context.write( new Text(md5), new Text(output));
>>>> }
>>>> }
>>>>
>>>> public static void main(String[] args) throws Exception{
>>>>
>>>> final Configuration conf = new Configuration();
>>>>
>>>> MongoConfigUtil.setInputURI(conf,"mongodb://localhost:27017/mongo_hadoop.example");
>>>>
>>>> MongoConfigUtil.setCreateInputSplits(conf, false);
>>>> System.out.println("Configuration: " + conf);
>>>>
>>>> final Job job = new Job(conf, "Mongo Import");
>>>> Path out = new Path("/user/mongo_data");
>>>> FileOutputFormat.setOutputPath(job, out);
>>>> job.setJarByClass(ImportWeblogsFromMongo.class);
>>>> job.setMapperClass(ReadWeblogsFromMongo.class);
>>>> job.setOutputKeyClass(Text.class);
>>>> job.setOutputValueClass(Text.class);
>>>> job.setInputFormatClass(MongoInputFormat.class);
>>>> job.setOutputFormatClass(TextOutputFormat.class);
>>>> job.setNumReduceTasks(0);
>>>> System.exit(job.waitForCompletion(true) ? 0 : 1 );
>>>> }
>>>> }
>>>>
>>>>
>>>>
>>>> On Wed, Sep 11, 2013 at 11:50 PM, Russell Jurney <
>>>> russell.jur...@gmail.com> wrote:
>>>>
>>>>> The docs are at
>>>>> https://github.com/mongodb/mongo-hadoop/tree/master/hive
>>>>>
>>>>> You need to build mongo-hadoop, and then use the documented syntax to
>>>>> create BSON tables in Hive.
>>>>>
>>>>>
>>>>> On Wed, Sep 11, 2013 at 11:11 AM, Jitendra Yadav <
>>>>> jeetuyadav200...@gmail.com> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> 1. you may use Hadoop-mongodb connector, create a map reduce program
>>>>>> to process your data from mongodb to hive.
>>>>>>
>>>>>> https://github.com/mongodb/mongo-hadoop
>>>>>>
>>>>>>
>>>>>> 2. As an alternative you can also use pig mongodb combination to get
>>>>>> the data from mongodb through pig, then after you can create a table
>>>>>> in hive that will points to the pig output file on hdfs.
>>>>>>
>>>>>> https://github.com/mongodb/mongo-hadoop/blob/master/pig/README.md
>>>>>>
>>>>>> Regards
>>>>>> Jitendra
>>>>>> On 9/11/13, Jérôme Verdier <verdier.jerom...@gmail.com> wrote:
>>>>>> > Hi,
>>>>>> >
>>>>>> > You can use Talend to import data from mongodb to hive
>>>>>> >
>>>>>> > More informations here : http://www.talend.com/products/big-data
>>>>>> >
>>>>>> >
>>>>>> > 2013/9/11 Sandeep Nemuri <nhsande...@gmail.com>
>>>>>> >
>>>>>> >> Hi every one ,
>>>>>> >>                        I am trying to import data from mongodb to
>>>>>> hive .
>>>>>> >> i
>>>>>> >> got some jar files to connect mongo and hive .
>>>>>> >> now how to import the data from mongodb to hive ?
>>>>>> >>
>>>>>> >> Thanks in advance.
>>>>>> >>
>>>>>> >> --
>>>>>> >> --Regards
>>>>>> >>   Sandeep Nemuri
>>>>>> >>
>>>>>> >
>>>>>> >
>>>>>> >
>>>>>> > --
>>>>>> > *Jérôme VERDIER*
>>>>>> > 06.72.19.17.31
>>>>>> > verdier.jerom...@gmail.com
>>>>>> >
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Russell Jurney twitter.com/rjurney russell.jur...@gmail.com
>>>>> datasyndrome.com
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> --Regards
>>>>   Sandeep Nemuri
>>>>
>>>
>>>
>>>
>>> --
>>> Nitin Pawar
>>>
>>
>>
>>
>> --
>> --Regards
>>   Sandeep Nemuri
>>
>
>
>
> --
> Nitin Pawar
>



-- 
--Regards
  Sandeep Nemuri

Reply via email to