I have a situation where i have seven bolts and one spout & i want to
distribute the tuples according to the field ID.
For eg. if ID=21 I want the tuple to be processed by first bolt
              ID=31 I want that tuple to be processed by second bolt & so
on.

So is there a way to implement these. I was thinking about using fields
grouping but in that i can only define the field name but not the value of
that field, So if i use field grouping i don't think there would be a
guarantee that suppose for ID=21 the tuple would be processed by first bolt.
Kindly correct me if i'm wrong about field grouping & provide solution to
implement these kind of topology.
Thanks in advance.


On Fri, Aug 1, 2014 at 10:20 PM, amjad khan <amjadkhan987...@gmail.com>
wrote:

> My bolt tries to write data to hdfs but the whole data is not written it
> throws exception
>
> org.apache.hadoop.ipc.RemoteException: 
> org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on 
> /storm.txt File does not exist. Holder DFSClient_attempt_storm.txt does not 
> have any open files.
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1557)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1548
>
> Kindly help me if anyone has any idea about this.
>
>
>
> On Sat, Jul 26, 2014 at 12:47 PM, amjad khan <amjadkhan987...@gmail.com>
> wrote:
>
>> Output when using bolt that tries to write its data to hdfs.
>>
>> INFO org.apache.hadoop,ipc.Client - Retrying Connect to Server: localhost/
>> 131.0.0.1:43785 Already tried 6 time(s).
>> WARN Caught URI Exception
>> java.net.ConnectException Call to localhost/131.0.0.1:43785 Failed on
>> Connect Exception: java.net.ConnectException: Connection Refused
>>
>> IN MY Code:
>> Configuration config = new Configuration();
>> config.set("fs.defaultFS","hdfs://localhost:9000");
>> FileSystem fs = FileSystem.get(config);
>>
>>
>> /etc/hosts contain
>> 181.45.83.79 localhost
>>
>> core-site contain
>>
>> <name>fs.default.name</name>
>> <value>hdfs://localhost:9000</value>
>>
>>
>> Kindly tell me why it is trying to connect on 131.0.0.1 & why at port
>> 43785 .
>>
>>
>> The same code is working fine in java without implementing it in storm &
>> i'm using hadoop 1.0.2.
>>
>>
>> On Fri, Jul 18, 2014 at 11:33 AM, Parth Brahmbhatt <
>> pbrahmbh...@hortonworks.com> wrote:
>>
>>> Hi Amjad,
>>>
>>> Is there any reason you can not upgrade to hadoop 2.0? Hadoop 2.0 has
>>> made many improvements over 1.X versions and they are source compatible so
>>> any of your MR jobs will be unaffected as long as you recompile with 2.x.
>>>
>>> The code we pointed at assumes that all the classes for hadoop 2.X are
>>> present in your class path. if you are not using maven or some other build
>>> system and would like to add jars manually you probably will have tough
>>> time resolving conflicts so I would advise against it.
>>> If you still want to add jars manually my best guess would be to look
>>> under
>>> <YOUR_HADOO_INSTALLATION_DIR>/libexec/share/hadoop/
>>>
>>> Thanks
>>> Parth
>>> On Jul 18, 2014, at 10:56 AM, amjad khan <amjadkhan987...@gmail.com>
>>> wrote:
>>>
>>> Thanks for your reply taylor. I'm using hadoop1.0.2. Can u suggest me
>>> any alternative to connect to hadoop.
>>>
>>>
>>>
>>> On Fri, Jul 18, 2014 at 8:45 AM, P. Taylor Goetz <ptgo...@gmail.com>
>>> wrote:
>>>
>>>> What version of Hadoop are you using? Storm-hdfs requires Hadoop 2.x.
>>>>
>>>> - Taylor
>>>>
>>>> On Jul 18, 2014, at 6:07 AM, amjad khan <amjadkhan987...@gmail.com>
>>>> wrote:
>>>>
>>>> Thanks for your help parth
>>>>
>>>> When i trying to run the topology to write the data to hdfs it throws
>>>> exception Class Not Found:
>>>> org.apache.hadoop.client.hdfs.HDFSDataOutputStream$SyncFlags
>>>> Can anyone tell me what are the jars needed to execute the code to
>>>> write data to hdfs. Please tell me all the required jars.
>>>>
>>>>
>>>> On Wed, Jul 16, 2014 at 10:46 AM, Parth Brahmbhatt <
>>>> pbrahmbh...@hortonworks.com> wrote:
>>>>
>>>>> You can use
>>>>>
>>>>> https://github.com/ptgoetz/storm-hdfs
>>>>>
>>>>> It supports writing to HDFS with both Storm bolts and trident states.
>>>>> Thanks
>>>>> Parth
>>>>>
>>>>> On Jul 16, 2014, at 10:41 AM, amjad khan <amjadkhan987...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> Can anyone provide the code for bolt to write its data to hdfs. Kindly
>>>>> tell me the jar's required to run that bolt.
>>>>>
>>>>>
>>>>> On Mon, Jul 14, 2014 at 2:33 PM, Max Evers <mcev...@gmail.com> wrote:
>>>>>
>>>>>> Can you expand on your use case? What is the query selecting on? Is
>>>>>> the column you are querying on indexed?  Do you really need to look at 
>>>>>> the
>>>>>> entire 20 gb every 20ms?
>>>>>>  On Jul 14, 2014 6:39 AM, "amjad khan" <amjadkhan987...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I made a storm topoogy where spout was fetching data from mysql
>>>>>>> using select query. The select query was fired after every 30 msec but
>>>>>>> because the size of the table is more than 20 GB the select query takes
>>>>>>> more than 10 sec to execute therefore this is not working. I need to 
>>>>>>> know
>>>>>>> what are the possible alternatives for this situation. Kindly reply as 
>>>>>>> soon
>>>>>>> as possible.
>>>>>>>
>>>>>>> Thanks,
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> CONFIDENTIALITY NOTICE
>>>>> NOTICE: This message is intended for the use of the individual or
>>>>> entity to which it is addressed and may contain information that is
>>>>> confidential, privileged and exempt from disclosure under applicable law.
>>>>> If the reader of this message is not the intended recipient, you are 
>>>>> hereby
>>>>> notified that any printing, copying, dissemination, distribution,
>>>>> disclosure or forwarding of this communication is strictly prohibited. If
>>>>> you have received this communication in error, please contact the sender
>>>>> immediately and delete it from your system. Thank You.
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>>> CONFIDENTIALITY NOTICE
>>> NOTICE: This message is intended for the use of the individual or entity
>>> to which it is addressed and may contain information that is confidential,
>>> privileged and exempt from disclosure under applicable law. If the reader
>>> of this message is not the intended recipient, you are hereby notified that
>>> any printing, copying, dissemination, distribution, disclosure or
>>> forwarding of this communication is strictly prohibited. If you have
>>> received this communication in error, please contact the sender immediately
>>> and delete it from your system. Thank You.
>>>
>>
>>
>

Reply via email to