Thanks Tyson!
This blog is super helpful.
I've been able to get LoggingMetrics working to an extent, however if I try
to create multiple CountMetrics in the same function, I only see one show
up in my NimbusUI. Does anybody know why this is ?


On Thu, May 29, 2014 at 8:57 AM, Tyson Norris <tnor...@adobe.com> wrote:

>  I found this blog helpful:
> http://www.bigdata-cookbook.com/post/72320512609/storm-metrics-how-to
>
>  Best regards,
> Tyson
>
>  On May 29, 2014, at 8:41 AM, Raphael Hsieh <raffihs...@gmail.com> wrote:
>
>  Can someone explain to me what LoggingMetrics is ?
> I've heard of it and people have told me to use it, but I can't find any
> documentation on it or any resources on how to use it.
>
> Thanks
>
>
> On Thu, May 29, 2014 at 12:06 AM, Tyson Norris <tnor...@adobe.com> wrote:
>
>> Hi -
>> Thanks - it turns out that the JSON parsing is actually fine with HEAD,
>> although inaccurate without the required message format (comments mention
>> expecting an “s” property with timestamp value).
>>
>>  My problem was that I was not specifying the spout root properly,
>> i.e. --spoutroot /transactional/<spout id>/user/  (in my case I had
>> specified a path that was valid, but not a spout)
>>
>>  Now I get offset info properly via monitor.py - Thanks!
>>
>>  Tyson
>>
>>  On May 28, 2014, at 10:12 AM, Cody A. Ray <cody.a....@gmail.com> wrote:
>>
>>  Right, its trying to read your kafka messages and parse as JSON.  See
>> the error:
>>
>>  simplejson.scanner.JSONDecodeError: Expecting value: line 1 column 1
>> (char 0)
>>
>>  If you want to use the BrightTag branch, you'll need to go a couple
>> commits back. Try this:
>>
>>  git clone https://github.com/BrightTag/stormkafkamon
>>  git checkout 07eede9ec72329fe2cad893d087541b583e11148
>>
>>  -Cody
>>
>>
>> On Wed, May 28, 2014 at 10:39 AM, Tyson Norris <tnor...@adobe.com> wrote:
>>
>>> Thanks Cody -
>>> I tried the BrightTag fork and still have problems with
>>> storm 0.9.1-incubating and kafka 0.8.1, I get an error with my trident
>>> topology (haven’t tried non-trident yet):
>>> (venv)tnorris-osx:stormkafkamon tnorris$ ./monitor.py --topology
>>> TrendingTagTopology --spoutroot storm --friendly
>>> Traceback (most recent call last):
>>>   File "./monitor.py", line 112, in <module>
>>>     sys.exit(main())
>>>   File "./monitor.py", line 96, in main
>>>     zk_data = process(zc.spouts(options.spoutroot, options.topology))
>>>   File "/git/github/stormkafkamon/stormkafkamon/zkclient.py", line 76,
>>> in spouts
>>>     j = json.loads(self.client.get(self._zjoin([spout_root, c, p]))[0])
>>>   File
>>> "/git/github/stormkafkamon/venv/lib/python2.7/site-packages/simplejson/__init__.py",
>>> line 501, in loads
>>>     return _default_decoder.decode(s)
>>>   File
>>> "/git/github/stormkafkamon/venv/lib/python2.7/site-packages/simplejson/decoder.py",
>>> line 370, in decode
>>>     obj, end = self.raw_decode(s)
>>>   File
>>> "/git/github/stormkafkamon/venv/lib/python2.7/site-packages/simplejson/decoder.py",
>>> line 389, in raw_decode
>>>     return self.scan_once(s, idx=_w(s, idx).end())
>>> simplejson.scanner.JSONDecodeError: Expecting value: line 1 column 1
>>> (char 0)
>>> (venv)tnorris-osx:stormkafkamon tnorris$
>>>
>>>  I’m not too familiar with python but will try to debug it as time
>>> allows - let me know if you have advice.
>>>
>>>  Thanks
>>>  Tyson
>>>
>>>
>>>
>>>
>>>  On May 28, 2014, at 7:20 AM, Cody A. Ray <cody.a....@gmail.com> wrote:
>>>
>>>  You can also use stormkafkamon to track this stuff. Its not good for
>>> historical analysis like graphite/ganglia, but its good if you just want to
>>> see how things currently stand.
>>>
>>>  The original: https://github.com/otoolep/stormkafkamon
>>>
>>>  This didn't work for us without some updates (incompatibility with the
>>> latest python-kafka dep). Here are those updates:
>>> https://github.com/BrightTag/stormkafkamon/commit/07eede9ec72329fe2cad893d087541b583e11148
>>>
>>>  (Our branch has a couple more things that parse the kafka messages
>>> with our format (which embeds a timestamp) to determine how long (in time)
>>> storm is behind... planning to clean that up soon so it can be a bit more
>>> reusable)
>>>
>>>  https://github.com/BrightTag/stormkafkamon
>>>
>>>  -Cody
>>>
>>>
>>> On Wed, May 28, 2014 at 4:50 AM, Danijel Schiavuzzi <
>>> dani...@schiavuzzi.com> wrote:
>>>
>>>> Yes, Trident Kafka spouts give you the same metrics. Take a look at the
>>>> code to find out what's available.
>>>>
>>>>
>>>> On Wed, May 28, 2014 at 3:55 AM, Tyson Norris <tnor...@adobe.com>
>>>> wrote:
>>>>
>>>>> Do Trident variants of kafka spouts do something similar?
>>>>> Thanks
>>>>> Tyson
>>>>>
>>>>> > On May 27, 2014, at 3:19 PM, "Harsha" <st...@harsha.io> wrote:
>>>>> >
>>>>> > Raphael,
>>>>> >        kafka spout sends metrics for kafkaOffset and kafkaPartition
>>>>> you can look at those by using LoggingMetrics or setting up a ganglia.
>>>>> Kafka uses its own zookeeper to store state info per topic & group.id
>>>>> you can look at kafka offsets using
>>>>> > kafka/bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker
>>>>> > -Harsha
>>>>> >
>>>>> >
>>>>> >> On Tue, May 27, 2014, at 03:01 PM, Raphael Hsieh wrote:
>>>>> >> Is there a way to tell where in the kafka stream my topology is
>>>>> starting from?
>>>>> >> From my understanding Storm will use zookeeper in order to tell its
>>>>> place in the Kafka stream. Where can I find metrics on this ?
>>>>> >> How can I see how large the stream is? What how much data is
>>>>> sitting in the stream and what the most recent/oldest position is?
>>>>> >>
>>>>> >> Thanks
>>>>> >>
>>>>> >> --
>>>>> >> Raphael Hsieh
>>>>>
>>>>
>>>>
>>>>
>>>>  --
>>>> Danijel Schiavuzzi
>>>>
>>>> E: dani...@schiavuzzi.com
>>>> W: www.schiavuzzi.com
>>>> T: +385989035562
>>>> Skype: danijels7
>>>>
>>>
>>>
>>>
>>>  --
>>>  Cody A. Ray, LEED AP
>>> cody.a....@gmail.com
>>> 215.501.7891
>>>
>>>
>>>
>>
>>
>>  --
>>  Cody A. Ray, LEED AP
>> cody.a....@gmail.com
>> 215.501.7891
>>
>>
>>
>
>
>  --
> Raphael Hsieh
>
>
>
>
>
>


-- 
Raphael Hsieh

Reply via email to