Official support for http sink?

2016-10-17 Thread Shang Wu (CS)
Hi,

We are Symantec Cloud Security group and Apache Flume looks very promising for 
our backend components, especially its wide supports for those popular data 
infrastructures via sources and sinks.
Having said that, we are looking for a very generic approach for restful 
API-based data storage systems, such as http sink, now.

We’ve found two http sink implementations in github, as follows, and would like 
to know any plan to include either of it, into the next release of Apache Flume?

https://github.com/josealvarezmuguerza/flume-http-sink
https://github.com/hmrc/flume-http-sink

Thanks adv.
BR,
Shang Wu, Senior Architect, BlueCoat, a Symantec Company.


Re: Which theme flume.apache.org use?

2016-10-17 Thread Mike Percy
If you want to improve the web site look + feel, you can check out the
source code for the web site at
http://svn.apache.org/viewvc/flume/site/trunk/

If you want to submit a patch for the web site, please let us know what
you're trying to do on d...@flume.apache.org so we can provide some guidance.

HTH,
Mike

On Fri, Oct 14, 2016 at 8:46 AM, Robert Lu  wrote:

> Hi, all.
>
> I want to generate flume locally, but when I run `sphinx-build -b html .
> ../html` in flume/flume-ng-doc/sphinx. I get the document with default
> theme.
>
> How can I generate official theme document?
>
> PS: I want improve document theme.
>
> --
> Robert Lu 
> About me: http://about.me/RobberPhex
>
>


Re: how to make KafkaSource consume the existing messages

2016-10-17 Thread Ping PW Wang
Hi Attila, Hi Chris, Many thanks for your reply and good comments. Yes, besides New group.id and "auto.offset.reset=earliest", "enable.auto.commit=false" is also needed to make KafkaSouce read messages from the beginning.   After I added the following lines into flume agent configuration, restarting flume then it worked as I expected.  tier1.sources.source1.kafka.consumer.group.id = group-new2 tier1.sources.source1.kafka.consumer.auto.offset.reset = earliest tier1.sources.source1.kafka.consumer.enable.auto.commit = falseThanks a lot for help!Attila Simon ---2016-10-13 下午 06:32:02---Hi, One more thing. If you switch to the new group.id and would like toFrom: Attila Simon To: user@flume.apache.org, d...@flume.apache.orgDate: 2016-10-13 下午 06:32Subject: Re: how to make KafkaSource consume the existing messages
Hi,One more thing. If you switch to the new group.id and would like tomaintain the read from beginning behaviour every time flume restartthen you might try setting enable.auto.commit to false.Again Kafka normally won't store the events indefinitely.Cheers,AttilaOn Thu, Oct 13, 2016 at 11:45 AM, Attila Simon  wrote:> for the records cc dev@>> On Thu, Oct 13, 2016 at 11:43 AM, Attila Simon  wrote:>> Hi, auto.offset.reset aim to handle failure scenarios when Flume lost the>> track of offsets. When Flume is able to successfully consume the>> messages it also commits the last processed offset. When failure>> happens and  was set resetting offset would use the last>> committed value.>> I don't think that always starting from "zero" offset would be>> valuable (would result a lot of duplicates). So I assume you would>> like to have a recovery scenario. What you can do is setting the>> consumer group.id to something new so if kafka still has the messages>> - you can check that with command line kafka consumer setting the>> --from-beginning argument as kafka by default purges them periodically>> - then flume would reset the offset to the effective beginning since>> offsets are stored per group.id. Quoted from Kafka docs>> (http://kafka.apache.org/documentation#newconsumerconfigs):>> auto.offset.reset - What to do when there is no initial offset in>> Kafka or if the current offset does not exist any more on the server>> (e.g. because that data has been deleted): earliest: automatically reset the offset to the earliest offset>> latest: automatically reset the offset to the latest offset>> none: throw exception to the consumer if no previous offset is found>> for the consumer's group>> anything else: throw exception to the consumer. Cheers,>> Attila>> On Thu, Oct 13, 2016 at 10:00 AM, Ping PW Wang  wrote:>>> Hi,>>> I used KafkaSource to consume the messages from Kafka. I found only new>>> messages were received while the old existing message not. I tried to use a>>> new consumer group and update the parameter "auto.offset.reset = latest" to>>> "earliest", but this does not work.>> tier2.sources.source1.kafka.consumer.group.id = test-consumer-group-new>>> tier2.sources.source1.kafka.consumer.auto.offset.reset = earliest>> Anyone knows how to make KafkaSource consume the existing messages?>>> Thanks a lot for any advice!>> Best Regards,>> Wang Ping (王苹)>>> InfoSphere BigInsights, CDL>>> Email: wpw...@cn.ibm.com Phone: (8610)82453448 Mobile: (86)17090815725>>> Address: Ring Bldg.No.28 Building,ZhongGuanCun Software Park,No.8 Dong Bei>>> Wang West Road, Haidian District Beijing P.R.China 100193>>> 地址:北京市海淀区东北旺西路8号,中关村软件园28号楼 邮编:100193>>>