Re: NiFi JSON enrichment

2018-12-27 Thread Austin Heyne
James, A little late to the show but hopefully this is useful. What we typically do for data enrichment is we'll use an EvaluateJsonPath processor to pull JSON fields out into attributes under a common key, e.g. foo.model. We then have a PutRedis processor that grabs everything under foo and

Re: NIfi cluster on Docker

2018-12-18 Thread Austin Heyne
/2018 03:52 PM, Jean-Sebastien Vachon wrote: Thanks for the advice. What about Zookeeper? are you running it within ECS as well? or are you managing it outside of ECS? Are there other alternatives? Thanks ---- *From:*

Re: NIfi cluster on Docker

2018-12-18 Thread Austin Heyne
We're using Nifi in ECS and the only gotcha's that come to mind are making sure you have a big enough EBS drive on the instance to handle all your flow files. You might get better performance with local storage but EBS has been good enough for us and we push a lot of data through it. You also m

Re: GetHDFS from Azure Blob

2017-03-29 Thread Austin Heyne
-site.xml, hdfs-site.xml and fs.xml in the 'Hadoop Configuration Resources' path ensuring the fs.xml comes last. This will overwrite the fs.defaultFS value set in core-site.xml. Thanks everyone for the help, Austin On 03/28/2017 06:11 PM, Austin Heyne wrote: Thanks Bryan, We're on

Re: GetHDFS from Azure Blob

2017-03-28 Thread Austin Heyne
the key for ACCOUNT3. Not really sure if that helps your situation. -Bryan On Tue, Mar 28, 2017 at 4:14 PM, Austin Heyne wrote: Bryan, So I initially didn't think much of it (assumed it a typo, etc) but you've said that the access url for wasb that you've been using is wasb://

Re: GetHDFS from Azure Blob

2017-03-28 Thread Austin Heyne
m here is that I don't see a way to specify different keys for each WASB filesystem in the core-site.xml. Admittedly I have never tried to setup something like this with many different filesystems. -Bryan On Tue, Mar 28, 2017 at 3:50 PM, Austin Heyne wrote: Hi Andre, Yes, I'm awar

Re: GetHDFS from Azure Blob

2017-03-28 Thread Austin Heyne
.xml' file or will revert to a default configuration" Have you tried using this field to point to a file as described by Bryan? Cheers On 29 Mar 2017 05:21, "Austin Heyne" <mailto:ahe...@ccri.com>> wrote: Thanks Bryan, Working with the configuration you s

Re: GetHDFS from Azure Blob

2017-03-28 Thread Austin Heyne
3.jar httpclient-4.2.5.jar httpcore-4.2.4.jar jackson-core-2.2.3.jar jsr305-1.3.9.jar slf4j-api-1.7.5.jar Thanks, Bryan On Tue, Mar 28, 2017 at 1:15 PM, Austin Heyne wrote: Hi all, Thanks for all the help you've given me so far. Today I'm trying to pull files from an Azure blob s

GetHDFS from Azure Blob

2017-03-28 Thread Austin Heyne
Hi all, Thanks for all the help you've given me so far. Today I'm trying to pull files from an Azure blob store. I've done some reading on this and from previous tickets [1] and guides [2] it seems the recommended approach is to place the required jars, to use the HDFS Azure protocol, in 'Add

Re: Routing File based on CSV header schema

2017-03-16 Thread Austin Heyne
, James Wing wrote: Austin, I think you are on the right track with RouteOnContent. Any chance you can share a sample CSV header, the settings of your RouteOnContent processor, including the regex? Thanks, James On Thu, Mar 16, 2017 at 11:14 AM, Austin Heyne <mailto:ahe...@ccri.com>

Routing File based on CSV header schema

2017-03-16 Thread Austin Heyne
won't work as columns with the same data type could be in a different order. Is there a ready made solution for this problem that I missed or perhaps a more clever way to approach it? Thanks, Austin Heyne