Sven,
What version of Accumulo are you running? We have a ticket for this [1]
which has had a lot of discussion on it. Christopher Tubbs mentioned that
he had gotten this to work.
[1] https://issues.apache.org/jira/browse/ACCUMULO-1378
On Wed, Sep 16, 2015 at 9:20 AM, Sven Hodapp
>>
>> Currently I'm using version 1.7
>>
>> Regards,
>> Sven
>>
>> - Ursprüngliche Mail -
>> > Von: "Corey Nolet" <cjno...@gmail.com>
>> > An: "user" <user@accumulo.apache.org>
>> > Gesendet: Mittwoc
Vaibnav,
The difference in an OR iterator is that you will want it to return a
single key for all of the given OR terms so that the iterator in the stack
above it would see it was a single hit. It's essentially a merge at the
key level to stop duplicate results from being returned (thus appearing
Andrew,
Have you considered leveraging existing SQL query layers like Hive or
Spark's SQL/DataFrames API? There are some pretty massive optimizations
involved in that API making the push-down predicates / selections pretty
easy to adapt for Accumulo.
On Mon, Apr 27, 2015 at 8:37 PM, Andrew Wells
I'm always looking for places to help out and integrate/share designs
ideas. I look forward to chatting with you about Q4A at the hackathon
tomorrow!
Have you, by chance, seen the Spark SQL adapter for the Accumulo Recipes
Event Entity Stores [1]? At the very least, it's a good example of using
The Apache Accumulo project is happy to announce its 1.6.2 release.
Version 1.6.2 is the most recent bug-fix release in its 1.6.x release line.
This version includes numerous bug fixes as well as a performance
improvement over previous versions. Existing users of 1.6.x are encouraged
to upgrade
(under a semver patch increment, this should be just as
strong an assertion as the reverse)
http://people.apache.org/~busbey/compat_reports/accumulo/1.6.2_to_1.6.1/compat_report.html
On Fri, Jan 23, 2015 at 8:02 PM, Corey Nolet cjno...@apache.org wrote:
Devs,
Please consider the following
I had a ticket for that awhile back and I don't believe it was ever
completed. By default, it wants to dump out new config files for
everything- have it reusing a config file would mean not re-initializing
each time and reusing the same instance id + rfiles.
ACCUMULO-1378 was the it and it looks
Dylan,
I know your original post mentioned grabbing it through the client API but
there's not currently a way to do that. As Sean mentioned, you can do it if
you have access to the cluster. You can run the reflection Keith provided
by adding the files in $ACCUMULO_HOME/lib/ to your classpath and
The Fluo project is happy to announce the 1.0.0-alpha-1 release of Fluo.
Fluo is a transaction layer that enables incremental processing on top of
Accumulo. It integrates into Yarn using Apache Twill.
This is the first release of Fluo and is not ready for production use. We
invite developers to
The Apache Accumulo project is happy to announce its 1.6.1 release.
Version 1.6.1 is the most recent bug-fix release in its 1.6.x release line.
This version includes numerous bug fixes and performance improvements over
previous versions. Existing users of 1.6.x are encouraged to upgrade to
this
, 2014 at 6:50 PM, Corey Nolet-2 [via Apache Accumulo] [hidden
email] http://user/SendEmail.jtp?type=nodenode=11303i=0 wrote:
Awesome John! It's good to have this documented for future users. Keep us
updated!
On Sun, Aug 24, 2014 at 11:05 AM, JavaHokie [hidden email]
http://user/SendEmail.jtp
I'm thinking this could be a yarn.application.classpath configuration
problem in your yarn-site.xml. I meant to ask earlier- how are you building
your jar that gets deployed? Are you shading it? Using libjars?
On Sun, Aug 24, 2014 at 6:56 AM, JavaHokie soozandjohny...@gmail.com
wrote:
Hey
Awesome John! It's good to have this documented for future users. Keep us
updated!
On Sun, Aug 24, 2014 at 11:05 AM, JavaHokie soozandjohny...@gmail.com
wrote:
Hi Corey,
Just to wrap things up, AccumuloMultipeTableInputFormat is working really
well. This is an outstanding feature I can
Awesome! I was going to recommend checking out the code last night so that
you could put some logging statements in there. You've probably noticed
this already but the MapWritable does not have static type parameters so it
dumps out the fully qualified class name so that it can instantiate it back
Hey John,
Could you give an example of one of the ranges you are using which causes
this to happen?
On Fri, Aug 22, 2014 at 11:02 PM, John Yost soozandjohny...@gmail.com
wrote:
Hey Everyone,
The AccumuloMultiTableInputFormat is an awesome addition to the Accumulo
API and I am really
The table configs get serialized as base64 and placed in the job's
Configuration under the key AccumuloInputFormat.ScanOpts.TableConfigs.
Could you verify/print what's being placed in this key in your
configuration?
On Sat, Aug 23, 2014 at 12:15 AM, JavaHokie soozandjohny...@gmail.com
wrote:
The tests I'm running aren't using the native Hadoop libs either. If you
don't mind, a little more code as to how you are setting up your job would
be useful. That's weird the key in the config would be null. Are you using
the job.getConfiguration()?
On Sat, Aug 23, 2014 at 12:31 AM, JavaHokie
at 1:11 AM, Corey Nolet cjno...@gmail.com wrote:
Job.getInstance(configuration) copies the configuration and makes its own.
Try doing your debug statement from earlier on job.getConfiguration() and
let's see what the base64 string looks like.
On Sat, Aug 23, 2014 at 1:00 AM, JavaHokie
That code I posted should be able to validate where you are getting hung
up. Can you try running that on the machine and seeing if it prints the
expected tables/ranges?
Also, are you running the job live? What does the configuration look like
for the job on your resource manager? Can you see if
Andrew,
Our recommendation on this has typically been to reverse the sort order of
the keys on ingest.
On Mon, Jun 30, 2014 at 12:24 PM, Andrew Wells awe...@clearedgeit.com
wrote:
Are there currently any good practices on doing this?
Especially when a rowId has a large number of Keys.
--
30, 2014 at 12:26 PM, Corey Nolet cjno...@gmail.com wrote:
Andrew,
Our recommendation on this has typically been to reverse the sort order
of the keys on ingest.
On Mon, Jun 30, 2014 at 12:24 PM, Andrew Wells awe...@clearedgeit.com
wrote:
Are there currently any good practices on doing
Jeff,
Unless you've got multiple different tables with different permissions to
manage for different physical Accumulo users, the connector should probably
be an instance variable in your service. It can be safely shared across all
the reads as long as the Accumulo user configured in the
. Still researching.
On Fri, May 16, 2014 at 4:19 PM, Corey Nolet cjno...@gmail.com wrote:
Can we assume this data has not yet been ingested? Do you have control
over the way in which you structure your table?
On Fri, May 16, 2014 at 1:54 PM, David Medinets david.medin...@gmail.com
wrote
Geoffry,
As Josh pointed out, you should only need the Hadoop libraries on the
client side to use the Text object. This means you won't have to go through
the pain of placing the xml files in your root bundles.
Did you try the JAAS export from the packages in your container? Did that
help?
I
=com.sun.security.auth.module
I also tried:
Import-Package: com.sun.security.auth.module
in my bundle that calls Hadoop.
Both of these were advised by one of the bndtools gurus--neither worked.
When I did the Import-Package other things broke.
On Wed, Apr 9, 2014 at 9:46 AM, Corey Nolet cjno...@gmail.com wrote
the Accumulo end of things in OSGi as well.
Wish me well
On Mon, Mar 31, 2014 at 8:34 PM, Corey Nolet cjno...@gmail.com wrote:
Geoffry,
What OSGi container are you using currently? The servicemix Hadoop
bundle should get you going with the Hadoop client dependencies at least
[1]. It looks
Geoffry,
What OSGi container are you using currently? The servicemix Hadoop bundle
should get you going with the Hadoop client dependencies at least [1]. It
looks like one of the servicemix guys created a Hadoop ticket for making
bundles of their jars as well [2], though it doesn't look like
Matt,
This should help:
CollectionPairText,Text cols = Collections.singleton(new
PairText,Text(new Text(cityOfBirth), null));
AccumuloInputFormat.fetchColumns(job, cols);
On Wed, Jan 15, 2014 at 7:29 PM, Dickson, Matt MR
matt.dick...@defence.gov.au wrote:
*UNOFFICIAL*
Thanks Keith. I've
Aaron,
We are currently re-working the AccumuloInputFormat for Accumulo 1.6 to
provide inputs from multiple tables (each with their own set of configured
iterators, ranges, columns). Check out ACCUMULO-391.
On Mon, Sep 16, 2013 at 11:41 AM, Aaron aarongm...@gmail.com wrote:
I was curious if
on a single table but the use case sounds interesting.
On Mon, Sep 16, 2013 at 3:55 PM, Corey Nolet cjno...@gmail.com wrote:
Aaron,
We are currently re-working the AccumuloInputFormat for Accumulo 1.6 to
provide inputs from multiple tables (each with their own set of configured
iterators
You are correct. Include as an attachment on the ticket.
On May 29, 2013 7:48 PM, Aaron aarongm...@gmail.com wrote:
Ok, is the best way just to add as an attachment? I did see that it got
rolled into a larger configurator issue..just upload the attachment to
the ticket? First time submitting
Would be useful to implement a -h flag that makes displays 1.4G or 500M
based on the input.
On Apr 23, 2013, at 10:31 AM, Keith Turner wrote:
On Tue, Apr 23, 2013 at 9:17 AM, Chris Sigman cypri...@gmail.com wrote:
Thanks Keith, I thought that might be the case considering the output,
The Thrift Proxy Server in 1.5 has had me excited so I started working on a C++
client API. I have the code checked in to GitHub if anyone is interested in
using or helping.
https://github.com/cjnolet/accumulo-cpp
to make this
better?
Thanks!
--
Corey Nolet
Senior Software Engineer
TexelTek, inc.
[Office] 301.880.7123
[Cell] 410-903-2110
that the iterator remains intact
through the entire scan, and it instead may be reconstructed, causing state
to be lost. I don't think this is the case for compaction time iterators, but
I'm not positive.
On Thu, Jan 3, 2013 at 5:41 PM, Corey Nolet cno...@texeltek.com wrote:
Hey Guys
to be reinitialized more
frequently. Maybe you are seeing this.
Keith
On Thu, Jan 3, 2013 at 5:41 PM, Corey Nolet cno...@texeltek.com wrote:
Hey Guys,
In Accumulo 1.3.5, I wrote a Top N table structure, services and a
FilteringIterator that would allow us to drop in several keys/values
associated
You can extend the output format to write to both and have the resulting record
writer underneath write to the correct endpoint depending on the items
submitted from the job.
On Oct 16, 2012, at 10:16 PM, Huanchen Zhang wrote:
Hello,
Hese I have a mapreduce job which needs to write to
Naturally, I could query the database before the ingest insert. But,
referring to slide 19 in Adam's presentation at
http://people.apache.org/~afuchs/slides/accumulo_table_design.pdf, the
read-modify-write design is not optimal.
--
Corey Nolet
Senior Software Engineer
TexelTek, inc.
[Office
39 matches
Mail list logo