Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Cassandra Wiki" for 
change notification.

The "HadoopSupport" page has been changed by JonathanEllis.
http://wiki.apache.org/cassandra/HadoopSupport?action=diff&rev1=3&rev2=4

--------------------------------------------------

- Cassandra version 0.6 and later support running Hadoop jobs against data in 
Cassandra, out of the box.  See 
https://svn.apache.org/repos/asf/cassandra/trunk/contrib/word_count/ for an 
example.  (Inserting the ''output'' of a Hadoop job into Cassandra has always 
been possible.)  Cassandra rows or row fragments (that is, pairs of (key, 
SortedMap of columns) are input to Map tasks for processing by your job, as 
specified by a `SlicePredicate` that describes which columns to fetch from each 
row.  Here's how this looks in the word_count example, which selects just one 
configurable columnName from each row:
+ Cassandra version 0.6 and later support running Hadoop jobs against data in 
Cassandra, out of the box.  See 
https://svn.apache.org/repos/asf/cassandra/trunk/contrib/word_count/ for an 
example.  (Inserting the ''output'' of a Hadoop job into Cassandra has always 
been possible.)  Cassandra rows or row fragments (that is, pairs of (key, 
`SortedMap` of columns) are input to Map tasks for processing by your job, as 
specified by a `SlicePredicate` that describes which columns to fetch from each 
row.  Here's how this looks in the word_count example, which selects just one 
configurable columnName from each row:
  
  {{{
              ConfigHelper.setColumnFamily(job.getConfiguration(), KEYSPACE, 
COLUMN_FAMILY);

Reply via email to