[ 
https://issues.apache.org/jira/browse/ACCUMULO-1197?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Josh Elser updated ACCUMULO-1197:
---------------------------------

    Description: 
The tracing functionality used by Accumulo provides nested regions of timing 
information for regular operations occurring inside of Accumulo, e.g. scans and 
compactions. There is basic functionality to view this information through the 
Accumulo monitor. This can be thought of as a distributed timing infrastructure 
for Accumulo which uses Accumulo to store its data.

Currently, this tracing doesn't fall through into HDFS. It would be awesome to 
actually introspect through the DFSClient, all the way down to the datanode 
writing to local disk. A large portion of the task would be investigating ways 
to inject the ability for calling applications (Accumulo, in this case) to 
provide the trace client through the Hadoop datanode code and record the 
necessary timings

Skills required would be a good understanding of Java. Some basic knowledge 
about Apache Hadoop would also be helpful.

  was:
The tracing functionality used by Accumulo provides nested regions of timing 
information for regular operations occurring inside of Accumulo, e.g. scans and 
compactions. There is basic functionality to view this information through the 
Accumulo monitor.

Currently, this tracing doesn't fall through into HDFS. It would be awesome to 
actually introspect through the DFSClient, all the way down to the datanode 
writing to local disk.

    
> Pass Accumulo trace functionality through the DFSClient
> -------------------------------------------------------
>
>                 Key: ACCUMULO-1197
>                 URL: https://issues.apache.org/jira/browse/ACCUMULO-1197
>             Project: Accumulo
>          Issue Type: Improvement
>          Components: trace
>            Reporter: Josh Elser
>            Assignee: Eric Newton
>              Labels: gsoc2013
>
> The tracing functionality used by Accumulo provides nested regions of timing 
> information for regular operations occurring inside of Accumulo, e.g. scans 
> and compactions. There is basic functionality to view this information 
> through the Accumulo monitor. This can be thought of as a distributed timing 
> infrastructure for Accumulo which uses Accumulo to store its data.
> Currently, this tracing doesn't fall through into HDFS. It would be awesome 
> to actually introspect through the DFSClient, all the way down to the 
> datanode writing to local disk. A large portion of the task would be 
> investigating ways to inject the ability for calling applications (Accumulo, 
> in this case) to provide the trace client through the Hadoop datanode code 
> and record the necessary timings
> Skills required would be a good understanding of Java. Some basic knowledge 
> about Apache Hadoop would also be helpful.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to