[ 
https://issues.apache.org/jira/browse/YARN-8617?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16714696#comment-16714696
 ] 

Bibin A Chundatt commented on YARN-8617:
----------------------------------------

[~Prabhu Joseph]

Looked into the issue again ..YARN-2583 contains 2 parts 

# limit number of files per node
  public static final String NM_LOG_AGGREGATION_NUM_LOG_FILES_SIZE_PER_APP
      = NM_PREFIX + "log-aggregation.num-log-files-per-app";
# Delete files old than expiry time.

{code}
        if (appDir.isDirectory() &&
            appDir.getModificationTime() < cutoffMillis) {
{code}

{quote}
The AggrgeatedLogDeletionService does deletion for Running Job based upon the 
file modification time which always will be latest as the rolled logs are 
getting updated into the node1 file regularly
{quote}

For long running service the *application folder* eg 
:user/logs/application_1234 modification time gets updated on every upload 
cycle.
This could cause nodefile to remain in hdfs if no new containers are allocated 
to same node.



> Aggregated Application Logs accumulates for long running jobs
> -------------------------------------------------------------
>
>                 Key: YARN-8617
>                 URL: https://issues.apache.org/jira/browse/YARN-8617
>             Project: Hadoop YARN
>          Issue Type: New Feature
>          Components: log-aggregation
>    Affects Versions: 2.7.4
>            Reporter: Prabhu Joseph
>            Priority: Major
>
> Currently AggregationDeletionService will delete older aggregated log files 
> once when they are complete. This will cause logs to accumulate for Long 
> Running Jobs like Llap, Spark Streaming.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org

Reply via email to