CCCYounger opened a new issue, #9880:
URL: https://github.com/apache/dolphinscheduler/issues/9880

   ### Search before asking
   
   - [X] I had searched in the 
[issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and 
found no similar issues.
   
   
   ### What happened
   
   
我采用的是k8s部署(华为云cce),多个pod挂载同一个日志卷,发现存的文件没有区分是哪一个pod保存的,只能区分master、worker、api、alert,不能具体到哪一个worker,进一步检查后,发现每一小时的日志文件都是某一个pod的日志。
   这样会导致pod重启后无法查询日志,挂载日志卷就没有意义了。
   
   I used K8S deployment (Huawei cloud CCE), and multiple pods mounted the same 
log volume. I found that the saved files did not distinguish which POD saved 
them, but could only distinguish master, worker, API and alert, not specific 
worker. After further checking,It was found that the log file of each hour was 
the log file of a pod.
   In this case, the POD cannot query logs after the restart, and it is 
meaningless to mount the log volume.
   
   ### What you expected to happen
   
   
我查看了日志相关配置(logback-master.xml、logback-worker.xml、logback-api.xml、logback-alert.xml),配置的文件名没有区分具体的pod,这样会导致所有pod使用同一个文件。
   
   I looked at the log configuration (logback-master. XML, logback-worker. XML, 
logback-api. XML, logback-alert.xml),the configured file name does not 
discriminate between specific pods, which results in all pods using the same 
file.
   
   ### How to reproduce
   
   也许文件名加入${hostname}之类的名称会更好?
   
   Maybe the filename would be better with a name like ${hostname}?
   
   ### Anything else
   
   _No response_
   
   ### Version
   
   2.0.5
   
   ### Are you willing to submit PR?
   
   - [ ] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://www.apache.org/foundation/policies/conduct)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: 
[email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to