Please update your whole documents, it seems a lot of mismatches exist in different sections.
Sheng Wu 吴晟 Twitter, wusheng1108 han liu <[email protected]> 于2019年12月14日周六 下午2:54写道: > Okay, I have modified these two issues. > > 1. New field called maxThreadThresold. > Added new input box to "add-task-modal" tab in prototype > Added new conditions for limiting the size of the field in doc chapter > "Conditions that can create thread monitoring tasks" > Added description of this field in doc chapter "Thread monitoring table > structures" > > 2. Modify endpoint id to the endpoint name. > This field was modified in doc chapter "Thread monitoring table > structures" > Modify the input box to "add-task-modal" tab in prototype > > I will continue to modify for other issues in the comment. > > > Sheng Wu <[email protected]> 于2019年12月14日周六 上午9:46写道: > > > han liu <[email protected]> 于2019年12月13日周五 下午10:40写道: > > > > > I see what you mean. I think this is a good feature, and I will > summarize > > > this into the design doc. > > > > > > If you have any good suggestions,please let me know. > > > > > > > A simple way should be enough, set up sampling rule based on first span > > operation name, rather than endpoint it. > > In this case, plus #4056[1], there will be no id for local span or exit > > span. but those two are used in the first span in the async scenario. > > > > [1] https://github.com/apache/skywalking/issues/4056 > > > > > > Sheng Wu 吴晟 > > Twitter, wusheng1108 > > > > > > > > > > Sheng Wu <[email protected]> 于2019年12月13日周五 下午10:24写道: > > > > > > > Hi Han Liu > > > > > > > > One more reminder, a trace id in one instance could have multiple > > threads > > > > sampling in theory, such as across threads scenarios. We also should > > set > > > a > > > > threshold for this. Max 3 threads for one trace id maybe? > > > > > > > > Sheng Wu 吴晟 > > > > Twitter, wusheng1108 > > > > > > > > > > > > Sheng Wu <[email protected]> 于2019年12月13日周五 下午1:03写道: > > > > > > > > > Hi Han Liu and everyone > > > > > > > > > > I have submitted a design draft to the doc. Please take a look, if > > you > > > > > have an issue, please let me known. We could set up a online > meeting > > > too. > > > > > > > > > > Sheng Wu <[email protected]>于2019年12月12日 周四下午8:49写道: > > > > > > > > > >> Hi Han Liu > > > > >> > > > > >> I have replied the design with the most important key points I > > expect. > > > > >> Let's discuss those. After we are on the same page, we could > > continue > > > on > > > > >> more details. > > > > >> > > > > >> Sheng Wu 吴晟 > > > > >> Twitter, wusheng1108 > > > > >> > > > > >> > > > > >> han liu <[email protected]> 于2019年12月12日周四 下午2:26写道: > > > > >> > > > > >>> Due to formatting issues with previous mailboxes, they have been > > > > replaced > > > > >>> with new ones. > > > > >>> > > > > >>> I have completed some of the features in the google doc, and can > > > > provide > > > > >>> your comments and improvements. I will continue to improve the > > > > following > > > > >>> functions in the documentation. > > > > >>> The documentation is the same as you previously sent me. To > prevent > > > > >>> trouble, I'll post the link again here. > > > > >>> > > > > >>> > > > > > > > > > > https://docs.google.com/document/d/1rxMf1WN3PaFaZp7r8JmtwfdkmjLTcFW_ETAZv5FIU-s/edit# > > > > >>> > > > > >>> Sheng Wu <[email protected]> 于2019年12月10日周二 上午10:46写道: > > > > >>> > > > > >>> > 741550557 <[email protected]> 于2019年12月9日周一 下午9:42写道: > > > > >>> > > > > > >>> > > Thank for your reply, the issues you mentioned are very > > critical > > > > and > > > > >>> > > meaningful. > > > > >>> > > There I will answer what you mentioned. Sorry, I'm not good > at > > > > >>> comment > > > > >>> > > mode, so I use different colors and “ “ prefix to QA. > > > > >>> > > > > > > >>> > > > > > > >>> > > As we already have designed limit mechanism at backend and > > agent > > > > >>> > > side(according to your design), also the number would not be > > > > big(10 > > > > >>> most > > > > >>> > > likely), we just need a list to storage the trace-id(s) > > > > >>> > > > > > > >>> > > > > > > >>> > > If just need a list to storage trace-id(s), so how can I map > to > > > the > > > > >>> > > thread? I hope to use the map to quickly find thread info > from > > > > >>> trace-id. > > > > >>> > > How can I get thread-stack information from your way? Could > you > > > > >>> please > > > > >>> > > help elaborate? > > > > >>> > > > > > > >>> > > > > > >>> > Why do you need to do that? You just save a list of thread ids > > > which > > > > >>> should > > > > >>> > do thread dump, or remove some thread id from them when the > trace > > > id > > > > is > > > > >>> > finished. > > > > >>> > This is easy to do this by doing a loop search in the list. > > Right? > > > > >>> > Thread-stack is in the list, they are stored in an element. > Also, > > > > they > > > > >>> are > > > > >>> > in a list too. > > > > >>> > > > > > >>> > I think you were thinking the same all stack in a single map? > > That > > > > will > > > > >>> > cause a very dangerous memory and GC risk. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > Could you explain the (2), what do you mean `stop`? I think > if > > > > your > > > > >>> > > sampling mechanism should include the sampling duration. > > > > >>> > > > > > > >>> > > > > > > >>> > > As far as the communication between the sniffer and the OAP > > > > server, I > > > > >>> > hope > > > > >>> > > the sniffer only needs to obtain the thread-monitor task that > > > needs > > > > >>> to be > > > > >>> > > monitored at this time. The termination condition can be > > stopped > > > by > > > > >>> the > > > > >>> > > sniffer or the OAP server. > > > > >>> > > If It’s just an OAP server notification, it may be more > > > > complicated. > > > > >>> > Cause > > > > >>> > > OAP server need record sniffer has received the current > > command, > > > > and > > > > >>> > > sniffer is not stable, such as sniffer has shutdown when > > > receiving > > > > >>> the > > > > >>> > > command, at this time, no thread information I have been > > > collected. > > > > >>> > > > > > > >>> > > > > > > >>> > > I think that the active calculation termination by the OAP > > server > > > > can > > > > >>> > make > > > > >>> > > the monitoring more controllable, of course, the client can > > also > > > > >>> actively > > > > >>> > > report the end. > > > > >>> > > I think it’s necessary to provide a protection mechanism for > > the > > > > >>> sniffer > > > > >>> > > side, and it can be released quickly when the business peak > > > period > > > > >>> or the > > > > >>> > > probe suddenly occupies a lot of CPU / memory resources. > > > Therefore, > > > > >>> the > > > > >>> > > function of stopping monitoring can be provided in the UI > > > > interface, > > > > >>> so > > > > >>> > > that the sniffer can recover. > > > > >>> > > Sampling duration is required, but only as a default > > termination > > > > >>> > > thread-monitor condition. > > > > >>> > > > > > > >>> > > > > > >>> > But you should know, in the real case, the thread dump monitor > > is a > > > > >>> > sampling mechanism, you are even hard to know where they are > > > > happening. > > > > >>> > Then you have to send the stop notification to every instance. > > > > >>> > Even you could send the notification, but could you explain how > > you > > > > >>> know to > > > > >>> > stop? > > > > >>> > The scenario is, you are facing an issue, which trace and > metrics > > > > can't > > > > >>> > explain, so you active thread dump, right? At the same time, > you > > > want > > > > >>> to > > > > >>> > stop? > > > > >>> > > > > > >>> > CPU and memory resources should be guaranteed by design level, > > such > > > > as > > > > >>> > 1. Limited thread dump task for one service. > > > > >>> > 2. Limited thread dump traces in the certain time window. > > > > >>> > For example, the OAP backend/UI would say, you only could > > > > >>> > 1. Set 3 thread dump commands in the same time window. > > > > >>> > 2. Every command will require the sampling thread dump number > > > should > > > > be > > > > >>> > less than 5 traces. At the same time, in order to make this > > > sampling > > > > >>> works, > > > > >>> > only active sampling thread dump after the trace executed more > > than > > > > >>> > 200ms(value is an example only). > > > > >>> > 3. Thread dump could be sent to the backend duration sampling > to > > > > >>> reduce the > > > > >>> > memory cache. > > > > >>> > 4. Thread dump period should not less than 5ms, recommend 20ms > > > > >>> > 5. How depth the thread dump should do > > > > >>> > > > > > >>> > We need a very detailed design, above are just my thoughts, in > > > order > > > > to > > > > >>> > share the idea, the safe of the agent should not be by UI > button. > > > > >>> > Otherwise, your online system will be very dangerous, which is > > not > > > > the > > > > >>> > design goal of SkyWalking. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > The sampling period depends on how you are going to > visualize > > > it. > > > > >>> > > > > > > >>> > > > > > > >>> > > Yes, I agree. I hope can provide a select/input let trace > count > > > and > > > > >>> time > > > > >>> > > windows can be configurable in UI. Of course, this is my > > current > > > > >>> idea, > > > > >>> > and > > > > >>> > > if there have other plains, I will adopt it. > > > > >>> > > > > > > >>> > > > > > > >>> > > Highly doubt about this, reduce the memory, maybe, only > reduce > > > if > > > > >>> the > > > > >>> > > codes > > > > >>> > > are running the loop or facing lock issue. But if it is > > neither > > > of > > > > >>> these > > > > >>> > > two, they are different. > > > > >>> > > Also, please consider the CPU cost of the comparison of the > > > stack. > > > > >>> You > > > > >>> > > need > > > > >>> > > a performance benchmark to verify if you want this. > > > > >>> > > > > > > >>> > > > > > > >>> > > I didn’t understand that first sentence. In my personal > > > experience, > > > > >>> most > > > > >>> > > of the cases are blocking in the lock(socket/local) and > running > > > > >>> loop. I > > > > >>> > > have not imagined any other cases? > > > > >>> > > For the second sentence, I think I can add a > > thread-stack-element > > > > >>> field > > > > >>> > to > > > > >>> > > storage the top-level element of last stack information. When > > get > > > > >>> stack > > > > >>> > > information next time, I can compare the current top-level > > > element > > > > >>> that > > > > >>> > is > > > > >>> > > the same with that field. > > > > >>> > > I do this mainly to reduce duplicate thread-stack information > > > form > > > > >>> taking > > > > >>> > > up too much memory space, this is a way to optimizing memory > > > space. > > > > >>> It > > > > >>> > can > > > > >>> > > consider remove it, or do you have a better memory-saving > > > solution? > > > > >>> After > > > > >>> > > all, memory and CPU resources are very valuable in the > sniffer. > > > > >>> > > > > > > >>> > > > > > >>> > I know you mean about reducing the memory, but do you consider > > how > > > > >>> much CPU > > > > >>> > you will cost do a full thread dump comparison? The thread dump > > > could > > > > >>> > easily be hundreds of lines in Java. > > > > >>> > I mean this is a tradeoff, CPU or memory. If you are just using > > > > limited > > > > >>> > memory, before you could send the snapshot to backend while > > > > collecting > > > > >>> new, > > > > >>> > even could save into the disk(if really necessary). > > > > >>> > In my experience, compress is always very high risk in the > agent, > > > if > > > > >>> you > > > > >>> > want to do that, you need a benchmark test to improve that, > this > > > CPU > > > > >>> cost > > > > >>> > is small enough. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > The trace number and time window should be configurable, > that > > > is I > > > > >>> mean > > > > >>> > > more complex. Inthe current SamplingServcie, only n traces > > per 3 > > > > >>> > seconds. > > > > >>> > > But here, it is a dynamic rule. > > > > >>> > > > > > > >>> > > > > > > >>> > > I expect that it can be configured at the UI level for > special > > > > trace > > > > >>> > count > > > > >>> > > and time windows as I said above. > > > > >>> > > For SamplingService, my previous tech design was not rigorous > > > > >>> enough, and > > > > >>> > > there were indeed problems. > > > > >>> > > Maybe we need to extend a new SamplingService, build a > mapping > > > base > > > > >>> on > > > > >>> > > endpoint-id and AtomicInteger. > > > > >>> > > For `first 5 traces of this endpoint in the next 5 mins`, > just > > > need > > > > >>> to > > > > >>> > > increment it. > > > > >>> > > For sampling, maybe use another schedule task to reset > > > > AtomicInteger > > > > >>> > value. > > > > >>> > > > > > > >>> > > > > > >>> > You could avoid map, by using ArrayList with > > > > >>> RangeAtomicInteger(SkyWalking > > > > >>> > provides that) to let the trace context to get the slot. > > > > >>> > Also, you are considering `active sampling after trace > execution > > > time > > > > >>> more > > > > >>> > than xxx ms`, you should add remove mechanism during runtime. > > > > >>> > Anyway, try your best to avoid using Map, especially this map > > could > > > > be > > > > >>> > changed in the runtime. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > I think at least should be a level one new page called > > > > >>> configuration or > > > > >>> > > command page, which could set up the multiple sampling rule > > and > > > > >>> > visualize > > > > >>> > > the existing tasks and related sampling data. > > > > >>> > > > > > > >>> > > > > > > >>> > > I think it’s necessary to add a new page to the configuration > > > > >>> > > thread-monitor task, I think the specific UI display should > be > > > > >>> designed > > > > >>> > in > > > > >>> > > detail. > > > > >>> > > For example, what I expected is similar to the trace page. > The > > > left > > > > >>> side > > > > >>> > > displays the configuration, and the right side quickly > displays > > > the > > > > >>> > related > > > > >>> > > trace list. When clicked, it quickly links to the trace page > > and > > > > >>> displays > > > > >>> > > the sidebox display. > > > > >>> > > I ’m not good at this. Do you have any good plans? > > > > >>> > > > > > > >>> > > > > > >>> > UI is the thing that is hard to discuss by text, so I am pretty > > > sure, > > > > >>> we > > > > >>> > need some demo(could not be the codes, that is I mean drew by a > > > tool) > > > > >>> > It is OK to show a trace with thread dumps on another page, > even > > > > better > > > > >>> > linking to your task ID. > > > > >>> > But this kind of abstract description is hard to continue, no > > > > details I > > > > >>> > mean. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > And I feel that the two of us have a different understanding > of > > > the > > > > >>> > > configuration object. I think it is more of a task than a > > > command. > > > > I > > > > >>> > don't > > > > >>> > > know which way is better? > > > > >>> > > I suddenly thought of a problem. I think that some real > > problems > > > > are > > > > >>> > often > > > > >>> > > triggered at a specific period, such as a fixed business peak > > > > >>> period, and > > > > >>> > > we cannot guarantee that the user will operate on the UI. > > > > >>> > > So should the task mechanism be adopted to ensure that it can > > be > > > > run > > > > >>> at a > > > > >>> > > specific period? > > > > >>> > > > > > > >>> > > > > > >>> > This makes sense to me, and it is a just enhance feature. It is > > > just > > > > a > > > > >>> > start time sampling rule. > > > > >>> > > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > We don't have separated thread monitor view table, how about > > we > > > > add > > > > >>> an > > > > >>> > > icon > > > > >>> > > at the segment list, and add icon at the first span of this > > > > segment > > > > >>> in > > > > >>> > > trace detail view? > > > > >>> > > I think the latter one should be an entrance of the thread > > view. > > > > >>> > > > > > > >>> > > > > > > >>> > > I think it's a good idea. The link I mentioned in one of the > > > > answers > > > > >>> > > above, I think it is also a convenient entry point. > > > > >>> > > The switch button I mentioned earlier is only a data > filtering > > > item > > > > >>> in > > > > >>> > the > > > > >>> > > query of the trace list and does not need a separate table > UI. > > > > >>> > > > > > > >>> > > > > > >>> > As you intend to have a separated page for thread sampling, it > is > > > OK > > > > to > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > If you have some visualization idea, drawn by any tool you > > like > > > > >>> > supporting > > > > >>> > > comment, we could discuss it there. In my mind, we should > > > support > > > > >>> > > visualize > > > > >>> > > the thread dump stack through the time windows, and support > > > > >>> aggregate > > > > >>> > them > > > > >>> > > by choosing the continued stack snapshots on the time > window. > > > > >>> > > > > > > >>> > > > > > > >>> > > I think we should find a front-end who is better at > discussing > > > > >>> together > > > > >>> > > because this depends on how the front-end UI can be > displayed. > > > > >>> > > BTW: I can provide code for the OAP server and sniffer, and > the > > > > >>> frontend > > > > >>> > > may need to look for help in the community alone. Hope that > any > > > > >>> front-end > > > > >>> > > friends can participate in the topic discussion. > > > > >>> > > > > > > >>> > > > > > >>> > Once you have the demo, I could loop our UI committers in for > UI > > > side > > > > >>> > development. But UI committers may not be familiar with thread > > dump > > > > >>> context > > > > >>> > story. We need to resolve that first. > > > > >>> > Let's start up a demo, such as some slides on Google doc? > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > > > > > >>> > > > > > > >>> > > The above is my answer to all the questions, and I look > forward > > > to > > > > >>> your > > > > >>> > > reply at any time. As more and more discussions took place, > the > > > > >>> details > > > > >>> > > became more and more complete. This is good. > > > > >>> > > Everyone is welcome to discuss together if you have any > > questions > > > > or > > > > >>> good > > > > >>> > > ideas, please let me know. > > > > >>> > > > > > > >>> > > > > > >>> > I think we could move the discussion to the design doc as the > > next > > > > >>> step. > > > > >>> > > > > > >>> > Please use this > > > > >>> > > > > > >>> > > > > > >>> > > > > > > > > > > https://docs.google.com/document/d/1rxMf1WN3PaFaZp7r8JmtwfdkmjLTcFW_ETAZv5FIU-s/edit# > > > > >>> > Trite the design including > > > > >>> > 1. Key features > > > > >>> > 2. Protocol > > > > >>> > 3. Work mechanism > > > > >>> > 4. UI design, prototype > > > > >>> > and anything you think important before writing codes. > > > > >>> > > > > > >>> > This is SkyWalking CLI design doc, you could use it as a > > reference. > > > > >>> > > > > > >>> > > > > > >>> > > > > > > > > > > https://docs.google.com/document/d/1WBnRNF0ABxaSdBZo6Gv2hMzCQzj04YAePUdOyLWHWew/edit# > > > > >>> > > > > > >>> > > > > > >>> > > > > > > >>> > > > > > > >>> > > 原始邮件 > > > > >>> > > 发件人:Sheng [email protected] > > > > >>> > > 收件人:[email protected] > > > > >>> > > 发送时间:2019年12月9日(周一) 10:50 > > > > >>> > > 主题:Re: A proposal for Skywalking(thread monitor) > > > > >>> > > > > > > >>> > > > > > > >>> > > Hi Thanks for writing this proposal with a detailed design. > My > > > > >>> comments > > > > >>> > > are inline. 741550557 [email protected] 于2019年12月8日周日 > > 下午11:22写道: > > > > >>> Thanks > > > > >>> > > for your reply, I have carefully read these issues you > > mentioned, > > > > >>> and > > > > >>> > > these issues mentioned are very meaningful and critical. I > will > > > > >>> give you > > > > >>> > > technical details about the issues you mentioned below. I > find > > > > these > > > > >>> > > issues are related, so I will explain them in different > > > > dimensions. > > > > >>> > use > > > > >>> > > a different protocol to transmission trace and thread-stack: > > 1. > > > > add > > > > >>> a > > > > >>> > > boolean field in segment data, to record has thread > monitored. > > > and > > > > >>> is > > > > >>> > good > > > > >>> > > for filter monitored trace in UI. 2. add new BootService, > > > storage > > > > >>> Map to > > > > >>> > > record relate trace-id and trace-stack information. As we > > > already > > > > >>> have > > > > >>> > > designed limit mechanism at backend and agent side(according > to > > > > your > > > > >>> > > design), also the number would not be big(10 most likely), we > > > just > > > > >>> need a > > > > >>> > > list to storage the trace-id(s) 3. listen > > > > >>> > > TracingContextListener#afterFinished if the current segment > has > > > > >>> thread > > > > >>> > > monitored, mark current trace-id don’t need to monitor > anymore. > > > > >>> (Cause > > > > >>> > if > > > > >>> > > for-each the step 2 map, the remove operation will fail and > > throw > > > > >>> > > exception). 4. when thread-monitor main thread running, It > > will > > > > >>> for-each > > > > >>> > > step 2 map and check is it don’t need monitor anymore, I > will > > > put > > > > >>> data > > > > >>> > > into new data carrier. 5. generate new thread-monitor gRPC > > > > >>> protocol to > > > > >>> > > send data from the data carrier. The agent side design seems > > > > pretty > > > > >>> > good. > > > > >>> > > the server receives thread-stack logic: 1. storage > > stack-stack > > > > >>> > > informations and trace-id/segment-id relations on a > different > > > > >>> table. 2. > > > > >>> > > check thread-monitor is need to be stop on receiving data or > > > > >>> schedule. > > > > >>> > > Could you explain the (2), what do you mean `stop`? I think > if > > > your > > > > >>> > > sampling mechanism should include the sampling duration. > > > reduce > > > > >>> CPU > > > > >>> > and > > > > >>> > > memory in sniffer: 1. through the configuration of thread > > > > >>> monitoring in > > > > >>> > > the UI, you can configure the performance loss. For example, > > set > > > > the > > > > >>> > > monitoring level: fast monitoring (100ms), medium speed > > > monitoring > > > > >>> > > (500ms), slow speed monitoring (1000ms). The sampling > period > > > > >>> depends on > > > > >>> > > how you are going to visualize it. 2. add new integer field > on > > > per > > > > >>> > > thread-stack, if current thread-stack last element same as > > last > > > > >>> time, > > > > >>> > > don’t need storage, just increment it. I think it will save > a > > > lot > > > > of > > > > >>> > > memory space. Highly doubt about this, reduce the memory, > > maybe, > > > > only > > > > >>> > > reduce if the codes are running the loop or facing lock > issue. > > > But > > > > >>> if it > > > > >>> > is > > > > >>> > > neither of these two, they are different. Also, please > consider > > > the > > > > >>> CPU > > > > >>> > > cost of the comparison of the stack. You need a performance > > > > >>> benchmark to > > > > >>> > > verify if you want this. 3. create new VM args to setting > > > > >>> thread-monitor > > > > >>> > > pool size, It dependence on user, maybe default 3? (this can > > be > > > > >>> > discussed > > > > >>> > > later) I think UI limit is enough. 3 seems good to me. 4. > > limit > > > > >>> > > thread-stack-element size to 100, I think it can resolve most > > of > > > > the > > > > >>> > > scenes already. It also can create a new VM args if need. > > > > multiple > > > > >>> > > sampling methods can choose :(just my current thoughts, can > add > > > > >>> more) > > > > >>> > 1. > > > > >>> > > base on current client SamplingServcie, extra a new factor > > holder > > > > to > > > > >>> > > increment, and reset on schedule. Yours may be a little more > > > > complex > > > > >>> > than > > > > >>> > > the current SamplingServcie, right? Based on the next rule. > 2. > > > > >>> `first 5 > > > > >>> > > traces of this endpoint in the next 5 mins`, it a good idea. > My > > > > >>> > > understanding is that within a few minutes, each instance can > > > send > > > > a > > > > >>> > > specified number of traces. The trace number and time window > > > > should > > > > >>> be > > > > >>> > > configurable, that is I mean more complex. Inthe current > > > > >>> SamplingServcie, > > > > >>> > > only n traces per 3 seconds. But here, it is a dynamic rule. > > > UI > > > > >>> > settings > > > > >>> > > and sniffer perception: 1. create a new button on the > > dashboard > > > > >>> page, It > > > > >>> > > can create or stop a thread-monitor. It can be dynamic load > > > > >>> > thread-monitor > > > > >>> > > status when reselecting endpoint. I think at least should > be > > a > > > > >>> level > > > > >>> > one > > > > >>> > > new page called configuration or command page, which could > set > > up > > > > the > > > > >>> > > multiple sampling rule and visualize the existing tasks and > > > related > > > > >>> > > sampling data. 2. sniffer creates a new scheduled task to > > check > > > > the > > > > >>> > > current service has need monitor endpoint each 5 seconds. (I > > see > > > > >>> current > > > > >>> > > sniffer has command functions, feel that principle is the > same > > > as > > > > >>> the > > > > >>> > > scheduler) Seems reasonable. thread-monitor on the > > UI:(That’s > > > > >>> just my > > > > >>> > > initial thoughts, I think there will have a better way to > > show) > > > > 1. > > > > >>> When > > > > >>> > > switch to the trace page, I think we need to add a new switch > > > > >>> button to > > > > >>> > > filter thread-monitor trace. 2. make a new thread-monitor > icon > > > on > > > > >>> the > > > > >>> > same > > > > >>> > > segment. It means it has thread-stack information. We don't > > > have > > > > >>> > > separated thread monitor view table, how about we add an icon > > at > > > > the > > > > >>> > > segment list, and add icon at the first span of this segment > in > > > > trace > > > > >>> > > detail view? I think the latter one should be an entrance of > > the > > > > >>> thread > > > > >>> > > view. 3. show on the information sidebox when the user clicks > > the > > > > >>> > > thread-monitor segment(any span). create a new tab, like the > > log > > > > >>> tab. > > > > >>> > If > > > > >>> > > you have some visualization idea, drawn by any tool you like > > > > >>> supporting > > > > >>> > > comment, we could discuss it there. In my mind, we should > > support > > > > >>> > visualize > > > > >>> > > the thread dump stack through the time windows, and support > > > > aggregate > > > > >>> > them > > > > >>> > > by choosing the continued stack snapshots on the time window. > > > > >>> They're > > > > >>> > > just a description of my current implementation details for > > > > >>> > thread-monitor > > > > >>> > > if these seem to work. I can do some time planning for these > > > > tasks. > > > > >>> > Sorry, > > > > >>> > > my English is not very well, hope you can understand. Maybe > > > these > > > > >>> seem > > > > >>> > to > > > > >>> > > have some problem, any good idea or suggestion are welcome. > > Very > > > > >>> > > appreciated you to lead this new direction. It is a long term > > > task > > > > >>> but > > > > >>> > > should be interesting. :) Good work, carry on. 原始邮件 > > > 发件人:Sheng > > > > >>> > > [email protected] 收件人: > [email protected] > > > > >>> > > 发送时间:2019年12月8日(周日) 08:31 主题:Re: A proposal for > > > Skywalking(thread > > > > >>> > > monitor) First of all, thanks for your proposal. Thread > > > > >>> monitoring is > > > > >>> > > super important for application performance. So basically, I > > > agree > > > > >>> with > > > > >>> > > this proposal. But for tech details, I think we need more > > > > >>> discussion in > > > > >>> > > the following ways 1. Do you want to add thread status to > the > > > > >>> trace? If > > > > >>> > > so, why don't consider this as a UI level join? Because we > > could > > > > >>> know > > > > >>> > > thread id in the trace when we create a span, right? Then we > > > have > > > > >>> all > > > > >>> > the > > > > >>> > > thread dump(if), we could ask UI to query specific thread > > > context > > > > >>> based > > > > >>> > > on timestamp and thread number(s). 2. For thread dump, I > don't > > > > know > > > > >>> > > whether you do the performance evaluation for this OP. From > my > > > > >>> > > experiences, `get all need thread monitor segment every 100 > > > > >>> > milliseconds` > > > > >>> > > is a very high cost in your application and agent. So, you > may > > > > need > > > > >>> to > > > > >>> > be > > > > >>> > > careful about doing this. 3. Endpoint related thread dump > with > > > > some > > > > >>> > > sampling mechanisms makes more sense to me. And this should > be > > > > >>> activated > > > > >>> > > by UI. We should only provide a conditional thread dump > > sampling > > > > >>> > > mechanism, such as `first 5 traces of this endpoint in the > > next > > > 5 > > > > >>> mins`. > > > > >>> > > Jian Tan I think DaoCloud also has customized this feature > in > > > your > > > > >>> > > internal SkyWalking. Could you share what you do? Sheng Wu > 吴晟 > > > > >>> Twitter, > > > > >>> > > wusheng1108 741550557 [email protected] 于2019年12月8日周日 > > 上午12:14写道: > > > > >>> Hello > > > > >>> > > everyone, I would like to share a new feature with > skywalking, > > > > >>> called > > > > >>> > > “thread monitor”. Background When our company used > skywalking > > to > > > > APM > > > > >>> > > earlier, we found that many traces did not have enough span > to > > > > fill > > > > >>> up, > > > > >>> > > doubting whether there were some third-party frameworks that > > we > > > > >>> didn't > > > > >>> > > enhance or programmers API usage errors such as java > CountDown > > > > >>> number > > > > >>> > is 3 > > > > >>> > > but there are only 2 countdowns. So we decide to write a new > > > > >>> feature to > > > > >>> > > monitor executing trace thread stack, then we can get more > > > > >>> information > > > > >>> > on > > > > >>> > > the trace, quick known what’s happening on that trace. > > Structure > > > > >>> > > Agent(thread monitor) — gRPC protocol — OAP Server(Storage) > — > > > > >>> > > Skywalking-Rocketbot-UI More detail OAP Server: 1. Storage > > witch > > > > >>> traces > > > > >>> > > need to monitor(i suggest storage on the endpoint, add new > > > boolean > > > > >>> field > > > > >>> > > named needThreadMonitor) 2. Provide GraphQL API to change > > > endpoint > > > > >>> > monitor > > > > >>> > > status. 3. Monitor Trace parse, storage thread stack if the > > > > segment > > > > >>> has > > > > >>> > > any thread info. Skywalking-Rocketbot-UI: 1. Add a new > switch > > > > >>> button on > > > > >>> > > the dashboard, It can read or modify endpoint status. 2. It > > will > > > > >>> show > > > > >>> > > every thread stack on click trace detail. Agent: 1. setup > two > > > new > > > > >>> > > BootService: 1) find any need thread monitor endpoint in > > current > > > > >>> > service, > > > > >>> > > start on a new schedule take and works on each minute. 2) > > start > > > > new > > > > >>> > > schedule task to get all need thread monitor segment each > 100 > > > > >>> > > milliseconds, and put a new thread dump task to a global > > thread > > > > >>> > > pool(fixed, count number default 3). 2. check endpoint need > > > thread > > > > >>> > monitor > > > > >>> > > on create entry/local > > span(TracingConext#createEntry/LocalSpan). > > > > If > > > > >>> > need, > > > > >>> > > It will be marked and put into thread monitor map. 3. when > > > > >>> > TraceingContext > > > > >>> > > finishes, It will get thread has monitored, and send all > > thread > > > > >>> stack to > > > > >>> > > server. Finally, I don’t know it is a good idea to get more > > > > >>> information > > > > >>> > on > > > > >>> > > trace? If you have any good ideas or suggestions on this, > > please > > > > >>> let me > > > > >>> > > know. Mrpro > > > > >>> > > > > > >>> > > > > >> -- > > > > > Sheng Wu 吴晟 > > > > > > > > > > Apache SkyWalking > > > > > Apache Incubator > > > > > Apache ShardingSphere, ECharts, DolphinScheduler podlings > > > > > Zipkin > > > > > Twitter, wusheng1108 > > > > > > > > > > > > > > >
