Re: [Xen-devel] Performance evaluation and Questions: Eliminating Xen (RTDS) scheduler overhead on dedicated CPU

2015-04-23 Thread Meng Xu
Hi Dario,

2015-04-23 8:48 GMT-04:00 Dario Faggioli :
> Hey, guys,
>
> I know, I know, I'm s much late to the party! Sorry, I got trapped
> into a thing that I really needed to finish... :-/
>
> I've got no intention to resurrect this old thread, just wanted to
> pointed out a few things.

What you pointed out is very interesting! I will have a look at them
and then show the numbers of the measurements.
(I will also need/have to redo some experiments to see how the TSC may
affect the results, which I didn't realize before. :-( )

>
> On Wed, 2015-04-08 at 16:52 -0400, Meng Xu wrote:
>> 2015-04-08 5:13 GMT-04:00 George Dunlap :
>> > On 04/07/2015 09:25 PM, Meng Xu wrote:
>> >> Hi George, Dario and Konrad,
>> >>
>> >> I finished a prototype of the RTDS scheduler with the dedicated CPU
>> >> feature and did some quick evaluation on this feature. Right now, I
>> >> need to refactor the code (because it is kind of messy when I was
>> >> exploring different approaches :() and will send out the clean patch
>> >> later (this week or next week). But the design follows our discussion
>> >> at 
>> >> http://lists.xenproject.org/archives/html/xen-devel/2015-03/msg02854.html.
>> >>
> The idea of 'dedicated CPU' makes sense. It's also always been quite
> common, in the Linux community, to see it as a real-time oriented
> feature. I personally don't agree much, as real-time is about
> determinism and dedicating a CPU to a task (in our case, that would mean
> dedicating a pCPU to a vCPU and then, in the guest, that vCPU to a task)
> does not automatically gives you determinism.
>
> Sure, it cuts off some overhead and some sources of unpredictable
> behavior (e.g., scheduler code), but not all of them (what about, for
> instance, caches shared with non-isolated pCPUs).

I'm working on eliminating the shared cache interference among guest
domains on non-isolated pCPUs. I have a prototype that can statically
partition the LLC into equal size partitions and assign it to guest
domains based on domain's configuration file. I'm running some
evaluations to show the effectiveness of the cache partitioning
approach (via page coloring) and will send you guys a slide about this
soon. :-)

>> > And in any case, as you say, it looks like the source of the overhead is
>> > the very frequent invocation of the RTDS scheduler.
>>
> Exactly! I'd put it this way: there are more urgent and more useful
> optimization, in general, but especially in RTDS, to be done before
> thinking at something like this.

Right.

>> > What I was expecting you to test, for the RTDS scheduler, was the
>> > wake-up latency.  Have you looked at that at all?
>>
> Indeed, that would be really interesting.
>
>> Ah, I didn't realize this... Do you have any concrete evaluation plan for 
>> this?
>> In my mind, I can issue hypercalls in domU to wake-up and sleep a vcpu
>> and measure how long it takes to wake up a vcpu. Maybe you have some
>> better idea in mind?
>> (The wake up latency of a vcpu will depends on the priority of the
>> vcpu and how heavy loaded the system is, in my speculation.)
>>
> Yes, that is something that could (should?) be done, as the wakeup
> latency of a vcpu is a lower bound for the wakeup latency of in-guest
> workloads, so we really want to know where we stand wrt to that, if we
> need to improve things, and if yes how.
>
> It's priority and load dependant... yes, of course, but that's why we
> have real-time schedulers for, isn't it? :-P Jokes apart, for the actual
> 'lower bound', we're clearly interesting to measure a vcpu when running
> alone on a pCPU, or with top priority.
>
> On the other hand, to look at wakeup latency from within the guest,
> cyclictest is the way to go:
> https://rt.wiki.kernel.org/index.php/Cyclictest
>
> What we want is to run it inside a guest, under different host and guest
> load conditions (and using different schedulers, varying the scheduling
> parameters, etc), and see what happens... Ever looked at that? I think
> it would be interesting.

I will have a look at the cyclictest and have some evaluation set up.
I totally agree that wakeup latency (on highest priority vcpu) is very
important for real-time applications. I will start a new thread about
the cyclictest result once I get it.

>
> I've done something similar while preparing this talk:
> https://archive.fosdem.org/2014/schedule/event/virtiaas16/
>
> But never got the chance to repeat the experiments (neither I did any
> further reasoning or investigation about how the timestamps are
> obtained, TSC emulation, etc., as George pointed out)
>
> That's all... Sorry again for chiming in only now. :-(

No problem at all! :-) Your advice and information are very useful!

Thanks and best regards,

Meng

---
Meng Xu
PhD Student in Computer and Information Science
University of Pennsylvania

___
Xen-devel mailing list
Xen-devel@lists.xen.org
http://lists.xen.org/xen-devel


Re: [Xen-devel] Performance evaluation and Questions: Eliminating Xen (RTDS) scheduler overhead on dedicated CPU

2015-04-23 Thread Dario Faggioli
Hey, guys,

I know, I know, I'm s much late to the party! Sorry, I got trapped
into a thing that I really needed to finish... :-/

I've got no intention to resurrect this old thread, just wanted to
pointed out a few things.

On Wed, 2015-04-08 at 16:52 -0400, Meng Xu wrote:
> 2015-04-08 5:13 GMT-04:00 George Dunlap :
> > On 04/07/2015 09:25 PM, Meng Xu wrote:
> >> Hi George, Dario and Konrad,
> >>
> >> I finished a prototype of the RTDS scheduler with the dedicated CPU
> >> feature and did some quick evaluation on this feature. Right now, I
> >> need to refactor the code (because it is kind of messy when I was
> >> exploring different approaches :() and will send out the clean patch
> >> later (this week or next week). But the design follows our discussion
> >> at 
> >> http://lists.xenproject.org/archives/html/xen-devel/2015-03/msg02854.html.
> >>
The idea of 'dedicated CPU' makes sense. It's also always been quite
common, in the Linux community, to see it as a real-time oriented
feature. I personally don't agree much, as real-time is about
determinism and dedicating a CPU to a task (in our case, that would mean
dedicating a pCPU to a vCPU and then, in the guest, that vCPU to a task)
does not automatically gives you determinism.

Sure, it cuts off some overhead and some sources of unpredictable
behavior (e.g., scheduler code), but not all of them (what about, for
instance, caches shared with non-isolated pCPUs). No, IMO, if you want
determinism, you should make the code deterministic, not get rid of
it! :-D

In fact, Linux has a feature similar the one Meng investigated, and that
has traditionally been used (at least until I was involved with Linux
scheduling) by HPC people, database engines and high frequency trading
use cases (which are also often categorized as 'real-time workloads' but
just aren't, IMO).

It's called isolcpus. For sure there was a boot time parameter for it,
and it looks like it is still there:
http://wiki.linuxcnc.org/cgi-bin/wiki.pl?The_Isolcpus_Boot_Parameter_And_GRUB2
http://www.linuxtopia.org/online_books/linux_kernel/kernel_configuration/re46.html
http://lxr.linux.no/linux+v3.19.1/Documentation/kernel-parameters.txt#L1530

I'm not sure whether they grew interfaces to setup this at runtime, but
doubt it.

For us, I'm not sure whether something like that would be useful. To be
fruitfully used together with something similar to Linux's isolcpus, it
need to look like how Meng is doing it, i.e., it ought to be possible to
handle single vCPUs, not full domains. However...

> > Secondly, adding an entirely new interface, as implementing the
> > "dedicated cpu" would require, on the other hand, is a fairly
> > significant cost.  It's costly for users to learn and configure the new
> > interface, it's costly to document, and once it's there we have to
> > continue to support it perhaps for a long time to come; and the feature
> > itself is also fairly complicated and increases the code maintenance.
> >
> > So the performance improvement you've shown so far I think is nowhere
> > near high enough a benefit to outweigh this cost.
> 
... I agree with George on this...

> OK. I see and agree.
> 
... and I'm happy you also do! :-D

> > And in any case, as you say, it looks like the source of the overhead is
> > the very frequent invocation of the RTDS scheduler.
>
Exactly! I'd put it this way: there are more urgent and more useful
optimization, in general, but especially in RTDS, to be done before
thinking at something like this.

>   You could probably
> > get the same kinds of benefits without adding any new interfaces by
> > reducing the amount of time the scheduler gets invoked when there are no
> > other tasks to run on that cpu.
> 
Exactly. And again, that is particularly relevant to RTDS, as numbers
show. Looking again at Linux world, this (i.e., avoiding invoking the
scheduler when there is only one task on a CPU) is also something
they've introduced rather recently.

It's called full dynticks:
https://www.kernel.org/doc/Documentation/timers/NO_HZ.txt
http://ertl.jp/~shinpei/conf/ospert13/slides/FredericWeisbecker.pdf
https://lwn.net/Articles/549580/
http://thread.gmane.org/gmane.linux.kernel/1485210 [*]

[*] check out Linus' replies... "awesome" as usual, he even managed to
rant about virtualization, all by himself!! :-P

That is IMO a line of action that may deserve some investigation.
RTDS-wise, for sure... the Credit-s, are not at all bad from that
perspective (as your numbers also show), but it might be possible to do
better.

> Yes. This is what Dagaen (cc.ed) is doing right now. He had a RFC
> patch and sent to me last week. We  are working on refining the patch
> before sending it out to the mailing list.
> 
I'll be super glad to see this! :-D

> >
> > What I was expecting you to test, for the RTDS scheduler, was the
> > wake-up latency.  Have you looked at that at all?
> 
Indeed, that would be really interesting.

> Ah, I didn't realize this... Do you have any concrete ev

Re: [Xen-devel] Performance evaluation and Questions: Eliminating Xen (RTDS) scheduler overhead on dedicated CPU

2015-04-08 Thread Meng Xu
2015-04-08 5:13 GMT-04:00 George Dunlap :
> On 04/07/2015 09:25 PM, Meng Xu wrote:
>> Hi George, Dario and Konrad,
>>
>> I finished a prototype of the RTDS scheduler with the dedicated CPU
>> feature and did some quick evaluation on this feature. Right now, I
>> need to refactor the code (because it is kind of messy when I was
>> exploring different approaches :() and will send out the clean patch
>> later (this week or next week). But the design follows our discussion
>> at http://lists.xenproject.org/archives/html/xen-devel/2015-03/msg02854.html.
>>
>> In a nutshell of the design, when a CPU is marked as dedicated CPU,
>> the scheduler on that CPU will return the dedicated VCPU on it with a
>> negative time so that it disable the scheduler timer on that CPU and
>> other CPUs will no longer send SCHEDULE_SOFTIRQ to the dedicated CPU.
>> The scheduler on the dedicated CPU may still be invoked when the
>> dedicated VCPU is blocked/unblocked by the domU. Once this situation
>> occurs, the schedule go though a fast pass to just return an idle
>> VCPU/the dedicated VCPU instead of go through the runq.
>>
>> I did the following evaluation to show the benefits of introducing
>> this dedicated CPU feature:
>>
>> I created a simple cpu-intensive task which just do the multiplication
>> for specified times:
>> start = rdtsc();
>> while ( i++ < cpu_measurement->multiply_times )
>> result += i * i;
>> finish = rdtsc();
>> latencies[k] = finish - start;
>>
>> I run this task and measure the execution time of above piece of code
>> on different environments: native Linux on bare metal, domU on Xen
>> with RTDS scheduler and domU on Xen with RTDS scheduler with dedicated
>> CPU feature, domU on Xen with Credit/Credit2 scheduler.
>>
>> The difference between the execution time in virtualization
>> environment and the execution time on native linux on bare metal is
>> the virtualization overhead introduced by Xen.
>>
>> I want to see that
>> 1) The virtualization overhead decreases a lot after the dedicated CPU
>> feature is employed for RTDS scheduler (because the execution of the
>> task will no longer suffer the scheduler overhead any more).
>> 2) The frequency of invoking the scheduler on the dedicated CPU
>> becomes very low once the dedicated CPU feature is applied.
>>
>> The result is as follows:
>> When the cpu-intensive task did the multiplication for 1024 times, the
>> execution time of the piece of code is:
>> 9264 cycles on native linux on bare metal;
>> 10320 cycles on Xen RTDS scheduler with dedicated CPU feature;
>> 10324 cycles on Xen RTDS scheduler without dedicated CPU feature;
>>
>> We didn't see the improvement of the dedicated CPU feature here
>> because the execution time is too short and it may not experience the
>> scheduler overhead yet.
>>
>> When the cpu-intensive task did the multiplication for  536870912
>> times, the execution time of the piece of code is:
>> 4838016028  cycles on native linux on bare metal;
>> 4839649567 cycles on Xen RTDS scheduler with dedicated CPU feature;
>> 4855509977 cycles on Xen RTDS scheduler without dedicated CPU feature;
>
> Hey Meng!  Thanks for looking at this.
>
> One thing: it's not entirely clear to me whether the numbers for
> "without dedicated CPU feature" are still with the equivalent of
> "pinning' -- i.e., is it guaranteed that no other vcpu will be run on
> the same cpu as the test program?

Yes. In the experiment, every VCPU is pinned to one isolated CPU. So
no other VCPU will run on the same cpu as the test program.

>
> Assuming that's the case, the numbers you give above show a 0.3%
> improvement for the "dedicated" cpu for cpu-intensive workloads.
>
>

Yes. This is the save when the cpu-intensive task did the
multiplication for  536870912 times.
Another thing to note is that the scheduling quantum of RTDS scheduler
is <= 1ms. If the scheduling quantum is smaller, the saving should
increase, but won't be too much in my speculation. (So I agree with
your conclusion that the benefits may not worth the complexity)

>> We can see that the dedicated CPU feature did save time for the
>> cpu-intensive task. Without the dedicated CPU feature, the hypervisor
>> scheduler may steal time from the domU and delay the execution of the
>> task inside domU.
>>
>> I did vary the number of multiplications of the above piece of code in
>> cpu-intensive task, and draw a figure to show the relation of the
>> overhead and the execution time of the task on native linux. The
>> figure can be found at
>> http://www.cis.upenn.edu/~mengxu/xen-ml/cpu-base-alone_multiply_0_0_100.virtOhVSwcetnative.pdf.
>> Please note the x-axis is the "log" value of the execution time. So
>> the overhead is actually linear to the execution time of the task.
>
> I'm not sure I can gain any useful information out of this graph.  A
> more useful comparison  would be to graph the execution time as an
> *overhead* compared to the Linux execution time.  F

Re: [Xen-devel] Performance evaluation and Questions: Eliminating Xen (RTDS) scheduler overhead on dedicated CPU

2015-04-08 Thread George Dunlap
On 04/07/2015 09:25 PM, Meng Xu wrote:
> Hi George, Dario and Konrad,
> 
> I finished a prototype of the RTDS scheduler with the dedicated CPU
> feature and did some quick evaluation on this feature. Right now, I
> need to refactor the code (because it is kind of messy when I was
> exploring different approaches :() and will send out the clean patch
> later (this week or next week). But the design follows our discussion
> at http://lists.xenproject.org/archives/html/xen-devel/2015-03/msg02854.html.
> 
> In a nutshell of the design, when a CPU is marked as dedicated CPU,
> the scheduler on that CPU will return the dedicated VCPU on it with a
> negative time so that it disable the scheduler timer on that CPU and
> other CPUs will no longer send SCHEDULE_SOFTIRQ to the dedicated CPU.
> The scheduler on the dedicated CPU may still be invoked when the
> dedicated VCPU is blocked/unblocked by the domU. Once this situation
> occurs, the schedule go though a fast pass to just return an idle
> VCPU/the dedicated VCPU instead of go through the runq.
> 
> I did the following evaluation to show the benefits of introducing
> this dedicated CPU feature:
> 
> I created a simple cpu-intensive task which just do the multiplication
> for specified times:
> start = rdtsc();
> while ( i++ < cpu_measurement->multiply_times )
> result += i * i;
> finish = rdtsc();
> latencies[k] = finish - start;
> 
> I run this task and measure the execution time of above piece of code
> on different environments: native Linux on bare metal, domU on Xen
> with RTDS scheduler and domU on Xen with RTDS scheduler with dedicated
> CPU feature, domU on Xen with Credit/Credit2 scheduler.
> 
> The difference between the execution time in virtualization
> environment and the execution time on native linux on bare metal is
> the virtualization overhead introduced by Xen.
> 
> I want to see that
> 1) The virtualization overhead decreases a lot after the dedicated CPU
> feature is employed for RTDS scheduler (because the execution of the
> task will no longer suffer the scheduler overhead any more).
> 2) The frequency of invoking the scheduler on the dedicated CPU
> becomes very low once the dedicated CPU feature is applied.
> 
> The result is as follows:
> When the cpu-intensive task did the multiplication for 1024 times, the
> execution time of the piece of code is:
> 9264 cycles on native linux on bare metal;
> 10320 cycles on Xen RTDS scheduler with dedicated CPU feature;
> 10324 cycles on Xen RTDS scheduler without dedicated CPU feature;
> 
> We didn't see the improvement of the dedicated CPU feature here
> because the execution time is too short and it may not experience the
> scheduler overhead yet.
> 
> When the cpu-intensive task did the multiplication for  536870912
> times, the execution time of the piece of code is:
> 4838016028  cycles on native linux on bare metal;
> 4839649567 cycles on Xen RTDS scheduler with dedicated CPU feature;
> 4855509977 cycles on Xen RTDS scheduler without dedicated CPU feature;

Hey Meng!  Thanks for looking at this.

One thing: it's not entirely clear to me whether the numbers for
"without dedicated CPU feature" are still with the equivalent of
"pinning' -- i.e., is it guaranteed that no other vcpu will be run on
the same cpu as the test program?

Assuming that's the case, the numbers you give above show a 0.3%
improvement for the "dedicated" cpu for cpu-intensive workloads.


> We can see that the dedicated CPU feature did save time for the
> cpu-intensive task. Without the dedicated CPU feature, the hypervisor
> scheduler may steal time from the domU and delay the execution of the
> task inside domU.
> 
> I did vary the number of multiplications of the above piece of code in
> cpu-intensive task, and draw a figure to show the relation of the
> overhead and the execution time of the task on native linux. The
> figure can be found at
> http://www.cis.upenn.edu/~mengxu/xen-ml/cpu-base-alone_multiply_0_0_100.virtOhVSwcetnative.pdf.
> Please note the x-axis is the "log" value of the execution time. So
> the overhead is actually linear to the execution time of the task.

I'm not sure I can gain any useful information out of this graph.  A
more useful comparison  would be to graph the execution time as an
*overhead* compared to the Linux execution time.  For instance, in the
numbers above, you'd have Linux = 1, RTDS+dedicated = 1.000337, RTDS =
1.00361.

But what it sounds like you're saying is that if you did such a graph,
the overhead would be pretty flat.  That's what I'd expect -- a fairly
constant overhead, regardless of how long you were running the test.

> As to the frequence of invoking the RTDS scheduler with/without the
> dedicated CPU feature, I add some code to trace which event triggers
> the scheduler on the dedicated cpu and how frequent it is.
> 
> Before we apply the dedicated CPU feature to the RTDS scheduler, the
> dedicated CPU 3 was invoked once
> ever