Re: [openstack-dev] We need a new version of hacking for Icehouse, or provide compatibility with oslo.sphinx in oslosphinx

2014-03-22 Thread Thomas Goirand
On 03/22/2014 12:58 AM, Joe Gordon wrote:
 
 So it sounds like we need:
 
 * Hacking 0.8.1 to fix the oslo.sphinx  oslosphinx issue for Icehouse.
 Since we cap hacking versions at 0.9 [1] this will  get used in icehouse.
 * Hacking 0.9 to release all the new hacking goodness. This will be
 targeted for use in Juno.

I agree with this plan.

Thomas


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [OpenStack-dev][NOVA][VCDriver][live-migration] VCDriver live migration problem

2014-03-22 Thread Shawn Hartsock
Hi Jay. We usually use [vmware] to tag discussion of VMware things. I
almost didn't see this message.

In short, there is a plan and we're currently blocked because we have
to address several other pressing issues in the driver before we can
address this one. Part of this is due to the fact that we can't press
harder on blueprints or changes to the VCDriver right now.

I actually reported this bug and we've discussed this at
https://wiki.openstack.org/wiki/Meetings/VMwareAPI the basic problem
is that live-migration actually works but you can't presently
formulate a command that activates the feature from the CLI under some
configurations. That's because of the introduction of clusters in the
VCDriver in Havana.

To fix this, we have to come up with a way to target a host inside the
cluster (as I pointed out in the bug) or we have to have some way for
a live migration to occur between clusters and a way to validate that
this can happen first.

As for the priority of this bug, it's been set to Medium which puts it
well behind many of the Critical or High tasks on our radar. As for
fixing the bug, no new outward behaviors or API are going to be
introduced and this was working at one point and now it's stopped. To
call this a new feature seems a bit strange.

So, moving forward... perhaps we need to re-evaluate the priority
order on some of these things. I tabled Juno planning during the last
VMwareAPI subteam meeting but I plan on starting the discussion next
week. We have a priority order for blueprints that we set as a team
and these are publicly recorded in our meeting logs and on the wiki.
I'll try to do better advertising these things. You are of course
invited... and yeah... if you're interested in what we're fixing next
in the VCDriver that next IRC meeting is where we'll start the
discussion.

On Sat, Mar 22, 2014 at 1:18 AM, Jay Lau jay.lau@gmail.com wrote:
 Hi,

 Currently we cannot do live migration with VCDriver in nova, live migration
 is really an important feature, so any plan to fix this?

 I noticed that there is already bug tracing this but seems no progress since
 last year's November: https://bugs.launchpad.net/nova/+bug/1192192

 Here just bring this problem up to see if there are any plan to fix this.
 After some investigation, I think that this might deserve to be a blueprint
 but not a bug.

 We may need to resolve issues for the following cases:
 1) How to live migration with only one nova compute? (one nova compute can
 manage multiple clusters and there can be multi hosts in one cluster)
 2) Support live migration between clusters
 3) Support live migration between resource pools
 4) Support live migration between hosts
 5) Support live migration between cluster and host
 6) Support live migration between cluster and resource pool
 7) Support live migration between resource pool and host
 8) Might be more cases.

 Please show your comments if any and correct me if anything is not correct.

 --
 Thanks,

 Jay

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




-- 
# Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Nova] Updates to Juno blueprint review process

2014-03-22 Thread Sean Dague
On 03/21/2014 02:55 PM, Stefano Maffulli wrote:
 On 03/20/2014 03:50 PM, Jay Lau wrote:
 It is better that we can have some diagram workflow just like
 Gerrit_Workflow https://wiki.openstack.org/wiki/Gerrit_Workflow to
 show the new process.
 
 Indeed, I think it would help.
 
 While I'm here, and for the records, I think that creating a new
 workflow 'temporarily' only until we have Storyboard usable, is a *huge*
 mistake. It seems to me that you're ignoring or at least underestimating
 the amount of *people* that will need to be retrained, the amount of
 documentation that need to be fixed/adjusted. And the confusion that
 this will create on the 'long tail' developers.
 
 A change like this, done with a couple of announcements on a mailing
 list and a few mentions on IRC is not enough to steer the ~400
 developers who may be affected by this change. And then we'll have to
 manage the change again when we switch to Storyboard. If I were you, I'd
 focus on getting storyboard ready to use asap, instead.
 
 There, I said it, and I'm now going back to my cave.
 
 .stef

Honestly, I largely disagree. This is applying some process where there
clearly wasn't any before. For people that aren't creating new
blueprints, nothing changes, except the blueprints they see in launchpad
are 100x more useful to understand what features the community is
interested in.

For people submitting blueprints, there is now actual guidance, instead
of randomly throw things into launchpad and not realize why your patch
is -2ed at milestone 3.

Storyboard remains vaporware. I will be enormously happy when it is not.
But you can't have people sit around waiting on vaporware, ever. Because
software that doesn't exist and people don't have to use always
magically seems to solve problems better than software that does exist.

I don't think anything about this process is incompatible with
Storyboard. The workflow will be very similar given any tracker. And I
expect the workflow might evolve over time if Storyboard can help more.
I honestly think spec review was completely outside of it's use case. So
I look forward to evolving the process with Storyboard once it's
actually a tool we use and not a Unicorn we talk about.

-Sean

-- 
Sean Dague
Samsung Research America
s...@dague.net / sean.da...@samsung.com
http://dague.net



signature.asc
Description: OpenPGP digital signature
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] An analysis of code review in Nova

2014-03-22 Thread Shawn Hartsock
On Fri, Mar 14, 2014 at 6:58 PM, Dan Smith d...@danplanet.com wrote:
 Review latency will be directly affected by how good the refactoring
 changes are staged. If they are small, on-topic and easy to validate,
 they will go quickly. They should be linearized unless there are some
 places where multiple sequences of changes make sense (i.e. refactoring
 a single file that results in no changes required to others).


I'm going to bring this to the next
https://wiki.openstack.org/wiki/Meetings/VMwareAPI we can start
working on how we'll set the order for this kind of work. Currently we
have a whole blueprint for refactoring a single method. That seems
silly. I'll want to come up with a plan around how to restructure the
driver so we can avoid some of the messes we've seen in the past.

I want to avoid one big refactor effort that drags on, but I also want
to address bigger problems we have inside the driver. For example,
vm_util.py seems to have become burdened with work that it shouldn't
have. It also performs a great number of unnecessary round trips using
a vm_ref to pull individual vm details over one at a time. Introducing
a VirtualMachine object that held all these references would simplify
some operations (I'm not the first person to suggest this and it
wasn't novel to me either when it was presented.)

It would seem Juno-1 would be the time to make these changes and we
need to serialize this work to keep reviewers from losing their
marbles trying to track it all. I would like to work out a plan for
this in conjunction with interested core-reviewers who would be
willing to more or less sponsor this work. Because of the problems
Matt points out, I don't want to tackle this in a haphazard or
piece-meal way since it could completely disrupt any new blueprint
work people may have targeted for Juno.

Having said that, on this driver, new blueprints in the last several
cycles have introduced serious feature regressions. Several minor bug
fixes have altered and introduced key architectural components that
have broken multiple critical features. In my professional opinion
this has a root cause based on the drivers tightly coupled and
non-cohesive internal design.

The driver design is tightly coupled in that a change in one area
forces multiple updates and changes *throughout* the rest of the
driver. This is true in testing as well. The testing design often
requires you to trace the entire codebase if you add a single optional
parameter to a single method. This does not have to be true.

The driver design is non-cohesive in that important details and
related information is spread throughout the driver. You must be aware
at all times (for example) whether or not your current operation
requires you to check if your vm_ref is outdated (we just worked on
several last minute critical bugs for RC1 where myself and others
pulled all nighters to fix the issue in a bad case of Heroic
Programming).

I would like to stop the http://c2.com/cgi/wiki?CodeVomit please. May we?

-- 
# Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [Ceilometer][QA][Tempest][Infra] Ceilometer tempest testing in gate

2014-03-22 Thread Sean Dague
On 03/21/2014 05:11 PM, Joe Gordon wrote:
 
 
 
 On Fri, Mar 21, 2014 at 4:04 AM, Sean Dague s...@dague.net
 mailto:s...@dague.net wrote:
 
 On 03/20/2014 06:18 PM, Joe Gordon wrote:
 
 
 
  On Thu, Mar 20, 2014 at 3:03 PM, Alexei Kornienko
  alexei.kornie...@gmail.com mailto:alexei.kornie...@gmail.com
 mailto:alexei.kornie...@gmail.com
 mailto:alexei.kornie...@gmail.com wrote:
 
  Hello,
 
  We've done some profiling and results are quite interesting:
  during 1,5 hour ceilometer inserted 59755 events (59755 calls to
  record_metering_data)
  this calls resulted in total 2591573 SQL queries.
 
  And the most interesting part is that 291569 queries were ROLLBACK
  queries.
  We do around 5 rollbacks to record a single event!
 
  I guess it means that MySQL backend is currently totally
 unusable in
  production environment.
 
 
  It should be noticed that SQLAlchemy is horrible for performance, in
  nova we usually see sqlalchemy overheads of well over 10x (time
  nova.db.api call vs the time MySQL measures when slow log is recording
  everything).
 
 That's not really a fair assessment. Python object inflation takes time.
 I do get that there is SQLA overhead here, but even if you trimmed it
 out you would not get the the mysql query time.
 
 
 To give an example from nova:
 
 doing a nova list with no servers:
 
 stack@devstack:~/devstack$ nova --timing list 
 
 | GET
 http://10.0.0.16:8774/v2/a82ededa9a934b93a7184d06f302d745/servers/detail
 | 0.0817470550537 |
 
 So nova command takes 0.0817470550537 seconds.
 
 Inside the nova logs (when putting a timer around all nova.db.api calls
 [1] ), nova.db.api.instance_get_all_by_filters takes 0.06 seconds:
 
 2014-03-21 20:58:46.760 DEBUG nova.db.api
 [req-91879f86-7665-4943-8953-41c92c42c030 demo demo]
 'instance_get_all_by_filters' 0.06 seconds timed
 /mnt/stack/nova/nova/db/api.py:1940
 
 But the sql slow long reports the same query takes only 0.001006 seconds
 with a lock_time of 0.000269 for a total of  0.00127 seconds.
 
 # Query_time: 0.001006  Lock_time: 0.000269 Rows_sent: 0
  Rows_examined: 0
 
 
 So in this case only 2% of the time
 that  nova.db.api.instance_get_all_by_filters takes is spent inside of
 mysql. Or to put it differently  nova.db.api.instance_get_all_by_filters
 is 47 times slower then the raw DB call underneath.
 
 Yes I agree that that turning raw sql data into python objects should
 take time, but I just don't think it should take 98% of the time.
 
 [1] 
 https://github.com/jogo/nova/commit/7743ee366bbf8746f1c0f634f29ebf73bff16ea1
 
 That being said, having Ceilometer's write path be highly tuned and not
 use SQLA (and written for every back end natively) is probably
 appropriate.
 
 
 While I like this idea, they loose free postgresql support by dropping
 SQLA. But that is a solvable problem.

Joe, you're just trolling now, right? :)

I mean you picked the most pathological case possible. An empty table
with no data ever returned. So no actual work was done anywhere, and
this is just measure side effects which in no way are commensurate with
actual read / write profiles of a real system.

I 100% agree that SQLA provides overhead. However, removing SQLA is the
last in a series of optimizations that you do on a system. Because
taking it out doesn't solve having bad data usage (getting more data
than you need), bad schema, or bad queries. I would expect substantial
gains could be made tackling those first.

If after that, fast path drivers sounded like a good idea, go for it.

But realize that a fast path driver is more work to write and maintain.
And has the energy hasn't gone into optimizing things yet, I think a
proposal to put even more work on the team to write a new set of harder
to maintain drivers, is just a non starter.

All I'm asking is that we need profiling. Ceilometer is suppose to be
high performance / low overhead metrics collection. We have some
indication that it's not meeting that desire based on our gate runs.
Which means we can reproduce it. Which is great, because reproducing
means things are fixable, and we can easily know if we did fix it.

Optimizing is hard, but I think it's the right time to do it. Not just
with elasticity, but with old fashion analysis.

-Sean

-- 
Sean Dague
Samsung Research America
s...@dague.net / sean.da...@samsung.com
http://dague.net



signature.asc
Description: OpenPGP digital signature
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [NOVA][VMWare][live-migration] VCDriver live migration problem

2014-03-22 Thread Jay Lau
Thanks Shawn, I have updated the title with VMWare.

Yes, I know that live migration works. But the problem is when a cluster
admin want to live migrate a VM instance, s/he will not know the target
host where to migrate, as s/he cannot get target host from nova compute
because currently VCDriver can only report cluster or resource pool as
hypervisor host but not ESX server.

IMHO, the VCDriver should support live migration between cluster, resource
pool and ESX host, so we may need do at least the following enhancements:
1) Enable live migration with even one nova compute. My current thinking is
that enhance target host as host:node when live migrate a VM instance and
the live migration task need
2) Enable VCDriver report all ESX servers.

We can discuss more during next week's IRC meeting.

Thanks!


2014-03-22 17:13 GMT+08:00 Shawn Hartsock harts...@acm.org:

 Hi Jay. We usually use [vmware] to tag discussion of VMware things. I
 almost didn't see this message.

 In short, there is a plan and we're currently blocked because we have
 to address several other pressing issues in the driver before we can
 address this one. Part of this is due to the fact that we can't press
 harder on blueprints or changes to the VCDriver right now.

 I actually reported this bug and we've discussed this at
 https://wiki.openstack.org/wiki/Meetings/VMwareAPI the basic problem
 is that live-migration actually works but you can't presently
 formulate a command that activates the feature from the CLI under some
 configurations. That's because of the introduction of clusters in the
 VCDriver in Havana.

 To fix this, we have to come up with a way to target a host inside the
 cluster (as I pointed out in the bug) or we have to have some way for
 a live migration to occur between clusters and a way to validate that
 this can happen first.

 As for the priority of this bug, it's been set to Medium which puts it
 well behind many of the Critical or High tasks on our radar. As for
 fixing the bug, no new outward behaviors or API are going to be
 introduced and this was working at one point and now it's stopped. To
 call this a new feature seems a bit strange.

 So, moving forward... perhaps we need to re-evaluate the priority
 order on some of these things. I tabled Juno planning during the last
 VMwareAPI subteam meeting but I plan on starting the discussion next
 week. We have a priority order for blueprints that we set as a team
 and these are publicly recorded in our meeting logs and on the wiki.
 I'll try to do better advertising these things. You are of course
 invited... and yeah... if you're interested in what we're fixing next
 in the VCDriver that next IRC meeting is where we'll start the
 discussion.

 On Sat, Mar 22, 2014 at 1:18 AM, Jay Lau jay.lau@gmail.com wrote:
  Hi,
 
  Currently we cannot do live migration with VCDriver in nova, live
 migration
  is really an important feature, so any plan to fix this?
 
  I noticed that there is already bug tracing this but seems no progress
 since
  last year's November: https://bugs.launchpad.net/nova/+bug/1192192
 
  Here just bring this problem up to see if there are any plan to fix this.
  After some investigation, I think that this might deserve to be a
 blueprint
  but not a bug.
 
  We may need to resolve issues for the following cases:
  1) How to live migration with only one nova compute? (one nova compute
 can
  manage multiple clusters and there can be multi hosts in one cluster)
  2) Support live migration between clusters
  3) Support live migration between resource pools
  4) Support live migration between hosts
  5) Support live migration between cluster and host
  6) Support live migration between cluster and resource pool
  7) Support live migration between resource pool and host
  8) Might be more cases.
 
  Please show your comments if any and correct me if anything is not
 correct.
 
  --
  Thanks,
 
  Jay
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 



 --
 # Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




-- 
Thanks,

Jay
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] Preparing for 2013.2.3 -- branches freeze March 27th

2014-03-22 Thread Thomas Goirand
On 03/22/2014 03:32 AM, Adam Gandelman wrote:
 Hi All-
 
 We'll be freezing the stable/havana branches for integrated projects this
 Thursday March 27th in preparation for the 2013.2.3 stable release on
 Thursday April 3rd.  You can view the current queue of proposed patches
 on gerrit [1].  I'd like to request all interested parties review current
 bugs affecting Havana and help ensure any relevant fixes be proposed
 soon and merged by Thursday, or notify the stable-maint team of
 anything critical that may land late.
 
 Thanks,
 Adam
 
 [1] https://review.openstack.org/#/q/status:open+branch:stable/havana,n,z

Hi,

There's indeed 2 bugs which I would very much like to be fixed.

The first one would be the backport of oauth2 to oauthlib in Keystone,
which is IMO critical for security reasons:
https://review.openstack.org/#/c/70750/

It is my view that core reviewers should have paid more attention to
this backport, and that it shouldn't have taken that long... :/

Note that I currently embed a Debian specific patch in the package, but
I really would love to get rid of it, and use something that has gone
through the review process.

The 2nd one would be the removal of minified Javascript in Horizon:
https://review.openstack.org/#/c/67096/

I'm not sure that 2nd one has been backported yet, though it is
important that it lands into the stable release, because otherwise that
makes Horizon non-free. Therefore, I just created a patch for the
stable/havana branch over here:

https://review.openstack.org/82298

I'd really like to have it included in the next stable release of
Horizon. I don't think this one is hard to deal with, and I hope it will
go through even with the lack of time.

Cheers,

Thomas Goirand (zigo)

P.S: I'm not registered to the openstack-stable-maint list.


___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] An analysis of code review in Nova

2014-03-22 Thread Matt Riedemann



On 3/22/2014 5:19 AM, Shawn Hartsock wrote:

On Fri, Mar 14, 2014 at 6:58 PM, Dan Smith d...@danplanet.com wrote:

Review latency will be directly affected by how good the refactoring
changes are staged. If they are small, on-topic and easy to validate,
they will go quickly. They should be linearized unless there are some
places where multiple sequences of changes make sense (i.e. refactoring
a single file that results in no changes required to others).



I'm going to bring this to the next
https://wiki.openstack.org/wiki/Meetings/VMwareAPI we can start
working on how we'll set the order for this kind of work. Currently we
have a whole blueprint for refactoring a single method. That seems
silly. I'll want to come up with a plan around how to restructure the
driver so we can avoid some of the messes we've seen in the past.


I think the point of starting with refactoring the nested method mess in 
the spawn method was it (1) seemed relatively trivial (think fast review 
turnaround) and (2) should be a good bang for the buck kind of change, 
as a lot of the original complaint was related to how hard it is to 
verify changes in the giant spawn method are tested - which you also 
point out below.




I want to avoid one big refactor effort that drags on, but I also want
to address bigger problems we have inside the driver. For example,


I also want to avoid a big refactor effort dragging on, and I like the 
thinking on design changes, but are they doing going to be happening at 
the same time?  Or is the complete re-design going to supersede the 
refactoring?  My only concern is biting off more than can be chewed in 
juno-1.


Plus there is the refactor to use oslo.vmware, where does that fit into 
this?



vm_util.py seems to have become burdened with work that it shouldn't
have. It also performs a great number of unnecessary round trips using
a vm_ref to pull individual vm details over one at a time. Introducing
a VirtualMachine object that held all these references would simplify
some operations (I'm not the first person to suggest this and it
wasn't novel to me either when it was presented.)

It would seem Juno-1 would be the time to make these changes and we
need to serialize this work to keep reviewers from losing their
marbles trying to track it all. I would like to work out a plan for
this in conjunction with interested core-reviewers who would be
willing to more or less sponsor this work. Because of the problems
Matt points out, I don't want to tackle this in a haphazard or
piece-meal way since it could completely disrupt any new blueprint
work people may have targeted for Juno.


Yeah, definitely need a plan here.  I'd like to see things prioritized 
based on what can be fixed in a relatively isolated way which gives a 
good return on coding/reviewing investment, e.g. pulling those nested 
methods out of spawn so they can be unit tested individually with mock 
rather than a large, seemingly rigid and scaffolded test framework.




Having said that, on this driver, new blueprints in the last several
cycles have introduced serious feature regressions. Several minor bug
fixes have altered and introduced key architectural components that
have broken multiple critical features. In my professional opinion
this has a root cause based on the drivers tightly coupled and
non-cohesive internal design.

The driver design is tightly coupled in that a change in one area
forces multiple updates and changes *throughout* the rest of the
driver. This is true in testing as well. The testing design often
requires you to trace the entire codebase if you add a single optional
parameter to a single method. This does not have to be true.


Yup, this is my major complaint and ties into what I'm saying above, I 
find it really difficult to determine most of the time where a change is 
tested.  Because of the nature of the driver code and my lack of 
actually writing features in it, as a reviewer I don't know if a change 
in X is going to break Y, so I rely on solid test coverage and the 
testing needs to be more natural to follow than it currently is.




The driver design is non-cohesive in that important details and
related information is spread throughout the driver. You must be aware
at all times (for example) whether or not your current operation
requires you to check if your vm_ref is outdated (we just worked on
several last minute critical bugs for RC1 where myself and others
pulled all nighters to fix the issue in a bad case of Heroic
Programming).

I would like to stop the http://c2.com/cgi/wiki?CodeVomit please. May we?



I know this isn't going to be easy so I'm really glad you're planning on 
tackling it in Juno.  I'll tentatively sign up to help sponsor this but 
I'm not going to be able to commit all of my review bandwidth to a ton 
of changes for refactor and re-design.  Hopefully targets will become 
more clear as the team gets the plans in place.


--

Thanks,

Matt Riedemann



Re: [openstack-dev] [nova] An analysis of code review in Nova

2014-03-22 Thread Chris Behrens
I'd like to get spawn broken up sooner rather than later, personally. It has 
additional benefits of being able to do better orchestration of builds from 
conductor, etc.

On Mar 14, 2014, at 3:58 PM, Dan Smith d...@danplanet.com wrote:

 Just to answer this point, despite the review latency, please don't be
 tempted to think one big change will get in quicker than a series of
 little, easy to review, changes. All changes are not equal. A large
 change often scares me away to easier to review patches.
 
 Seems like, for Juno-1, it would be worth cancelling all non-urgent
 bug fixes, and doing the refactoring we need.
 
 I think the aim here should be better (and easier to understand) unit
 test coverage. Thats a great way to drive good code structure.
 
 Review latency will be directly affected by how good the refactoring
 changes are staged. If they are small, on-topic and easy to validate,
 they will go quickly. They should be linearized unless there are some
 places where multiple sequences of changes make sense (i.e. refactoring
 a single file that results in no changes required to others).
 
 As John says, if it's just a big change everything patch, or a ton of
 smaller ones that don't fit a plan or process, then it will be slow and
 painful (for everyone).
 
 +1 sounds like a good first step is to move to oslo.vmware
 
 I'm not sure whether I think that refactoring spawn would be better done
 first or second. My gut tells me that doing spawn first would mean that
 we could more easily validate the oslo refactors because (a) spawn is
 impossible to follow right now and (b) refactoring it to smaller methods
 should be fairly easy. The tests for spawn are equally hard to follow
 and refactoring it first would yield a bunch of more unit-y tests that
 would help us follow the oslo refactoring.
 
 However, it sounds like the osloificastion has maybe already started and
 that refactoring spawn will have to take a backseat to that.
 
 --Dan
 
 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


[openstack-dev] OpenStack/GSoC 2014 - Reviewing Applications

2014-03-22 Thread Davanum Srinivas
Dear Students,

We received the following proposals:

AMES-Cloud (Nagashree Bhat)
A pre-caching system for OpenStack (Anastasis Andronidis)
Proposal for Implementing an application-level FWaaS driver
((Zorp) Dániel Csubák)
Cassandra and Redis storage backend for Openstack marconi (Jeremy Henriques)
Openstack-OSLO :Add a New Backend to Oslo.Cache (sai krishna)
Implement a Fuzz testing framework that can be run on Tempest
(Manishanker Talusani)
How to detect network anomalies from telemetry data within Openstack (mst89)
Implement a re-usable shared library for VMware(oslo.vmware) (Masaru Nomura)
Cross-services Scheduler project (Artem Shepelev)
OpenStack/Marconi: Py3k support (Nataliia)
Improve benchmarking context mechanism in Rally (Kumar Rishabh)
Develop a benchmarking suite and new storage backends to OpenStack
Marconi (Prashanth Raghu)
Adding Redis as a Storage Backend to OpenStack Marconi (Chenchong Qin)
Auto Benchmarking System for OpenStack (RobberPhex)
Developing Benchmarks for Virtual Machines of OpenStack with Rally
(Tzanetos Balitsaris)
Add a new storage backend to the OpenStack Message Queuing Service
(Victoria Martínez de la Cruz)

Dear Mentors 
(ddutta/flwang/julim/hughsaunders/greghaynes/annegentle/sriramhere/arnaudleg/coroner/boris_42/blint/ybudupi/cppcabrera),

Please log into the Google GSoC web-site and review **all** proposals
within a week (our own deadline - say 29th).

1) Please click Wish to mentor on the left hand side, if you are
willing to mentor this project.
   Yes, we should have more than one mentor. This is a way to mention
that you are willing to mentor,
   actual assignment will happen later i believe. So please switch
this on for all projects you are
   interested in.

2) If you have a question about a proposal and need a response from
the candidate, please leave
   a comment for them. If you want to ask a observation/question to
other mentors or me, leave a
   comment marked Private

3) To assign a score. Please click on the number of start in My
score: at the bottom
5 = amazing applicant, could be a module maintainer on completing
the program
4 = strong applicant, will do a good job
3 = good applicant, but is somewhat inexperienced
2 = is unlikely to do a good job
1 = not a good candidate

Please feel free to leave detailed notes on candidates you know well
for other mentors (Please mark comments as Private)

If you are not able to help with this exercise, please let me know ASAP.

If anyone else would like to step up and be a Mentor for any of these
projects/students, please register in the Google GSoC site and let me
know your username.

Thanks,
dims



-- 
Davanum Srinivas :: http://davanum.wordpress.com

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [NOVA][VMWare][live-migration] VCDriver live migration problem

2014-03-22 Thread Shawn Hartsock
Jay,

On point number 2 I proposed
https://blueprints.launchpad.net/nova/+spec/vmware-auto-inventory soon
after we merged the Cluster patch because of just the issues you are
highlighting. I may have named it badly and it's purpose maybe
outdated now... so I should definitely work on clarifying things like
that.

My original plan for what I called auto-inventory was to enable the
VCDriver to pull out individual ESX hosts and present them as nodes to
Nova.  So I mention this just to say, many of us have the same ideas
but we've not acted on them yet. I think the time will be right to
start working on some of these issues in Juno.

Talk to you Wednesday. We won't finish the discussion then, but we can
at least start having them.

On Sat, Mar 22, 2014 at 8:22 AM, Jay Lau jay.lau@gmail.com wrote:
 Thanks Shawn, I have updated the title with VMWare.

 Yes, I know that live migration works. But the problem is when a cluster
 admin want to live migrate a VM instance, s/he will not know the target host
 where to migrate, as s/he cannot get target host from nova compute because
 currently VCDriver can only report cluster or resource pool as hypervisor
 host but not ESX server.

 IMHO, the VCDriver should support live migration between cluster, resource
 pool and ESX host, so we may need do at least the following enhancements:
 1) Enable live migration with even one nova compute. My current thinking is
 that enhance target host as host:node when live migrate a VM instance and
 the live migration task need
 2) Enable VCDriver report all ESX servers.

 We can discuss more during next week's IRC meeting.

 Thanks!


 2014-03-22 17:13 GMT+08:00 Shawn Hartsock harts...@acm.org:

 Hi Jay. We usually use [vmware] to tag discussion of VMware things. I
 almost didn't see this message.

 In short, there is a plan and we're currently blocked because we have
 to address several other pressing issues in the driver before we can
 address this one. Part of this is due to the fact that we can't press
 harder on blueprints or changes to the VCDriver right now.

 I actually reported this bug and we've discussed this at
 https://wiki.openstack.org/wiki/Meetings/VMwareAPI the basic problem
 is that live-migration actually works but you can't presently
 formulate a command that activates the feature from the CLI under some
 configurations. That's because of the introduction of clusters in the
 VCDriver in Havana.

 To fix this, we have to come up with a way to target a host inside the
 cluster (as I pointed out in the bug) or we have to have some way for
 a live migration to occur between clusters and a way to validate that
 this can happen first.

 As for the priority of this bug, it's been set to Medium which puts it
 well behind many of the Critical or High tasks on our radar. As for
 fixing the bug, no new outward behaviors or API are going to be
 introduced and this was working at one point and now it's stopped. To
 call this a new feature seems a bit strange.

 So, moving forward... perhaps we need to re-evaluate the priority
 order on some of these things. I tabled Juno planning during the last
 VMwareAPI subteam meeting but I plan on starting the discussion next
 week. We have a priority order for blueprints that we set as a team
 and these are publicly recorded in our meeting logs and on the wiki.
 I'll try to do better advertising these things. You are of course
 invited... and yeah... if you're interested in what we're fixing next
 in the VCDriver that next IRC meeting is where we'll start the
 discussion.

 On Sat, Mar 22, 2014 at 1:18 AM, Jay Lau jay.lau@gmail.com wrote:
  Hi,
 
  Currently we cannot do live migration with VCDriver in nova, live
  migration
  is really an important feature, so any plan to fix this?
 
  I noticed that there is already bug tracing this but seems no progress
  since
  last year's November: https://bugs.launchpad.net/nova/+bug/1192192
 
  Here just bring this problem up to see if there are any plan to fix
  this.
  After some investigation, I think that this might deserve to be a
  blueprint
  but not a bug.
 
  We may need to resolve issues for the following cases:
  1) How to live migration with only one nova compute? (one nova compute
  can
  manage multiple clusters and there can be multi hosts in one cluster)
  2) Support live migration between clusters
  3) Support live migration between resource pools
  4) Support live migration between hosts
  5) Support live migration between cluster and host
  6) Support live migration between cluster and resource pool
  7) Support live migration between resource pool and host
  8) Might be more cases.
 
  Please show your comments if any and correct me if anything is not
  correct.
 
  --
  Thanks,
 
  Jay
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 



 --
 # Shawn.Hartsock - twitter: @hartsock - 

Re: [openstack-dev] [nova] An analysis of code review in Nova

2014-03-22 Thread Shawn Hartsock
On Sat, Mar 22, 2014 at 2:03 PM, Chris Behrens cbehr...@codestud.com wrote:
 I'd like to get spawn broken up sooner rather than later, personally. It has 
 additional benefits of being able to do better orchestration of builds from 
 conductor, etc.


I plan on delivery of that occurring as soon as or within a week after
Juno opens for commits. I understand we should see this in a week.
I've been holding on the refactor to avoid splitting reviewer
attention.

-- 
# Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] OpenStack/GSoC 2014 - Reviewing Applications

2014-03-22 Thread Sriram Subramanian
Davanum, are the scores private or is there a way to make them private?


On Sat, Mar 22, 2014 at 12:07 PM, Davanum Srinivas dava...@gmail.comwrote:

 Dear Students,

 We received the following proposals:

 AMES-Cloud (Nagashree Bhat)
 A pre-caching system for OpenStack (Anastasis Andronidis)
 Proposal for Implementing an application-level FWaaS driver
 ((Zorp) Dániel Csubák)
 Cassandra and Redis storage backend for Openstack marconi (Jeremy
 Henriques)
 Openstack-OSLO :Add a New Backend to Oslo.Cache (sai krishna)
 Implement a Fuzz testing framework that can be run on Tempest
 (Manishanker Talusani)
 How to detect network anomalies from telemetry data within Openstack
 (mst89)
 Implement a re-usable shared library for VMware(oslo.vmware) (Masaru
 Nomura)
 Cross-services Scheduler project (Artem Shepelev)
 OpenStack/Marconi: Py3k support (Nataliia)
 Improve benchmarking context mechanism in Rally (Kumar Rishabh)
 Develop a benchmarking suite and new storage backends to OpenStack
 Marconi (Prashanth Raghu)
 Adding Redis as a Storage Backend to OpenStack Marconi (Chenchong Qin)
 Auto Benchmarking System for OpenStack (RobberPhex)
 Developing Benchmarks for Virtual Machines of OpenStack with Rally
 (Tzanetos Balitsaris)
 Add a new storage backend to the OpenStack Message Queuing Service
 (Victoria Martínez de la Cruz)

 Dear Mentors
 (ddutta/flwang/julim/hughsaunders/greghaynes/annegentle/sriramhere/arnaudleg/coroner/boris_42/blint/ybudupi/cppcabrera),

 Please log into the Google GSoC web-site and review **all** proposals
 within a week (our own deadline - say 29th).

 1) Please click Wish to mentor on the left hand side, if you are
 willing to mentor this project.
Yes, we should have more than one mentor. This is a way to mention
 that you are willing to mentor,
actual assignment will happen later i believe. So please switch
 this on for all projects you are
interested in.

 2) If you have a question about a proposal and need a response from
 the candidate, please leave
a comment for them. If you want to ask a observation/question to
 other mentors or me, leave a
comment marked Private

 3) To assign a score. Please click on the number of start in My
 score: at the bottom
 5 = amazing applicant, could be a module maintainer on completing
 the program
 4 = strong applicant, will do a good job
 3 = good applicant, but is somewhat inexperienced
 2 = is unlikely to do a good job
 1 = not a good candidate

 Please feel free to leave detailed notes on candidates you know well
 for other mentors (Please mark comments as Private)

 If you are not able to help with this exercise, please let me know ASAP.

 If anyone else would like to step up and be a Mentor for any of these
 projects/students, please register in the Google GSoC site and let me
 know your username.

 Thanks,
 dims



 --
 Davanum Srinivas :: http://davanum.wordpress.com

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




-- 
Thanks,
-Sriram
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] OpenStack/GSoC 2014 - Reviewing Applications

2014-03-22 Thread Davanum Srinivas
Sriram,

We'll have to check with a student if there are able to see it during
this phase of the GSoC where mentors are evaluating the proposals.
There is some information/guidance here:
http://www.google-melange.com/document/show/gsoc_program/google/gsoc2011/userguide#depth_appreview
http://en.flossmanuals.net/melange/student-application-period/

-- dims

On Sat, Mar 22, 2014 at 6:34 PM, Sriram Subramanian
sri...@sriramhere.com wrote:
 Davanum, are the scores private or is there a way to make them private?


 On Sat, Mar 22, 2014 at 12:07 PM, Davanum Srinivas dava...@gmail.com
 wrote:

 Dear Students,

 We received the following proposals:

 AMES-Cloud (Nagashree Bhat)
 A pre-caching system for OpenStack (Anastasis Andronidis)
 Proposal for Implementing an application-level FWaaS driver
 ((Zorp) Dániel Csubák)
 Cassandra and Redis storage backend for Openstack marconi (Jeremy
 Henriques)
 Openstack-OSLO :Add a New Backend to Oslo.Cache (sai krishna)
 Implement a Fuzz testing framework that can be run on Tempest
 (Manishanker Talusani)
 How to detect network anomalies from telemetry data within Openstack
 (mst89)
 Implement a re-usable shared library for VMware(oslo.vmware) (Masaru
 Nomura)
 Cross-services Scheduler project (Artem Shepelev)
 OpenStack/Marconi: Py3k support (Nataliia)
 Improve benchmarking context mechanism in Rally (Kumar Rishabh)
 Develop a benchmarking suite and new storage backends to OpenStack
 Marconi (Prashanth Raghu)
 Adding Redis as a Storage Backend to OpenStack Marconi (Chenchong Qin)
 Auto Benchmarking System for OpenStack (RobberPhex)
 Developing Benchmarks for Virtual Machines of OpenStack with Rally
 (Tzanetos Balitsaris)
 Add a new storage backend to the OpenStack Message Queuing Service
 (Victoria Martínez de la Cruz)

 Dear Mentors
 (ddutta/flwang/julim/hughsaunders/greghaynes/annegentle/sriramhere/arnaudleg/coroner/boris_42/blint/ybudupi/cppcabrera),

 Please log into the Google GSoC web-site and review **all** proposals
 within a week (our own deadline - say 29th).

 1) Please click Wish to mentor on the left hand side, if you are
 willing to mentor this project.
Yes, we should have more than one mentor. This is a way to mention
 that you are willing to mentor,
actual assignment will happen later i believe. So please switch
 this on for all projects you are
interested in.

 2) If you have a question about a proposal and need a response from
 the candidate, please leave
a comment for them. If you want to ask a observation/question to
 other mentors or me, leave a
comment marked Private

 3) To assign a score. Please click on the number of start in My
 score: at the bottom
 5 = amazing applicant, could be a module maintainer on completing
 the program
 4 = strong applicant, will do a good job
 3 = good applicant, but is somewhat inexperienced
 2 = is unlikely to do a good job
 1 = not a good candidate

 Please feel free to leave detailed notes on candidates you know well
 for other mentors (Please mark comments as Private)

 If you are not able to help with this exercise, please let me know ASAP.

 If anyone else would like to step up and be a Mentor for any of these
 projects/students, please register in the Google GSoC site and let me
 know your username.

 Thanks,
 dims



 --
 Davanum Srinivas :: http://davanum.wordpress.com

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




 --
 Thanks,
 -Sriram

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




-- 
Davanum Srinivas :: http://davanum.wordpress.com

___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [nova] An analysis of code review in Nova

2014-03-22 Thread Shawn Hartsock
On Sat, Mar 22, 2014 at 11:55 AM, Matt Riedemann
mrie...@linux.vnet.ibm.com wrote:


 On 3/22/2014 5:19 AM, Shawn Hartsock wrote:

 On Fri, Mar 14, 2014 at 6:58 PM, Dan Smith d...@danplanet.com wrote:

 Review latency will be directly affected by how good the refactoring
 changes are staged. If they are small, on-topic and easy to validate,
 they will go quickly. They should be linearized unless there are some
 places where multiple sequences of changes make sense (i.e. refactoring
 a single file that results in no changes required to others).



 I'm going to bring this to the next
 https://wiki.openstack.org/wiki/Meetings/VMwareAPI we can start
 working on how we'll set the order for this kind of work. Currently we
 have a whole blueprint for refactoring a single method. That seems
 silly. I'll want to come up with a plan around how to restructure the
 driver so we can avoid some of the messes we've seen in the past.


 I think the point of starting with refactoring the nested method mess in the
 spawn method was it (1) seemed relatively trivial (think fast review
 turnaround) and (2) should be a good bang for the buck kind of change, as
 a lot of the original complaint was related to how hard it is to verify
 changes in the giant spawn method are tested - which you also point out
 below.



 I want to avoid one big refactor effort that drags on, but I also want
 to address bigger problems we have inside the driver. For example,


 I also want to avoid a big refactor effort dragging on, and I like the
 thinking on design changes, but are they doing going to be happening at the
 same time?  Or is the complete re-design going to supersede the refactoring?
 My only concern is biting off more than can be chewed in juno-1.

 Plus there is the refactor to use oslo.vmware, where does that fit into
 this?


 vm_util.py seems to have become burdened with work that it shouldn't
 have. It also performs a great number of unnecessary round trips using
 a vm_ref to pull individual vm details over one at a time. Introducing
 a VirtualMachine object that held all these references would simplify
 some operations (I'm not the first person to suggest this and it
 wasn't novel to me either when it was presented.)

 It would seem Juno-1 would be the time to make these changes and we
 need to serialize this work to keep reviewers from losing their
 marbles trying to track it all. I would like to work out a plan for
 this in conjunction with interested core-reviewers who would be
 willing to more or less sponsor this work. Because of the problems
 Matt points out, I don't want to tackle this in a haphazard or
 piece-meal way since it could completely disrupt any new blueprint
 work people may have targeted for Juno.


 Yeah, definitely need a plan here.  I'd like to see things prioritized based
 on what can be fixed in a relatively isolated way which gives a good return
 on coding/reviewing investment, e.g. pulling those nested methods out of
 spawn so they can be unit tested individually with mock rather than a large,
 seemingly rigid and scaffolded test framework.


Just because you said something about it :-)

I wrote this up off-the-cuff:
https://blueprints.launchpad.net/nova/+spec/vmware-vm-ref-refactor

This is about the size of each of these refactors that I'm thinking
of. I don't want to get much bigger than this and I don't want to
force a round-trip up to oslo.vmware and back to Nova for this size of
change. I don't really think we have to table the inclusion of
oslo.vmware integration in Juno in order to do this kind of work.
However, just as we pointed out at the top of the thread, there's a
danger that this kind of work can stall new feature adds so the timing
and inclusion of it is sensitive.



 Having said that, on this driver, new blueprints in the last several
 cycles have introduced serious feature regressions. Several minor bug
 fixes have altered and introduced key architectural components that
 have broken multiple critical features. In my professional opinion
 this has a root cause based on the drivers tightly coupled and
 non-cohesive internal design.

 The driver design is tightly coupled in that a change in one area
 forces multiple updates and changes *throughout* the rest of the
 driver. This is true in testing as well. The testing design often
 requires you to trace the entire codebase if you add a single optional
 parameter to a single method. This does not have to be true.


 Yup, this is my major complaint and ties into what I'm saying above, I find
 it really difficult to determine most of the time where a change is tested.
 Because of the nature of the driver code and my lack of actually writing
 features in it, as a reviewer I don't know if a change in X is going to
 break Y, so I rely on solid test coverage and the testing needs to be more
 natural to follow than it currently is.



 The driver design is non-cohesive in that important details and
 related information is spread throughout 

Re: [openstack-dev] [NOVA][VMWare][live-migration] VCDriver live migration problem

2014-03-22 Thread Jay Lau
Thanks Shawn, what you proposed is exactly I want ;-) Cool!

We can discuss more during the IRC meeting.

Thanks!


2014-03-22 20:22 GMT+08:00 Jay Lau jay.lau@gmail.com:

 Thanks Shawn, I have updated the title with VMWare.

 Yes, I know that live migration works. But the problem is when a cluster
 admin want to live migrate a VM instance, s/he will not know the target
 host where to migrate, as s/he cannot get target host from nova compute
 because currently VCDriver can only report cluster or resource pool as
 hypervisor host but not ESX server.

 IMHO, the VCDriver should support live migration between cluster, resource
 pool and ESX host, so we may need do at least the following enhancements:
 1) Enable live migration with even one nova compute. My current thinking
 is that enhance target host as host:node when live migrate a VM instance
 and the live migration task need
 2) Enable VCDriver report all ESX servers.

 We can discuss more during next week's IRC meeting.

 Thanks!


 2014-03-22 17:13 GMT+08:00 Shawn Hartsock harts...@acm.org:

 Hi Jay. We usually use [vmware] to tag discussion of VMware things. I
 almost didn't see this message.

 In short, there is a plan and we're currently blocked because we have
 to address several other pressing issues in the driver before we can
 address this one. Part of this is due to the fact that we can't press
 harder on blueprints or changes to the VCDriver right now.

 I actually reported this bug and we've discussed this at
 https://wiki.openstack.org/wiki/Meetings/VMwareAPI the basic problem
 is that live-migration actually works but you can't presently
 formulate a command that activates the feature from the CLI under some
 configurations. That's because of the introduction of clusters in the
 VCDriver in Havana.

 To fix this, we have to come up with a way to target a host inside the
 cluster (as I pointed out in the bug) or we have to have some way for
 a live migration to occur between clusters and a way to validate that
 this can happen first.

 As for the priority of this bug, it's been set to Medium which puts it
 well behind many of the Critical or High tasks on our radar. As for
 fixing the bug, no new outward behaviors or API are going to be
 introduced and this was working at one point and now it's stopped. To
 call this a new feature seems a bit strange.

 So, moving forward... perhaps we need to re-evaluate the priority
 order on some of these things. I tabled Juno planning during the last
 VMwareAPI subteam meeting but I plan on starting the discussion next
 week. We have a priority order for blueprints that we set as a team
 and these are publicly recorded in our meeting logs and on the wiki.
 I'll try to do better advertising these things. You are of course
 invited... and yeah... if you're interested in what we're fixing next
 in the VCDriver that next IRC meeting is where we'll start the
 discussion.

 On Sat, Mar 22, 2014 at 1:18 AM, Jay Lau jay.lau@gmail.com wrote:
  Hi,
 
  Currently we cannot do live migration with VCDriver in nova, live
 migration
  is really an important feature, so any plan to fix this?
 
  I noticed that there is already bug tracing this but seems no progress
 since
  last year's November: https://bugs.launchpad.net/nova/+bug/1192192
 
  Here just bring this problem up to see if there are any plan to fix
 this.
  After some investigation, I think that this might deserve to be a
 blueprint
  but not a bug.
 
  We may need to resolve issues for the following cases:
  1) How to live migration with only one nova compute? (one nova compute
 can
  manage multiple clusters and there can be multi hosts in one cluster)
  2) Support live migration between clusters
  3) Support live migration between resource pools
  4) Support live migration between hosts
  5) Support live migration between cluster and host
  6) Support live migration between cluster and resource pool
  7) Support live migration between resource pool and host
  8) Might be more cases.
 
  Please show your comments if any and correct me if anything is not
 correct.
 
  --
  Thanks,
 
  Jay
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 



 --
 # Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




 --
 Thanks,

 Jay




-- 
Thanks,

Jay
___
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


Re: [openstack-dev] [NOVA][VMWare][live-migration] VCDriver live migration problem

2014-03-22 Thread laserjetyang
I think we might need to discuss the VMware driver to refactor progress in
IRC meeting as well, and what is our overall plan. We have been keeping
seeing the vmware code broken.


On Sun, Mar 23, 2014 at 7:49 AM, Jay Lau jay.lau@gmail.com wrote:

 Thanks Shawn, what you proposed is exactly I want ;-) Cool!

 We can discuss more during the IRC meeting.

 Thanks!


 2014-03-22 20:22 GMT+08:00 Jay Lau jay.lau@gmail.com:

 Thanks Shawn, I have updated the title with VMWare.

 Yes, I know that live migration works. But the problem is when a
 cluster admin want to live migrate a VM instance, s/he will not know the
 target host where to migrate, as s/he cannot get target host from nova
 compute because currently VCDriver can only report cluster or resource pool
 as hypervisor host but not ESX server.

 IMHO, the VCDriver should support live migration between cluster,
 resource pool and ESX host, so we may need do at least the following
 enhancements:
 1) Enable live migration with even one nova compute. My current thinking
 is that enhance target host as host:node when live migrate a VM instance
 and the live migration task need
 2) Enable VCDriver report all ESX servers.

 We can discuss more during next week's IRC meeting.

 Thanks!


 2014-03-22 17:13 GMT+08:00 Shawn Hartsock harts...@acm.org:

 Hi Jay. We usually use [vmware] to tag discussion of VMware things. I
 almost didn't see this message.

 In short, there is a plan and we're currently blocked because we have
 to address several other pressing issues in the driver before we can
 address this one. Part of this is due to the fact that we can't press
 harder on blueprints or changes to the VCDriver right now.

 I actually reported this bug and we've discussed this at
 https://wiki.openstack.org/wiki/Meetings/VMwareAPI the basic problem
 is that live-migration actually works but you can't presently
 formulate a command that activates the feature from the CLI under some
 configurations. That's because of the introduction of clusters in the
 VCDriver in Havana.

 To fix this, we have to come up with a way to target a host inside the
 cluster (as I pointed out in the bug) or we have to have some way for
 a live migration to occur between clusters and a way to validate that
 this can happen first.

 As for the priority of this bug, it's been set to Medium which puts it
 well behind many of the Critical or High tasks on our radar. As for
 fixing the bug, no new outward behaviors or API are going to be
 introduced and this was working at one point and now it's stopped. To
 call this a new feature seems a bit strange.

 So, moving forward... perhaps we need to re-evaluate the priority
 order on some of these things. I tabled Juno planning during the last
 VMwareAPI subteam meeting but I plan on starting the discussion next
 week. We have a priority order for blueprints that we set as a team
 and these are publicly recorded in our meeting logs and on the wiki.
 I'll try to do better advertising these things. You are of course
 invited... and yeah... if you're interested in what we're fixing next
 in the VCDriver that next IRC meeting is where we'll start the
 discussion.

 On Sat, Mar 22, 2014 at 1:18 AM, Jay Lau jay.lau@gmail.com wrote:
  Hi,
 
  Currently we cannot do live migration with VCDriver in nova, live
 migration
  is really an important feature, so any plan to fix this?
 
  I noticed that there is already bug tracing this but seems no progress
 since
  last year's November: https://bugs.launchpad.net/nova/+bug/1192192
 
  Here just bring this problem up to see if there are any plan to fix
 this.
  After some investigation, I think that this might deserve to be a
 blueprint
  but not a bug.
 
  We may need to resolve issues for the following cases:
  1) How to live migration with only one nova compute? (one nova compute
 can
  manage multiple clusters and there can be multi hosts in one cluster)
  2) Support live migration between clusters
  3) Support live migration between resource pools
  4) Support live migration between hosts
  5) Support live migration between cluster and host
  6) Support live migration between cluster and resource pool
  7) Support live migration between resource pool and host
  8) Might be more cases.
 
  Please show your comments if any and correct me if anything is not
 correct.
 
  --
  Thanks,
 
  Jay
 
  ___
  OpenStack-dev mailing list
  OpenStack-dev@lists.openstack.org
  http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
 



 --
 # Shawn.Hartsock - twitter: @hartsock - plus.google.com/+ShawnHartsock

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org
 http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev




 --
 Thanks,

 Jay




 --
 Thanks,

 Jay

 ___
 OpenStack-dev mailing list
 OpenStack-dev@lists.openstack.org