I think I understand your confusing Justin. Extensions are not there to bind APIs together. The examples I gave were probably a bit misleading. Extensions are there to support niche functionality and to allow developers to innovate without having to wait for some centralized group to approve.
You're right, things should become clearer as we move towards code :-) -jOrGe W. On Feb 18, 2011, at 5:08 PM, Justin Santa Barbara wrote: I find this even more confusing than before. On the one hand, we talk about a suite of independent APIs, and on the other hand we talk about binding them together using extensions. We talk about standardizing around one API, and we talk about letting a thousand flowers bloom as extensions. I'm going to wait till there's concrete code here before commenting further, I think, so that we can talk in specifics. Justin On Fri, Feb 18, 2011 at 2:32 PM, Erik Carlin <erik.car...@rackspace.com<mailto:erik.car...@rackspace.com>> wrote: The way I see it, there isn't a singular OpenStack API (even today there is swift, nova, and glance). OpenStack is a suite of IaaS each with their own API – so there is a SUITE of standard OS APIs. And each OS service should strive to define the canonical API for automating that particular service. If I just want to run an image repo, I deploy glance. If my SAN guy can't get storage provisioned fast enough, I deploy the OS block storage service (once we have it). And if I want a full cloud suite, I deploy all the services. They are loosely coupled and (ideally) independent building blocks. Whether one chooses to front the different service endpoints with a proxy to unify them or have separate service endpoints is purely a deployment decision. Either way, there are no competing OS APIs. Support for 3rd party APIs (e.g. EC2) is secondary IMO, and to some degree, detrimental. Standards are defined largely in part by ubiquity. We want OS to become ubiquitous and we want the OS APIs to become defacto. Supporting additional APIs (or even variations of the same API like AMQP per the other thread) doesn't help us here. I would love to see the community rally behind a per service standard OS REST API that we can own and drive. To that end, the goal as I see it is to launch canonical OpenStack Compute (nova) and Image (glance) APIs with Cactus. In Diablo, we would then work to introduce separate network and block storage services with REST APIs as well. All APIs would be independently versioned and stable. I'm ALL for per language OpenStack bindings that implement support for the entire suite of services. Re: extensions, it's actually the technical aspects that are driving it. There is a tension between standards and innovation that needs to be resolved. In addition, we need to be able to support niche functionality (e.g. Rackspace may want to support API operations related to managed services) without imposing it on everyone. These problems are not new. We've seen the same exact thing with OpenGL and they have a very successful extension model that has solved this. Jorge studied this when did his PhD and has designed extensions with that in mind. He has a presentation on extensions here if you haven't seen it. I think extensions are critically important and would encourage dialog amongst the community to come to a consensus on this. Per my points above, I would prefer to avoid separate APIs for the same service. Let's see if we can get behind a per service API that becomes THE defacto standard way for automating that service. Erik From: Justin Santa Barbara <jus...@fathomdb.com<mailto:jus...@fathomdb.com>> Date: Fri, 18 Feb 2011 09:57:12 -0800 To: Paul Voccio <paul.voc...@rackspace.com<mailto:paul.voc...@rackspace.com>> Cc: "openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net>" <openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net>> Subject: Re: [Openstack] OpenStack Compute API 1.1 > How is the 1.1 api proposal breaking this? Because if we launch an OpenStack API, the expectation is that this will be the OpenStack API :-) If we support a third-party API (CloudServers or EC2), then people will continue to use their existing wrappers (e.g. jclouds) Once there's an OpenStack API, then end-users will want to find a library for that, and we don't want that to be a poor experience. To maintain a good experience, we either can't break the API, or we need to write and maintain a lot of proxying code to maintain compatibility. We know we're not ready for the first commitment, and I don't think we get enough to justify the second. > I think the proxy would make sense if you wanted to have a single api. Not > all service providers will but I see this as entirely optional, not required > to use the services. But then we have two OpenStack APIs? Our ultimate end users don't use the API, they use a wrapper library. They want a stable library that works and is kept up to date with recent changes and don't care about what's going on under the covers. Wrapper library authors want an API that is (1) one API and (2) stable with reasonable evolution, otherwise they'll abandon their wrapper or not update it. > The extensions mechanism is the biggest change, iirc. I'm not a big fan of the extensions idea, because it feels more like a reflection of a management goal, rather than a technical decision ("OpenStack is open to extensions") Supporting separate APIs feels like a better way to do that. I'm very open to be corrected here, but I think we need to see code that wants to use the extension API and isn't better done as a separate API. Right now I haven't seen any patches, and that makes me uneasy. On Fri, Feb 18, 2011 at 9:29 AM, Paul Voccio <paul.voc...@rackspace.com<mailto:paul.voc...@rackspace.com>> wrote: The spec for 1.0 and 1.1 are pretty close. The extensions mechanism is the biggest change, iirc. I think the proxy would make sense if you wanted to have a single api. Not all service providers will but I see this as entirely optional, not required to use the services. The push to get a completed compute api is the desire move away from the ec2 api to something that we can guide, extend and vote on as a community. The sooner we do the the better. How is the 1.1 api proposal breaking this? From: Justin Santa Barbara <jus...@fathomdb.com<mailto:jus...@fathomdb.com>> Date: Fri, 18 Feb 2011 09:10:19 -0800 To: Paul Voccio <paul.voc...@rackspace.com<mailto:paul.voc...@rackspace.com>> Cc: Jay Pipes <jaypi...@gmail.com<mailto:jaypi...@gmail.com>>, "openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net>" <openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net>> Subject: Re: [Openstack] OpenStack Compute API 1.1 Jay: The AMQP->REST was the re-architecting I was referring to, which would not be customer-facing (other than likely introducing new bugs.) Spinning off the services, if this is visible at the API level, is much more concerning to me. So Paul, I think the proxy is good because it acknowledges the importance of keeping a consistent API. But - if our API isn't finalized - why push it out at all, particularly if we're then going to have the overhead of maintaining another translation layer? For Cactus, let's just support EC2 and/or CloudServers 1.0 API compatibility (again a translation layer, but one we probably have to support anyway.) Then we can design the right OpenStack API at our leisure and meet all of our goals: a stable Cactus and stable APIs. If anyone ends up coding to a Cactus OpenStack API, we shouldn't have them become second-class citizens 3 months later. Justin On Fri, Feb 18, 2011 at 6:31 AM, Paul Voccio <paul.voc...@rackspace.com<mailto:paul.voc...@rackspace.com>> wrote: Jay, I understand Justin's concern if we move /network and /images and /volume to their own endpoints then it would be a change to the customer. I think this could be solved by putting a proxy in front of each endpoint and routing back to the appropriate service endpoint. I added another image on the wiki page to describe what I'm trying to say. http://wiki.openstack.org/api_transition I think might not be as bad of a transition since the compute worker would receive a request for a new compute node then it would proxy over to the admin or public api of the network or volume node to request information. It would work very similar to how the queues work now. pvo On 2/17/11 8:33 PM, "Jay Pipes" <jaypi...@gmail.com<mailto:jaypi...@gmail.com>> wrote: >Sorry, I don't view the proposed changes from AMQP to REST as being >"customer facing API changes". Could you explain? These are internal >interfaces, no? > >-jay > >On Thu, Feb 17, 2011 at 8:13 PM, Justin Santa Barbara ><jus...@fathomdb.com<mailto:jus...@fathomdb.com>> wrote: >> An API is for life, not just for Cactus. >> I agree that stability is important. I don't see how we can claim to >> deliver 'stability' when the plan is then immediately to destablize >> everything with a very disruptive change soon after, including customer >> facing API changes and massive internal re-architecting. >> >> >> On Thu, Feb 17, 2011 at 4:18 PM, Jay Pipes >> <jaypi...@gmail.com<mailto:jaypi...@gmail.com>> wrote: >>> >>> On Thu, Feb 17, 2011 at 6:57 PM, Justin Santa Barbara >>> <jus...@fathomdb.com<mailto:jus...@fathomdb.com>> wrote: >>> > Pulling volumes & images out into separate services (and moving from >>> > AMQP to >>> > REST) sounds like a huge breaking change, so if that is indeed the >>>plan, >>> > let's do that asap (i.e. Cactus). >>> >>> Sorry, I have to disagree with you here, Justin :) The Cactus release >>> is supposed to be about stability and the only features going into >>> Cactus should be to achieve API parity of the OpenStack Compute API >>> with the Rackspace Cloud Servers API. Doing such a huge change like >>> moving communication from AMQP to HTTP for volume and network would be >>> a change that would likely undermine the stability of the Cactus >>> release severely. >>> >>> -jay >> >> Confidentiality Notice: This e-mail message (including any attached or embedded documents) is intended for the exclusive and confidential use of the individual or entity to which this message is addressed, and unless otherwise expressly indicated, is confidential and privileged information of Rackspace. Any dissemination, distribution or copying of the enclosed material is prohibited. If you receive this transmission in error, please notify us immediately by e-mail at ab...@rackspace.com<mailto:ab...@rackspace.com>, and delete the original message. Your cooperation is appreciated. Confidentiality Notice: This e-mail message (including any attached or embedded documents) is intended for the exclusive and confidential use of the individual or entity to which this message is addressed, and unless otherwise expressly indicated, is confidential and privileged information of Rackspace. Any dissemination, distribution or copying of the enclosed material is prohibited. If you receive this transmission in error, please notify us immediately by e-mail at ab...@rackspace.com<mailto:ab...@rackspace.com>, and delete the original message. Your cooperation is appreciated. _______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net> Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp _______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net<mailto:openstack@lists.launchpad.net> Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp
_______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp