oxy driver, i.e., something
that wraps a vendor-specific accelerator service and makes it friendly to a
multi-tenant OpenStack cloud? Quantum co-processors might be a compelling
example which fit this model.
Cheers,
On Sun., 20 May 2018, 23:28 Chris Friesen,
wrote:
> On 05/19/2018 05:58 PM, Bla
G'day Jay,
On 20 May 2018 at 08:37, Jay Pipes wrote:
> If it's not the VM or baremetal machine that is using the accelerator, what
> is?
It will be a VM or BM, but I don't think accelerators should be tied
to the life of a single instance if that isn't technically necessary
(i.e., they are hot-p
Relatively Cyborg-naive question here...
I thought Cyborg was going to support a hot-plug model. So I certainly
hope it is not the expectation that accelerators will be encoded into
Nova flavors? That will severely limit its usefulness.
On 19 May 2018 at 23:30, Jay Pipes wrote:
> On 05/18/2018 0
Please do not default to deleting it, otherwise someone will eventually be
back here asking why an irate user has just lost data. The better scenario
is that the rebuild will fail (early - before impact to the running
instance) with a quota error.
Cheers,
On Thu., 15 Mar. 2018, 00:46 Matt Riedema
+1!
It may also be worth testing a step where Nova & Neutron remain at N-1.
On 20 December 2017 at 04:58, Matt Riedemann wrote:
> During discussion in the TC channel today [1], we got talking about how
> there is a perception that you must upgrade all of the services together for
> anything to w
The former - we're running Cells so only have a single region currently
(except for Swift where we have multiple proxy endpoints around the
country, all backed by a global cluster, but they have to be different
regions to put them all in the service catalog). See
https://trello.com/b/9fkuT1eU/necta
On 14 December 2017 at 17:36, Clint Byrum wrote:
> The batch size for "upgrade the whole cloud" is too big. Let's help our
> users advance components one at a time, and then we won't have to worry
> so much about doing the whole integrated release dance so often.
Is there any data about how opera
Hi all - please note this conversation has been split variously across
-dev and -operators.
One small observation from the discussion so far is that it seems as
though there are two issues being discussed under the one banner:
1) maintain old releases for longer
2) do stable releases less frequent
I missed this session but the discussion strikes a chord as this is
something I've been saying on my user survey every 6 months.
On 11 November 2017 at 09:51, John Dickinson wrote:
> What I heard from ops in the room is that they want (to start) one release a
> year who's branch isn't deleted aft
A related bug that hasn't seen any love for some time:
https://bugs.launchpad.net/nova/+bug/1613199
On 6 October 2017 at 07:47, James Penick wrote:
> Hey Pino,
>
> mriedem pointed me to the vendordata code [1] which shows some fields are
> passed (such as project ID) and that SSL is supported. S
On 29 September 2017 at 22:26, Bob Ball wrote:
> The concepts of PCI and SR-IOV are, of course, generic, but I think out of
> principal we should avoid a hypervisor-specific integration for vGPU (indeed
> Citrix has been clear from the beginning that the vGPU integration we are
> proposing is i
Hi Prema
On 28 September 2017 at 07:10, Premysl Kouril wrote:
> Hi, I work with Jakub (the op of this thread) and here is my two
> cents: I think what is critical to realize is that KVM virtual
> machines can have substantial memory overhead of up to 25% of memory,
> allocated to KVM virtual mach
On 27 September 2017 at 23:19, Jakub Jursa wrote:
> 'hw:cpu_policy=dedicated' (while NOT setting 'hw:numa_nodes') results in
> libvirt pinning CPU in 'strict' memory mode
>
> (from libvirt xml for given instance)
> ...
>
>
>
>
> ...
>
> So yeah, the instance is not able to allocate
On 27 September 2017 at 22:40, Belmiro Moreira
wrote:
> In the past we used the tabs but latest Horizon versions use the visibility
> column/search instead.
> The issue is that we would like the old images to continue to be
> discoverable by everyone and have a image list that only shows the lates
Also CC-ing os-ops as someone else may have encountered this before
and have further/better advice...
On 27 September 2017 at 18:40, Blair Bethwaite
wrote:
> On 27 September 2017 at 18:14, Stephen Finucane wrote:
>> What you're probably looking for is the 'reserved_hos
On 27 September 2017 at 18:14, Stephen Finucane wrote:
> What you're probably looking for is the 'reserved_host_memory_mb' option. This
> defaults to 512 (at least in the latest master) so if you up this to 4192 or
> similar you should resolve the issue.
I don't see how this would help given the
Hi Belmiro,
On 20 Sep. 2017 7:58 pm, "Belmiro Moreira" <
moreira.belmiro.email.li...@gmail.com> wrote:
> Discovering the latest image release is hard. So we added an image
property "recommended"
> that we update when a new image release is available. Also, we patched
horizon to show
> the "recomme
I've been watching this thread and I think we've already seen an
excellent and uncontroversial suggestion towards simplifying initial
deployment of OS - that was to push towards encoding Constellations
into the deployment and/or config management projects.
On 26 September 2017 at 15:44, Adam Lawso
Please don't make these 400s - it should not be a client error to be
unaware of the service status ahead of time.
On 12 July 2017 at 11:18, Matt Riedemann wrote:
> I'm looking for some broader input on something being discussed in this
> change:
>
> https://review.openstack.org/#/c/464280/21/nova
On 27 June 2017 at 23:47, Sean Dague wrote:
> I still think I've missed, or not grasped, during this thread how a SIG
> functions differently than a WG, besides name. Both in theory and practice.
I think for the most part SIG is just a more fitting moniker for some
of these groups. E.g. I would s
On 27 June 2017 at 23:47, Thierry Carrez wrote:
> Setting up a common ML for common discussions (openstack-sigs) will
> really help, even if there will be some pain setting them up and getting
> the right readership to them :)
It's worth a try! I agree it will probably take a while, and I would
s
Resend for openstack-dev with proper list perms...
-- Forwarded message --
From: Blair Bethwaite
Date: 27 June 2017 at 23:24
Subject: [scientific] IRC Meeting (Tues 2100 UTC): Science app catalogues,
network security of research computing on OpenStack
To: user-committee
There is a not insignificant degree of irony in the fact that this
conversation has splintered so that anyone only reading openstack-operators
and/or user-committee is missing 90% of the picture Maybe I just need a
new ML management strategy.
I'd like to add a +1 to Sean's suggestion about WG/
Hi Alex,
On 2 June 2017 at 23:13, Alexandra Settle wrote:
> O I like your thinking – I’m a pandoc fan, so, I’d be interested in
> moving this along using any tools to make it easier.
I can't realistically offer much time on this but I would be happy to
help (ad-hoc) review/catalog/clean-up i
Hi Alex,
Likewise for option 3. If I recall correctly from the summit session
that was also the main preference in the room?
On 2 June 2017 at 11:15, George Mihaiescu wrote:
> +1 for option 3
>
>
>
> On Jun 1, 2017, at 11:06, Alexandra Settle wrote:
>
> Hi everyone,
>
>
>
> I haven’t had any fe
Hey all,
Hopefully you've all noticed this by now, the timing of the WG
sessions (lightening talks, meeting, BoF) has changed a little since
first published. I've just updated the etherpad to reflect that now:
https://etherpad.openstack.org/p/Scientific-WG-boston
Tues 11:15am - 11:55am - Scientif
Morning all -
Apologies for the shotgun email. But looks like we still have one or
two spots available for lightening talks if anyone has work they want
to share and/or discuss:
https://etherpad.openstack.org/p/Scientific-WG-Boston-Lightning
Best regards,
On 28 April 2017 at 06:19, George Mihai
-- Forwarded message --
From: Blair Bethwaite
Date: 6 May 2017 at 17:55
Subject: GPU passthrough success and failure records
To: "openstack-oper."
Hi all,
I've been (very slowly) working on some docs detailing how to setup an
OpenStack Nova Libvirt+QEMU-KV
On 2 May 2017 at 05:50, Jay Pipes wrote:
> Masahito Muroi is currently marked as the moderator, but I will indeed be
> there and happy to assist Masahito in moderating, no problem.
The more the merrier :-).
There is a rather unfortunate clash here with the Scientific-WG BoF
session. Location per
> allocations in the future).
>
> A separate system (hopefully Blazar) is needed to manage the time-based
> associations to inventories of resources over a period in the future.
>
> Best,
> -jay
>
>>> I'm not sure how the above is different from the constrain
Thanks Rochelle. I encourage everyone to dump thoughts into the
etherpad (https://etherpad.openstack.org/p/BOS-forum-special-hardware
- feel free to garden it as you go!) so we can have some chance of
organising a coherent session. In particular it would be useful to
know what is going to be most u
On 29 April 2017 at 01:46, Mike Dorman wrote:
> I don’t disagree with you that the client side choose-a-server-at-random is
> not a great load balancer. (But isn’t this roughly the same thing that
> oslo-messaging does when we give it a list of RMQ servers?) For us it’s more
> about the failu
We at Nectar are in the same boat as Mike. Our use-case is a little
bit more about geo-distributed operations though - our Cells are in
different States around the country, so the local glance-apis are
particularly important for caching popular images close to the
nova-computes. We consider these g
Hi all,
A quick FYI that this Forum session exists:
https://www.openstack.org/summit/boston-2017/summit-schedule/events/18803/special-hardware
(https://etherpad.openstack.org/p/BOS-forum-special-hardware) is a
thing this Forum.
It would be great to see a good representation from both the Nova and
Devil's advocate - what is "full enough"? Surely another channel is
essentially free and having flexibility in available timing is of utmost
importance?
On 8 Nov 2016 5:37 PM, "Tony Breeds" wrote:
> On Mon, Nov 07, 2016 at 05:52:43PM +0100, lebre.adr...@free.fr wrote:
> > Dear all,
> >
> > We ar
Hi Nikhil,
2000UTC might catch a few kiwis, but it's 6am everywhere on the east
coast of Australia, and even earlier out west. 0800UTC, on the other
hand, would be more sociable.
On 26 May 2016 at 15:30, Nikhil Komawar wrote:
> Thanks Sam. We purposefully chose that time to accommodate some of o
Hi all,
I've just been doing some user consultation and pondering a case for
use of the Qemu Guest Agent in order to get quiesced backups.
In doing so I found myself wondering why on earth I need to set an
image property in Glance (hw_qemu_guest_agent) to toggle such a thing
for any particular in
On 20 November 2014 05:25, wrote:
> --
>
> Message: 24
> Date: Wed, 19 Nov 2014 10:57:17 -0500
> From: Doug Hellmann
> To: "OpenStack Development Mailing List (not for usage questions)"
>
> Subject: Re: [openstack-dev] Quota management and enforcement across
Hi there,
We've been investigating some guest filesystem issues recently and noticed
what looks like a slight inconsistency in base image handling in
block-migration. We're on Grizzly from the associated Ubuntu cloud archive
and using qcow on local storage.
What we've noticed is that after block-
39 matches
Mail list logo