Re: [openstack-dev] [Murano] Nominations to Murano core
Thank you guys! I'll do my best to continue to provide valuable reviews and make quality contributions to Murano! On Fri, Jun 27, 2014 at 5:47 PM, Ruslan Kamaldinov rkamaldi...@mirantis.com wrote: Serg, Steve, welcome to Murano core team! -- Ruslan On Thu, Jun 26, 2014 at 8:06 PM, Georgy Okrokvertskhov gokrokvertsk...@mirantis.com wrote: +1 for both! Thanks for the great work! Regards, Gosha On Thu, Jun 26, 2014 at 7:05 AM, Timur Sufiev tsuf...@mirantis.com wrote: +1 for both. On Thu, Jun 26, 2014 at 3:29 PM, Stan Lagun sla...@mirantis.com wrote: +1 for both (or should I say +2?) Sincerely yours, Stan Lagun Principal Software Engineer @ Mirantis On Thu, Jun 26, 2014 at 1:49 PM, Alexander Tivelkov ativel...@mirantis.com wrote: +1 on both Serge and Steve -- Regards, Alexander Tivelkov On Thu, Jun 26, 2014 at 1:37 PM, Ruslan Kamaldinov rkamaldi...@mirantis.com wrote: I would like to nominate Serg Melikyan and Steve McLellan to Murano core. Serge has been a significant reviewer in the Icehouse and Juno release cycles. Steve has been providing consistent quality reviews and they continue to get more frequent and better over time. Thanks, Ruslan ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Timur Sufiev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Georgy Okrokvertskhov Architect, OpenStack Platform Products, Mirantis http://www.mirantis.com Tel. +1 650 963 9828 Mob. +1 650 996 3284 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Serg Melikyan, Senior Software Engineer at Mirantis, Inc. http://mirantis.com | smelik...@mirantis.com +7 (495) 640-4904, 0261 +7 (903) 156-0836 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] 'retry' option
On 28/06/14 22:49 +0100, Mark McLoughlin wrote: On Fri, 2014-06-27 at 17:02 +0100, Gordon Sim wrote: A question about the new 'retry' option. The doc says: By default, cast() and call() will block until the message is successfully sent. What does 'successfully sent' mean here? Unclear, ambiguous, probably driver dependent etc. The 'blocking' we're talking about here is establishing a connection with the broker. If the connection has been lost, then cast() will block until the connection has been re-established and the message 'sent'. Does it mean 'written to the wire' or 'accepted by the broker'? For the impl_qpid.py driver, each send is synchronous, so it means accepted by the broker[1]. What does the impl_rabbit.py driver do? Does it just mean 'written to the wire', or is it using RabbitMQ confirmations to get notified when the broker accepts it (standard 0-9-1 has no way of doing this). I don't know, but it would be nice if someone did take the time to figure it out and document it :) Seriously, some docs around the subtle ways that the drivers differ from one another would be helpful ... particularly if it exposed incorrect assumptions API users are currently making. +1 We should also keep this in mind for the amqp 1.0 driver. It's mandatory to have a detailed documentation of it. Specifically, I'm interested in having documented how its architecture differs from other drivers. The more we can write down about it, the better. If the intention is to block until accepted by the broker that has obvious performance implications. On the other hand if it means block until written to the wire, what is the advantage of that? Was that a deliberate feature or perhaps just an accident of implementation? The use case for the new parameter, as described in the git commit, seems to be motivated by wanting to avoid the blocking when sending notifications. I can certainly understand that desire. However, notifications and casts feel like inherently asynchronous things to me, and perhaps having/needing the synchronous behaviour is the real issue? It's not so much about sync vs async, but a failure mode. By default, if we lose our connection with the broker, we wait until we can re-establish it rather than throwing exceptions (requiring the API caller to have its own retry logic) or quietly dropping the message. The use case for ceilometer is to allow its RPCPublisher to have a publishing policy - block until the samples have been sent, queue (in an in-memory, fixed-length queue) if we don't have a connection to the broker, or drop it if we don't have a connection to the broker. https://review.openstack.org/77845 I do understand the ambiguity around what message delivery guarantees are implicit in cast() isn't ideal, but that's not what adding this 'retry' parameter was about. Besides documenting what each driver does, it might also be useful to document what the expected guarantees are. The way you just described them, in terms of connections, sound quite reasonable to me. Calls by contrast, are inherently synchronous, but at present the retry controls only the sending of the request. If the server fails, the call may timeout regardless of the value of 'retry'. Just in passing, I'd suggest that renaming the new parameter max_reconnects, would make it's current behaviour and values clearer. The name 'retry' sounds like a yes/no type value, and retry=0 v. retry=1 is the reverse of what I would intuitively expect. Sounds reasonable. Would you like to submit a patch? Quick turnaround is important, because if Ceilometer starts using this retry parameter before we rename it, I'm not sure it'll be worth the hassle. +1 Flavio -- @flaper87 Flavio Percoco pgpSNaEPjQ0Ig.pgp Description: PGP signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Jenkins] [Cinder] InvocationError in gate-cinder-python26 python27
Hi Stackers, I have been facing below issues at gate-cinder-python26 gate-cinder-python27 after uploading my patch. I assume this to be an infrastructure issue than an issue with my patch. Can someone please confirm ? . 2014-06-30 05:41:57.704 | check_uptodate.sh: cinder.conf.sample is not up to date. 2014-06-30 05:41:57.704 | check_uptodate.sh: Please run /home/jenkins/workspace/gate-cinder-python26/tools/config/generate_sample.sh. 2014-06-30 05:41:57.705 | ERROR: InvocationError: '/home/jenkins/workspace/gate-cinder-python26/tools/config/check_uptodate.sh' 2014-06-30 05:41:57.705 | ___ summary 2014-06-30 05:41:57.705 | ERROR: py26: commands failed 2014-06-30 05:41:57.707 | + result=1 2014-06-30 05:41:57.708 | + echo 'Begin pip freeze output from test virtualenv:' 2014-06-30 05:41:57.708 | Begin pip freeze output from test virtualenv: 2014-06-30 05:41:57.708 | + echo == 2014-06-30 05:41:57.708 | == 2014-06-30 05:41:57.708 | + .tox/py26/bin/pip freeze . This is my patch url - https://review.openstack.org/#/c/102511/ I have even raised a bug against this - https://bugs.launchpad.net/openstack-ci/+bug/1334623 Regards, Amit *CloudByte Inc.* http://www.cloudbyte.com/ ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Jenkins] [Cinder] InvocationError in gate-cinder-python26 python27
On 30/06/14 16:37, Amit Das wrote: I have been facing below issues at gate-cinder-python26 gate-cinder-python27 after uploading my patch. I assume this to be an infrastructure issue than an issue with my patch. Can someone please confirm ? . 2014-06-30 05:41:57.704 | check_uptodate.sh: cinder.conf.sample is not up to date. This usually means that Oslo (the common libraries used by many projects) has added an configuration option. It's in fact a problem with cinder, in that they need to make sure the sample configuration file is up to date, but it will affect all patches until it is fixed. I'm sure it will get sorted out quickly. Personally, I think generating and comparing a sample config every build is daft, and a sample configuration should be generated during sdist or something. Cheers, -- Steve Stop breathing down my neck! My breathing is merely a simulation. So is my neck! Stop it anyway. - EMH vs EMH, USS Prometheus ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] Performance of security group
Yes, the usage of fanout topic by VNI is also another big improvement we could do. That will fit perfectly for the l2-pop mechanism driver. Of course, that need a specific call on a start/re-sync to get initial state. That actually done by the l2-pop MD if the uptime of an agent is less than 'agent_boot_time' flag [1]. [1] https://github.com/openstack/neutron/blob/master/neutron/plugins/ml2/drivers/l2pop/mech_driver.py#L181 Édouard. On Fri, Jun 27, 2014 at 3:43 AM, joehuang joehu...@huawei.com wrote: Interesting idea to optimize the performance. Not only security group rule will leads to fanout message load, we need to review and check to see if all fanout usegae in Neutron could be optimized. For example, L2 population: self.fanout_cast(context, self.make_msg(method, fdb_entries=fdb_entries), topic=self.topic_l2pop_update) it would be better to use network+l2pop_update as the topic, and only the agents which there are VMs running on it will consume the message. Best Regards Chaoyi Huang( Joe Huang) -邮件原件- 发件人: Miguel Angel Ajo Pelayo [mailto:mangel...@redhat.com] 发送时间: 2014年6月27日 1:33 收件人: OpenStack Development Mailing List (not for usage questions) 主题: Re: [openstack-dev] [neutron]Performance of security group - Original Message - @Nachi: Yes that could a good improvement to factorize the RPC mechanism. Another idea: What about creating a RPC topic per security group (quid of the RPC topic scalability) on which an agent subscribes if one of its ports is associated to the security group? Regards, Édouard. Hmm, Interesting, @Nachi, I'm not sure I fully understood: SG_LIST [ SG1, SG2] SG_RULE_LIST = [SG_Rule1, SG_Rule2] .. port[SG_ID1, SG_ID2], port2 , port3 Probably we may need to include also the SG_IP_LIST = [SG_IP1, SG_IP2] ... and let the agent do all the combination work. Something like this could make sense? Security_Groups = {SG1:{IPs:[],RULES:[], SG2:{IPs:[],RULES:[]} } Ports = {Port1:[SG1, SG2], Port2: [SG1] } @Edouard, actually I like the idea of having the agent subscribed to security groups they have ports on... That would remove the need to include all the security groups information on every call... But would need another call to get the full information of a set of security groups at start/resync if we don't already have any. On Fri, Jun 20, 2014 at 4:04 AM, shihanzhang ayshihanzh...@126.com wrote: hi Miguel Ángel, I am very agree with you about the following point: * physical implementation on the hosts (ipsets, nftables, ... ) --this can reduce the load of compute node. * rpc communication mechanisms. -- this can reduce the load of neutron server can you help me to review my BP specs? At 2014-06-19 10:11:34, Miguel Angel Ajo Pelayo mangel...@redhat.com wrote: Hi it's a very interesting topic, I was getting ready to raise the same concerns about our security groups implementation, shihanzhang thank you for starting this topic. Not only at low level where (with our default security group rules -allow all incoming from 'default' sg- the iptable rules will grow in ~X^2 for a tenant, and, the security_group_rules_for_devices rpc call from ovs-agent to neutron-server grows to message sizes of 100MB, generating serious scalability issues or timeouts/retries that totally break neutron service. (example trace of that RPC call with a few instances http://www.fpaste.org/104401/14008522/ ) I believe that we also need to review the RPC calling mechanism for the OVS agent here, there are several possible approaches to breaking down (or/and CIDR compressing) the information we return via this api call. So we have to look at two things here: * physical implementation on the hosts (ipsets, nftables, ... ) * rpc communication mechanisms. Best regards, Miguel Ángel. - Mensaje original - Do you though about nftables that will replace {ip,ip6,arp,eb}tables? It also based on the rule set mechanism. The issue in that proposition, it's only stable since the begin of the year and on Linux kernel 3.13. But there lot of pros I don't list here (leverage iptables limitation, efficient update rule, rule set, standardization of netfilter commands...). Édouard. On Thu, Jun 19, 2014 at 8:25 AM, henry hly henry4...@gmail.com wrote: we have done some tests, but have different result: the performance is nearly the same for empty and 5k rules in iptable, but huge gap between enable/disable iptable hook on linux bridge On Thu, Jun 19, 2014 at 11:21 AM, shihanzhang ayshihanzh...@126.com wrote: Now I have not get accurate test data, but I can confirm the following points: 1. In compute
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
On Sat, Jun 28, 2014 at 08:26:44AM -0500, Matt Riedemann wrote: On 6/27/2014 7:35 AM, Daniel P. Berrange wrote: On Fri, Jun 27, 2014 at 07:40:51AM -0400, Sean Dague wrote: It's clear that lots of projects want 3rd Party CI information on patches. But it's also clear that 6 months into this experiment with a lot of 3rd Party CI systems, the Gerrit UI is really not great for this. That's an understatement about the UI :-) It seems what we actually want is a dashboard of these results. We want them available when we go to Gerrit, but we don't want them in Gerrit itself. What if 3rd Party CI didn't vote in Gerrit? What if it instead published to some 3rd party test reporting site (a thing that doesn't yet exist). Gerrit has the facility so that we could inject the dashboard content for this in Gerrit in a little table somewhere, but the data would fundamentally live outside of Gerrit. It would also mean that all the aggregate reporting of 3rd Party CI that's being done in custom gerrit scripts, could be integrated directly into such a thing. Agreed, it would be a great improvement in usability if we stopped all CI systems, including our default Jenkins, from ever commenting on reviews. At most gating CIs should +1/-1. Having a table of results displayed, pulling the data from an external result tracking system would be a great idea. Even better if this external system had a nice button you can press to trigger re-check, so we can stop using comments for that too. I would disagree with this idea since it's equivalent to 'recheck no bug' and that's naughty, because then we don't track race bugs as well. It could easily have a text field for entering a bug number. The point is to stop adding comments to gerrit that aren't related to code review. Regards, Daniel -- |: http://berrange.com -o-http://www.flickr.com/photos/dberrange/ :| |: http://libvirt.org -o- http://virt-manager.org :| |: http://autobuild.org -o- http://search.cpan.org/~danberr/ :| |: http://entangle-photo.org -o- http://live.gnome.org/gtk-vnc :| ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [neutron][third-party] Simple and robust CI script?
Howdy! Paging other 3rd party CI operators... I would like to run a simple and robust 3rd party CI. Simple as in a small number of moving parts, robust as in unlikely to make mistakes due to unexpected problems. I'm imagining: - 100 lines of shell for the implementation. - Minimum of daemons. (Just Jenkins? Ideally not even that...) - Robust detection of operational problems (CI bugs) vs system-under-test problems (changes that should be voted against). - Effective notifications on all errors or negative votes. Does anybody already have an implementation like this that they would like to share? (Is anybody else wanting something in this direction?) I've been iterating on 3rd party CI mechanisms (currently onto my 3rd from-scratch setup) and I have not been completely satisfied with any of them. I would love to hear from someone who has a minimalist implementation that they are happy with :). Cheers, -Luke ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] Use MariaDB by default on Fedora
I am trying to finish off https://review.openstack.org/#/c/90134 - percona xtradb cluster for debian based system. I have read into this thread that I can error out on Redhat systems when trying to install percona and tell them to use mariadb instead, percona isn't support here. Is this correct? Michael On Friday 27 June 2014 16:49:58 James Slagle wrote: On Fri, Jun 27, 2014 at 4:13 PM, Clint Byrum cl...@fewbar.com wrote: Excerpts from James Slagle's message of 2014-06-27 12:59:36 -0700: Things are a bit confusing right now, especially with what's been proposed. Let me try and clarify (even if just for my own sake). Currently the choices offered are: 1. mysql percona with the percona tarball Percona Xtradb Cluster, not mysql percona 2. mariadb galera with mariadb.org packages 3. mariadb galera with rdo packages And, we're proposing to add 4. mysql percona with percona packages: https://review.openstack.org/#/c/90134 5. mariadb galera with fedora packages https://review.openstack.org/#/c/102815/ 4 replaces 1, but only for Ubuntu/Debian, it doesn't work on Fedora/RH 5 replaces 3 (neither of which work on Ubuntu/Debian, obviously) Do we still need 1? Fedora/RH + percona tarball. I personally don't think so. Do we still need 2? Fedora/RH or Ubuntu/Debian with galera packages from maraidb.org. For the Fedora/RH case, I doubt it, people will just use 5. 3 will be gone (replaced by 5). So, yes, I'd like to see 5 as the default for Fedora/RH and 4 as the default for Ubuntu/Debian, and both those tested in CI. And get rid of (or deprecate) 1-3. I'm actually more confused now than before I read this. The use of numbers is just making my head spin. There are 5 choices, some of which are not totally clear. Hence the need to clean things up. It can be stated this way I think: On RPM systems, use MariaDB Galera packages. If packages are in the distro, use distro packages. If packages are not in the distro, use RDO packages. There won't be a need to install from the RDO repositories. Mariadb galera packages are in the main Fedora package repositories, and for RHEL, they're in the epel repositories. On DEB systems, use Percona XtraDB Cluster packages. If packages are in the distro, use distro packages. If packages are not in the distro, use upstream packages. If anything doesn't match those principles, it is a bug. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [cinder]a problem about the implement of limit-volume-copy-bandwidth
Hi stackers, I found some problems about the current implement of limit-volume-copy-bandwidth (this patch has been merged in last week.) Firstly, assume that I configurate volume_copy_bps_limit=10M, If the path is a block device, cgroup blkio can limit copy-bandwidth separately for every volume. But If the path is a regular file, according to the current implement, cgroup blkio have to limit total copy-bandwidth for all volume on the disk device which the file lies on. The reason is : In cinder/utils.py, the method get_blkdev_major_minor elif lookup_for_file: # lookup the mounted disk which the file lies on out, _err = execute('df', path) devpath = out.split(\n)[1].split()[0] return get_blkdev_major_minor(devpath, False) If invoke the method copy_volume concurrently, copy-bandwidth for a volume is less than 10M. In this case, the meaning of param volume_copy_bps_limit in cinder.conf is defferent. Secondly, In NFS, the result of cmd 'df' is like this: [root@yuzhou yuzhou]# df /mnt/111 Filesystem 1K-blocks Used Available Use% Mounted on 186.100.8.144:/mnt/nfs_storage 51606528 14676992 34308096 30% /mnt I think the method get_blkdev_major_minor can not deal with the devpath '186.100.8.144:/mnt/nfs_storage'. i.e can not limit volume copy bandwidth in nfsdriver. So I think maybe we should modify the current implement to make sure copy-bandwidth for every volume meet the configuration requirement. I suggest we use loop device associated with the regular file(losetup /dev/loop0 /mnt/volumes/vm.qcow2), then limit the bps of loop device.( cgset -r blkio.throttle.write_bps_device=7:0 1000 test) After copying volume, detach loop device. (losetup --detach /dev/loop0) Any suggestion about my improvement opinions? Thanks! Zhou Yu ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] [messaging] 'retry' option
On 06/28/2014 10:49 PM, Mark McLoughlin wrote: On Fri, 2014-06-27 at 17:02 +0100, Gordon Sim wrote: A question about the new 'retry' option. The doc says: By default, cast() and call() will block until the message is successfully sent. What does 'successfully sent' mean here? Unclear, ambiguous, probably driver dependent etc. The 'blocking' we're talking about here is establishing a connection with the broker. If the connection has been lost, then cast() will block until the connection has been re-established and the message 'sent'. Understood, but to my mind, that is really an implementation detail. Does it mean 'written to the wire' or 'accepted by the broker'? For the impl_qpid.py driver, each send is synchronous, so it means accepted by the broker[1]. What does the impl_rabbit.py driver do? Does it just mean 'written to the wire', or is it using RabbitMQ confirmations to get notified when the broker accepts it (standard 0-9-1 has no way of doing this). I don't know, but it would be nice if someone did take the time to figure it out and document it :) Having googled around a bit, it appears that kombu v3.* has a 'confirm_publish' transport option when using the 'pyamqp' transport. That isn't available in the 2.* versions, which appear to be what is used in oslo.messaging, and I can't find that option specified anywhere either in the oslo.messaging codebase. Running a series of casts using the latest impl_rabbit.py driver and examining the data on the wire also shows no confirms being sent. So for impl_rabbit, the send is not acknowledged, but the delivery to consumers is. For impl_qpid its the other way round; the send is acknowledged but the delivery to consumers is not (though a prefetch of 1 is used limiting the loss to one message). Seriously, some docs around the subtle ways that the drivers differ from one another would be helpful ... particularly if it exposed incorrect assumptions API users are currently making. I'm happy to try and contribute to that. If the intention is to block until accepted by the broker that has obvious performance implications. On the other hand if it means block until written to the wire, what is the advantage of that? Was that a deliberate feature or perhaps just an accident of implementation? The use case for the new parameter, as described in the git commit, seems to be motivated by wanting to avoid the blocking when sending notifications. I can certainly understand that desire. However, notifications and casts feel like inherently asynchronous things to me, and perhaps having/needing the synchronous behaviour is the real issue? It's not so much about sync vs async, but a failure mode. By default, if we lose our connection with the broker, we wait until we can re-establish it rather than throwing exceptions (requiring the API caller to have its own retry logic) or quietly dropping the message. Even when you have no failure, your calling thread has to wait until the point the send is deemed successful before returning. So it is synchronous with respect to whatever that success criteria is. In the case where success is deemed to be acceptance by the broker (which is the case for the impl_qpid.py driver at present, whether intentional or not), the call is fully synchronous. If on the other hand success is merely writing the message to the wire, then any failure may well cause message loss regardless of the retry option. The reconnect and retry in this case is only of limited value. It can avoid certain losses, but not others. The use case for ceilometer is to allow its RPCPublisher to have a publishing policy - block until the samples have been sent, queue (in an in-memory, fixed-length queue) if we don't have a connection to the broker, or drop it if we don't have a connection to the broker. https://review.openstack.org/77845 I do understand the ambiguity around what message delivery guarantees are implicit in cast() isn't ideal, but that's not what adding this 'retry' parameter was about. Sure, I understand that. The retry option is necessitated by an (existing) implicit behaviour. However in my view that behaviour is implementations specific and of limited value in terms of the semantic contract of the call. --Gordon. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [ceilometer] How to filter out meters whose resources have been deleted?
Hi, As time goes on, meters will be a huge list, and some meters whose resources have been deleted may be useless for me, can I filter them out from meter list? Thanks ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
On 06/29/2014 09:39 AM, Joshua Hesketh wrote: On 6/28/14 10:40 AM, James E. Blair wrote: An alternate approach would be to have third-party CI systems register jobs with OpenStack's Zuul rather than using their own account. This would mean only a single report of all jobs (upstream and 3rd-party) per-patchset. It significantly reduces clutter and makes results more accessible -- but even with one system we've never actually wanted to have Jenkins results in comments, so I think one of the other options would be preferred. Nonetheless, this is possible with a little bit of work. I agree this isn't the preferred solution, but I disagree with the little bit of work. This would require CI systems registering with gearman which would mean security issues. The biggest problem with this though is that zuul would be stuck waiting from results from 3rd parties which often have very slow return times. Right, one of the other issues is the quality of the CI results varies as well. I think one of the test result burn out issues right now is based on the fact that they are too rolled up as it is. For instance, a docs only change gets Tempest results, which humans know are irrelevant, but Jenkins insists they aren't. I think that if we rolled up more information, and waited longer, we'd be in a worse state. -Sean -- Sean Dague http://dague.net signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [rubick] Proposal to make py33 job voting for stackforge/rubick
Hi, To proceed with this, I have sent (presumably) appropriate change to review: https://review.openstack.org/#/c/103516/ -- Best regards, Oleg Gelbukh On Fri, Jun 27, 2014 at 6:40 PM, Jay Pipes jaypi...@gmail.com wrote: Sure, why not? :) On 06/27/2014 06:25 AM, Oleg Gelbukh wrote: Hello, As our commits consistently pass py33 tests for last month (although not so many changes were made), I propose to enable py33 job voting on stackforge/rubick repository. What do you think? -- Best regards, Oleg Gelbukh ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo][messaging] Further improvements and refactoring
-BEGIN PGP SIGNED MESSAGE- Hash: SHA512 On 06/27/2014 04:04 PM, Ihar Hrachyshka wrote: On 26/06/14 22:38, Alexei Kornienko wrote: Hello Jay, Benchmark for oslo.messaging is really simple: You create a client that sends messages infinitively and a server that processes them. After you can use rabbitmq management plugin to see average throughput in the queue. Simple example can be found here - https://github.com/andreykurilin/profiler-for-oslo-messaging I've mentioned some of this already in my previous mails but here it is again: Huge is descriptive but not quantitative :) Do you have any numbers that pinpoint the amount of time that is being spent reconstructing and declaring the queues, say, compared to the time spent doing transmission? I don't have precise slowdown percentage for each issue. I've just identified hotspots using cProfile and strace. This ten times number... is that an estimate or do you have hard numbers that show that? Just curious. Using a script that I've mentioned earlier I get average throughput on my PC ~700 cast calls per second. I've written a simple and stupid script that uses kombu directly (in a single threaded and synchronous way with single connection) It gave me throughput ~8000 messages per second. Thats why I say that library should work at least 10 times faster. It doesn't show that those major issues you've pointed out result in such large message processing speed dropdown though. Maybe there are other causes of slowdown we observe. Neither it shows that refactoring of the code will actually help to boost the library significantly. It doesn't show that those major issues are the *only* reason for slowdown. Bit it shows that those issues are biggest that are currently visible. My understanding is that your analysis is mostly based on running a profiler against the code. Network operations can be bottlenecked in other places. You compare 'simple script using kombu' with 'script using oslo.messaging'. You don't compare script using oslo.messaging before refactoring and 'after that. The latter would show whether refactoring was worth the effort. Your test shows that oslo.messaging performance sucks, but it's not definite that hotspots you've revealed, once fixed, will show huge boost. My concern is that it may turn out that once all the effort to refactor the code is done, we won't see major difference. So we need base numbers, and performance tests would be a great helper here. We'll get a major speed boost if we fix them (and possibly discover new issues the would prevent us of reaching full speed). Though having some hard numbers from using kombu directly is still a good thing. Is it possible that we introduce some performance tests into oslo.messaging itself? Some performance tests may be introduced but they would be more like functional tests since they require setup of actual messaging server (rabbit, etc.). Yes. I think we already have some. F.e. tests/drivers/test_impl_qpid.py attempts to use local Qpid server (backing up to fake server if it's not available). We could create a separate subtree in the library for functional tests. What do you mean exactly by performance tests? Just testing overall throughput with some basic scenarios or you mean finding hotspots in the code? The former. Once we have some base numbers, we may use them to consider whether changes you propose are worth the effort. Regards, Alexei Kornienko On 06/26/2014 11:22 PM, Jay Pipes wrote: Hey Alexei, thanks for sharing your findings. Comments inline. On 06/26/2014 10:08 AM, Alexei Kornienko wrote: Hello, Returning to performance issues of oslo.messaging. I've found 2 biggest issue in existing implementation of rabbit (kombu) driver: 1) For almost every message sent/received a new object of Consumer/Publisher class is created. And each object of this class tries to declare it's queue even if it's already declared. https://github.com/openstack/oslo.messaging/blob/master/oslo/messaging/_drivers/impl_rabbit.py#L159 This causes a huge slowdown. Huge is descriptive but not quantitative :) Do you have any numbers that pinpoint the amount of time that is being spent reconstructing and declaring the queues, say, compared to the time spent doing transmission? 2) with issue #1 is fixed (I've applied a small hack to fix it in my repo) the next big issue araise. For every rpc message received a reply is sent when processing is done (it seems that reply is sent even for cast calls which it really strange to me). Reply sent is using connection pool to speed up replies. Due to bad implementation of custom connection pool for every message sent underlying connection channel is closed and reopened: https://github.com/openstack/oslo.messaging/blob/master/oslo/messaging/_drivers/impl_rabbit.py#L689 Cause of this major issues oslo.messaging performance is at least 10 times slower
[openstack-dev] [mistral] Using __init__.py files
Hi, What would be your opinion on the question “Should we place any important functionality into __init__.py files or just use it for package level initialization and exporting variables from module level to a package level?”. I personally would prefer not to keep there anything like class Engine (which is one of the most important parts of Mistral now). It’s somewhat confusing to me, especially when I navigate through the project structure. It’s not a critical urgent thing, of course, but would be nice if you share your opinion. What do you guys think? Renat Akhmerov @ Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [mistral] Using __init__.py files
Renat, As far as I can tell, it is de-facto standard to not place anything at all to __init__.py across the majority of OpenStack projects. -- Best regards, Oleg Gelbukh On Mon, Jun 30, 2014 at 3:50 PM, Renat Akhmerov rakhme...@mirantis.com wrote: Hi, What would be your opinion on the question “Should we place any important functionality into __init__.py files or just use it for package level initialization and exporting variables from module level to a package level?”. I personally would prefer not to keep there anything like class Engine (which is one of the most important parts of Mistral now). It’s somewhat confusing to me, especially when I navigate through the project structure. It’s not a critical urgent thing, of course, but would be nice if you share your opinion. What do you guys think? Renat Akhmerov @ Mirantis Inc. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [olso] olso incubator logging issues
Every time I crack open a nova logs in detail, at least 2 new olso incubator log issues have been introduced. The current ones is clearly someone is over exploding arrays, as we're getting things like: 2014-06-29 13:36:41.403 19459 DEBUG nova.openstack.common.processutils [-] Running cmd (subprocess): [ ' e n v ' , ' L C _ A L L = C ' , ' L A N G = C ' , ' q e m u - i m g ' , ' i n f o ' , ' / o p t / s t a c k / d a t a / n o v a / i n s t a n c e s / e f f 7 3 1 3 a - 1 1 b 2 - 4 0 2 b - 9 c c d - 6 5 7 8 c b 8 7 9 2 d b / d i s k ' ] execute /opt/stack/new/nova/nova/openstack/common/processutils.py:160 (yes all those spaces are in there, which now effectively inhibits search). Also on every wsgi request to Nova API we get something like this: 2014-06-29 13:26:43.836 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute:get will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.837 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:security_groups will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.838 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:security_groups will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.838 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:keypairs will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.838 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:hide_server_addresses will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.838 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_volumes will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.842 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:config_drive will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.842 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:server_usage will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.842 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_status will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.843 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_server_attributes will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.843 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_ips_mac will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.843 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_ips will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.843 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:extended_availability_zone will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 2014-06-29 13:26:43.844 DEBUG nova.openstack.common.policy [req-86680d63-6d6c-4962-9274-1de7de8ca37d FixedIPsNegativeTestJson-768779075 FixedIPsNegativeTestJson-579905596] Rule compute_extension:disk_config will be now enforced enforce /opt/stack/new/nova/nova/openstack/common/policy.py:288 On *every* request. oslo code, by definition, is going to be used a lot, inside of tight
Re: [openstack-dev] [Neutron][IPv6] Neutron IPv6 in Icehouse and further
Hi, There is a patch for radvd https://review.openstack.org/#/c/102648/2 that you can use in addition to the devstack patch. You want to make sure that ipv6 is enabled and ra accepted with your VM’s image. Both patches are under development. To use dhcpv6, the current dhcp agent should be working. However, it might be broken due to a recent commit. If you found a Traceback in your dhcp agent log complaining about an uninitialized reference to the variable ‘mode', you may hit that issue. You can just initialize it to be ‘static’. In addition, only recently, dhcp messages maybe dropped before entering the VM. Therefore, you might need to disable the ipv6 rules manually by “ip6tables –F” after a VM is launched . Also make sure you are using the latest dnsmasq (2.6.8) Thanks, Robert On 6/27/14, 3:47 AM, Jaume Devesa devv...@gmail.commailto:devv...@gmail.com wrote: Hello Maksym, last week I had more or less the same questions than you and I investigate a little bit... Currently we have the ipv6_ra_mode and ipv6_address_mode in the subnet entity. The way you combine these two values will determine how and who will configure your VM´s IPv6 addresses. Not all the combinations are possible. This document[1] and the upstream-slaac-support spec[2] provide the possible combinations. Not sure which one is more updated... If you want to try IPv6 current support, you can use the Baodong Li's devstack patch [3], although is still in development. Follow the message commit instructions to provide a radvd daemon. That means that there is no RA advertiser in Neutron currently. There is a spec in review[4] to fill this gap. The changes for allow DHCPv6 in dnsmasq are in review in this patch[5]. This is what I found... I hope some IPv6 folks can correct me if this information is not accurate enough (or wrong) [1]: https://www.dropbox.com/s/9bojvv9vywsz8sd/IPv6%20Two%20Modes%20v3.0.pdf [2]: http://docs-draft.openstack.org/43/88043/9/gate/gate-neutron-specs-docs/82c251a/doc/build/html/specs/juno/ipv6-provider-nets-slaac.html [3]: https://review.openstack.org/#/c/87987 [4]: https://review.openstack.org/#/c/101306/ [5]: https://review.openstack.org/#/c/70649/ On 27 June 2014 00:51, Martinx - ジェームズ thiagocmarti...@gmail.commailto:thiagocmarti...@gmail.com wrote: Hi! I'm waiting for that too... Currently, I'm running IceHouse with static IPv6 address, with the topology VLAN Provider Networks and, to make it easier, I'm counting on the following blueprint: https://blueprints.launchpad.net/neutron/+spec/ipv6-provider-nets-slaac ...but, I'm not sure if it will be enough to enable basic IPv6 support (without using Neutron as Instance's default gateway)... Cheers! Thiago On 26 June 2014 19:35, Maksym Lobur mlo...@mirantis.commailto:mlo...@mirantis.com wrote: Hi Folks, Could you please tell what is the current state of IPv6 in Neutron? Does it have DHCPv6 working? What is the best point to start hacking from? Devstack stable/icehouse or maybe some tag? Are there any docs / raw deployment guides? I see some patches not landed yet [1] ... I assume it won't work without them, right? Somehow I can't open any of the code reviews from the [2] (Not Found) [1] https://review.openstack.org/#/q/status:open+project:openstack/neutron+branch:master+topic:%255E.*%255Cipv6.*,n,z [2] https://wiki.openstack.org/wiki/Neutron/IPv6 Best regards, Max Lobur, Python Developer, Mirantis, Inc. Mobile: +38 (093) 665 14 28tel:%2B38%20%28093%29%20665%2014%2028 Skype: max_lobur 38, Lenina ave. Kharkov, Ukraine www.mirantis.comhttp://www.mirantis.com www.mirantis.ruhttp://www.mirantis.ru ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.orgmailto:OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev -- Jaume Devesa Software Engineer at Midokura ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova] Meeting this week?
On 06/29/2014 08:01 PM, Michael Still wrote: Hi. The meeting this week would be on the 3rd of July, which I assume means that many people will be out of the office. Do people think its worth running the meeting or shall we give this week a miss? I will be traveling during the meeting so I'm +1 on giving it a miss. Thanks, Michael ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova] Meeting this week?
On Jun 30, 2014 9:23 AM, Andrew Laski andrew.la...@rackspace.com wrote: On 06/29/2014 08:01 PM, Michael Still wrote: Hi. The meeting this week would be on the 3rd of July, which I assume means that many people will be out of the office. Do people think its worth running the meeting or shall we give this week a miss? I will be traveling during the meeting so I'm +1 on giving it a miss. Same here. Thanks, Michael ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][scheduler] Proposal: FairShareScheduler.
Hi All, we have analyzed the nova-scheduler component (FilterScheduler) in our Openstack installation used by some scientific teams. In our scenario, the cloud resources need to be distributed among the teams by considering the predefined share (e.g. quota) assigned to each team, the portion of the resources currently used and the resources they have already consumed. We have observed that: 1) User requests are sequentially processed (FIFO scheduling), i.e. FilterScheduler doesn't provide any dynamic priority algorithm; 2) User requests that cannot be satisfied (e.g. if resources are not available) fail and will be lost, i.e. on that scenario nova-scheduler doesn't provide any queuing of the requests; 3) OpenStack simply provides a static partitioning of resources among various projects / teams (use of quotas). If project/team 1 in a period is systematically underutilizing its quota and the project/team 2 instead is systematically saturating its quota, the only solution to give more resource to project/team 2 is a manual change (to be done by the admin) to the related quotas. The need to find a better approach to enable a more effective scheduling in Openstack becomes more and more evident when the number of the user requests to be handled increases significantly. This is a well known problem which has already been solved in the past for the Batch Systems. In order to solve those issues in our usage scenario of Openstack, we have developed a prototype of a pluggable scheduler, named FairShareScheduler, with the objective to extend the existing OpenStack scheduler (FilterScheduler) by integrating a (batch like) dynamic priority algorithm. The architecture of the FairShareScheduler is explicitly designed to provide a high scalability level. To all user requests will be assigned a priority value calculated by considering the share allocated to the user by the administrator and the evaluation of the effective resource usage consumed in the recent past. All requests will be inserted in a priority queue, and processed in parallel by a configurable pool of workers without interfering with the priority order. Moreover all significant information (e.g. priority queue) will be stored in a persistence layer in order to provide a fault tolerance mechanism while a proper logging system will annotate all relevant events, useful for auditing processing. In more detail, some features of the FairshareScheduler are: a) It assigns dynamically the proper priority to every new user requests; b) The priority of the queued requests will be recalculated periodically using the fairshare algorithm. This feature guarantees the usage of the cloud resources is distributed among users and groups by considering the portion of the cloud resources allocated to them (i.e. share) and the resources already consumed; c) all user requests will be inserted in a (persistent) priority queue and then processed asynchronously by the dedicated process (filtering + weighting phase) when compute resources are available; d) From the client point of view the queued requests remain in “Scheduling” state till the compute resources are available. No new states added: this prevents any possible interaction issue with the Openstack clients; e) User requests are dequeued by a pool of WorkerThreads (configurable), i.e. no sequential processing of the requests; f) The failed requests at filtering + weighting phase may be inserted again in the queue for n-times (configurable). We have integrated the FairShareScheduler in our Openstack installation (release HAVANA). We're now working to adapt the FairShareScheduler to the new release IceHouse. Does anyone have experiences in those issues found in our cloud scenario? Could the FairShareScheduler be useful for the Openstack community? In that case, we'll be happy to share our work. Any feedback/comment is welcome! Cheers, Eric. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] oslosphinx hacking build-depending on each other
Hi, It'd be nice to fix the fact that oslosphinx hacking are build-depending on each other. How can we fix this? Thomas Goirand (zigo) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [rubick] Proposal to make py33 job voting for stackforge/rubick
On 06/27/2014 06:25 PM, Oleg Gelbukh wrote: Hello, As our commits consistently pass py33 tests for last month (although not so many changes were made), I propose to enable py33 job voting on stackforge/rubick repository. What do you think? -- Best regards, Oleg Gelbukh Hi, While gating with Python 3 is a very good thing, I have to raise the fact that both Debian Sid and Ubuntu are now using Python 3.4 (Python 3.3 was removed from Sid a few weeks ago). And when it comes to Wheezy, we're still stuck with Python 3.2. So please go ahead and gate with Python 3.3, but please also consider adding support for Python 3.4 at least, and Python 3.2 if you can. Cheers, Thomas Goirand (zigo) ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Horizon] SCSS conversion and SCSS upgrade patches
I had an IRC discussion with jtomasek and rdopieralski recently regarding Radomir's patch for converting to SCSS bootstrap. https://review.openstack.org/#/c/90371/ We were trying to sort out how soon this patch should merge. We'd like to discuss this at the team meeting tomorrow, but I'm sharing here to get an initial discussion started. It seems that in this version of bootstrap SCSS had some bugs. They are low impact, but pretty obvious bugs - I saw mouseover styling and disable button styling problems. The most straightforward fix to this is to upgrade bootstrap. I understand that Jiri is working on that this week, and would like to have the SCSS patch merged ASAP to help with that effort. My feeling is that we shouldn't merge the SCSS patch until we have a bootstrap patch ready to merge. I'd like to see dependent patches used to manage this so we can get both changes reviewed and merged at the same time. Radomir has shared that he thinks dependent patches are too painful to use for this situation. Also he'd like to see the SCSS bootstrap patch merged ASAP because the horizon split depends on that work as well. Doug Fish ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Horizon] SCSS conversion and SCSS upgrade patches
On Mon 30 Jun 2014 08:09:40 AM MDT, Douglas Fish wrote: I had an IRC discussion with jtomasek and rdopieralski recently regarding Radomir's patch for converting to SCSS bootstrap. https://review.openstack.org/#/c/90371/ We were trying to sort out how soon this patch should merge. We'd like to discuss this at the team meeting tomorrow, but I'm sharing here to get an initial discussion started. It seems that in this version of bootstrap SCSS had some bugs. They are low impact, but pretty obvious bugs - I saw mouseover styling and disable button styling problems. The most straightforward fix to this is to upgrade bootstrap. I understand that Jiri is working on that this week, and would like to have the SCSS patch merged ASAP to help with that effort. My feeling is that we shouldn't merge the SCSS patch until we have a bootstrap patch ready to merge. I'd like to see dependent patches used to manage this so we can get both changes reviewed and merged at the same time. Radomir has shared that he thinks dependent patches are too painful to use for this situation. Also he'd like to see the SCSS bootstrap patch merged ASAP because the horizon split depends on that work as well. Doug Fish ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev I agree with Radomir and will voice that same opinion during the meeting. Lets get the SCSS patch up, and then everyone can contribute to improvements, not just Jiri. -J -- Jason E. Rist Senior Software Engineer OpenStack Management UI Red Hat, Inc. openuc: +1.972.707.6408 mobile: +1.720.256.3933 Freenode: jrist github/identi.ca: knowncitizen ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][scheduler] Proposal: FairShareScheduler.
On 30 June 2014 16:05, Eric Frizziero eric.frizzi...@pd.infn.it wrote: In more detail, some features of the FairshareScheduler are: a) It assigns dynamically the proper priority to every new user requests; b) The priority of the queued requests will be recalculated periodically using the fairshare algorithm. This feature guarantees the usage of the cloud resources is distributed among users and groups by considering the portion of the cloud resources allocated to them (i.e. share) and the resources already consumed; c) all user requests will be inserted in a (persistent) priority queue and then processed asynchronously by the dedicated process (filtering + weighting phase) when compute resources are available; d) From the client point of view the queued requests remain in “Scheduling” state till the compute resources are available. No new states added: this prevents any possible interaction issue with the Openstack clients; e) User requests are dequeued by a pool of WorkerThreads (configurable), i.e. no sequential processing of the requests; f) The failed requests at filtering + weighting phase may be inserted again in the queue for n-times (configurable). We have integrated the FairShareScheduler in our Openstack installation (release HAVANA). We're now working to adapt the FairShareScheduler to the new release IceHouse. Does anyone have experiences in those issues found in our cloud scenario? Could the FairShareScheduler be useful for the Openstack community? In that case, we'll be happy to share our work. Sounds like an interesting option to have available. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] oslosphinx hacking build-depending on each other
On 2014-06-30 22:11:30 +0800 (+0800), Thomas Goirand wrote: It'd be nice to fix the fact that oslosphinx hacking are build-depending on each other. How can we fix this? They're only build-depending on one another (in the Debian sense) because you have made them to do so. Hacking does have a build-time dependency on oslosphinx for generating its documentation, yes. On the other hand oslosphinx only has a test-time dependency on hacking. It's the decision to also run your tests at build-time (presumably because Debian lacks any package build provision to differentiate doc build requirements from testing requirements) which make this at all circular. Perhaps you could skip running hacking tests on oslosphinx (you could run the rest of flake8 against it, just omit the hacking extension)? Alternatively, use build profiles... oslosphinx pretty clearly falls into the documentation tools description in https://wiki.debian.org/DebianBootstrap#Circular_dependencies.2Fstaged_builds so might be a good fit. -- Jeremy Stanley ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo][messaging] Further improvements and refactoring
On 06/30/2014 12:22 PM, Ihar Hrachyshka wrote: Alexei Kornienko wrote: Some performance tests may be introduced but they would be more like functional tests since they require setup of actual messaging server (rabbit, etc.). Yes. I think we already have some. F.e. tests/drivers/test_impl_qpid.py attempts to use local Qpid server (backing up to fake server if it's not available). I always get failures when there is a real qpidd service listening on the expected port. Does anyone else see this? ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][i18n] why isn't the client translated?
I noticed that there is no locale directory or setup.cfg entry for babel, which surprises me. The v1_1 shell in python-novaclient has a lot of messages marked for translation using the _() function but the v3 shell doesn't, presumably because someone figured out we don't translate the client messages anyway. I'm just wondering why we don't translate the client? -- Thanks, Matt Riedemann ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
Joshua Hesketh joshua.hesk...@rackspace.com writes: On 6/28/14 10:40 AM, James E. Blair wrote: An alternate approach would be to have third-party CI systems register jobs with OpenStack's Zuul rather than using their own account. This would mean only a single report of all jobs (upstream and 3rd-party) per-patchset. It significantly reduces clutter and makes results more accessible -- but even with one system we've never actually wanted to have Jenkins results in comments, so I think one of the other options would be preferred. Nonetheless, this is possible with a little bit of work. I agree this isn't the preferred solution, but I disagree with the little bit of work. This would require CI systems registering with gearman which would mean security issues. The biggest problem with this though is that zuul would be stuck waiting from results from 3rd parties which often have very slow return times. Security issues is a bit vague. They already register with Gerrit; I'm only suggesting that the point of aggregation would change. I'm anticipating that they would use authenticated SSL, with ACLs scoped to the names of jobs each system is permitted to run. From the perspective of overall security as well as network topology (ie, firewalls), very little changes. The main differences are third party CI systems don't have to run Zuul anymore, and we go back to having a smaller number of votes/comments. Part of the little bit of work I was referring to was adding a timeout. That should truly be not much work, and work we're planning on doing anyway to help with the tripleo cloud. But anyway, it's not important to design this out if we prefer another solution (and I prefer the table of results separated from comments). -Jim ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] Use MariaDB by default on Fedora
On 06/30/2014 11:16 AM, Michael Kerrin wrote: I am trying to finish off https://review.openstack.org/#/c/90134 - percona xtradb cluster for debian based system. I have read into this thread that I can error out on Redhat systems when trying to install percona and tell them to use mariadb instead, percona isn't support here. Is this correct? I think on rhel mysql/xtradb is expected to work from the percona tarball, as it used to happen on fedora yet this change [1] is eventually going to switch the default solution to mariadb/galera for rhel too which is consistent with the previous idea from James there could still be unexpected issues on distro for which we don't have CI but I wouldn't expect troubles caused by your change 1. https://review.openstack.org/#/c/102818/ -- Giulio Fidente GPG KEY: 08D733BA ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Horizon] SCSS conversion and SCSS upgrade patches
It's possible that the only purpose of this discussion will be to fix my thinking, but I still can't understand why we are so anxious to integrate a patch that makes Horizon look funny. I expect the patch Radomir has our for review will be quite stable. People should be able to contribute work based on it now using dependent patches. https://wiki.openstack.org/wiki/Gerrit_Workflow#Add_dependency Once we have a set of patches with good Horizon behavior we can merge the set. Doug Fish Jason Rist jr...@redhat.com wrote on 06/30/2014 09:22:20 AM: From: Jason Rist jr...@redhat.com To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org Cc: Douglas Fish/Rochester/IBM@IBMUS Date: 06/30/2014 09:22 AM Subject: Re: [openstack-dev] [Horizon] SCSS conversion and SCSS upgrade patches On Mon 30 Jun 2014 08:09:40 AM MDT, Douglas Fish wrote: I had an IRC discussion with jtomasek and rdopieralski recently regarding Radomir's patch for converting to SCSS bootstrap. https://review.openstack.org/#/c/90371/ We were trying to sort out how soon this patch should merge. We'd like to discuss this at the team meeting tomorrow, but I'm sharing here to get an initial discussion started. It seems that in this version of bootstrap SCSS had some bugs. They are low impact, but pretty obvious bugs - I saw mouseover styling and disable button styling problems. The most straightforward fix to this is to upgrade bootstrap. I understand that Jiri is working on that this week, and would like to have the SCSS patch merged ASAP to help with that effort. My feeling is that we shouldn't merge the SCSS patch until we have a bootstrap patch ready to merge. I'd like to see dependent patches used to manage this so we can get both changes reviewed and merged at the same time. Radomir has shared that he thinks dependent patches are too painful to use for this situation. Also he'd like to see the SCSS bootstrap patch merged ASAP because the horizon split depends on that work as well. Doug Fish ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev I agree with Radomir and will voice that same opinion during the meeting. Lets get the SCSS patch up, and then everyone can contribute to improvements, not just Jiri. -J -- Jason E. Rist Senior Software Engineer OpenStack Management UI Red Hat, Inc. openuc: +1.972.707.6408 mobile: +1.720.256.3933 Freenode: jrist github/identi.ca: knowncitizen ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
Dan Smith d...@danplanet.com wrote on 06/27/2014 12:33:48 PM: What if 3rd Party CI didn't vote in Gerrit? What if it instead published to some 3rd party test reporting site (a thing that doesn't yet exist). Gerrit has the facility so that we could inject the dashboard content for this in Gerrit in a little table somewhere, but the data would fundamentally live outside of Gerrit. It would also mean that all the aggregate reporting of 3rd Party CI that's being done in custom gerrit scripts, could be integrated directly into such a thing. If it really does show up right in Gerrit as if it were integrated, then that would be fine with me. I think the biggest problem we have right now is that a lot of the CI systems are very inconsistent in their reporting and we often don't realize when one of them *hasn't* voted. If the new thing could fill out the chart based on everything we expect to see a vote from, so that it's very clear that one is missing, then that's a net win right there. There is a similar old bug for that, with a good suggestion for how it could possibly be done: https://bugs.launchpad.net/openstack-ci/+bug/1251758 Kurt Taylor (krtaylor) OpenStack Development Lead - PowerKVM CI IBM Linux Technology Center ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] VMware networking
Hi Gary, Thanks for sending this out, comments inline. On 29 June 2014 00:15, Gary Kotton gkot...@vmware.com wrote: Hi, At the moment there are a number of different BP’s that are proposed to enable different VMware network management solutions. The following specs are in review: 1. VMware NSX-vSphere plugin: https://review.openstack.org/102720 2. Neutron mechanism driver for VMWare vCenter DVS network creation: https://review.openstack.org/#/c/101124/ 3. VMware dvSwitch/vSphere API support for Neutron ML2: https://review.openstack.org/#/c/100810/ In addition to this there is also talk about HP proposing some for of VMware network management. I believe this is blueprint [1]. This was proposed a while ago, but now it needs to go through the new BP review process. [1] - https://blueprints.launchpad.net/neutron/+spec/ovsvapp-esxi-vxlan Each of the above has specific use case and will enable existing vSphere users to adopt and make use of Neutron. Items #2 and #3 offer a use case where the user is able to leverage and manage VMware DVS networks. This support will have the following limitations: - Only VLANs are supported (there is no VXLAN support) - No security groups - #3 – the spec indicates that it will make use of pyvmomi ( https://github.com/vmware/pyvmomi). There are a number of disclaimers here: - This is currently blocked regarding the integration into the requirements project (https://review.openstack.org/#/c/69964/) - The idea was to have oslo.vmware leverage this in the future ( https://github.com/openstack/oslo.vmware) Item #1 will offer support for all of the existing Neutron API’s and there functionality. This solution will require a additional component called NSX (https://www.vmware.com/support/pubs/nsx_pubs.html). It's great to see this breakdown, it's very useful in order to identify the potential gaps and overlaps amongst the various efforts around ESX and Neutron. This will also ensure a path towards a coherent code contribution. It would be great if we could all align our efforts and have some clear development items for the community. In order to do this I’d like suggest that we meet to sync and discuss all efforts. Please let me know if the following sounds ok for an initial meeting to discuss how we can move forwards: - Tuesday 15:00 UTC - IRC channel #openstack-vmware I am available to join. We can discuss the following: 1. Different proposals 2. Combining efforts 3. Setting a formal time for meetings and follow ups Looking forwards to working on this stuff with the community and providing a gateway to using Neutron and further enabling the adaption of OpenStack. I think code contribution is only one aspect of this story; my other concern is that from a usability standpoint we would need to provide a clear framework for users to understand what these solutions can do for them and which one to choose. Going forward I think it would be useful if we produced an overarching blueprint that outlines all the ESX options being proposed for OpenStack Networking (and the existing ones, like NSX - formerly known as NVP, or nova-network), their benefits and drawbacks, their technical dependencies, system requirements, API supported etc. so that a user can make an informed decision when looking at ESX deployments in OpenStack. Thanks Gary Cheers, Armando ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
Sean Dague s...@dague.net wrote on 06/30/2014 06:03:50 AM: From: Sean Dague s...@dague.net To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org, Date: 06/30/2014 06:09 AM Subject: Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit On 06/29/2014 09:39 AM, Joshua Hesketh wrote: On 6/28/14 10:40 AM, James E. Blair wrote: An alternate approach would be to have third-party CI systems register jobs with OpenStack's Zuul rather than using their own account. This would mean only a single report of all jobs (upstream and 3rd-party) per-patchset. It significantly reduces clutter and makes results more accessible -- but even with one system we've never actually wanted to have Jenkins results in comments, so I think one of the other options would be preferred. Nonetheless, this is possible with a little bit of work. I agree this isn't the preferred solution, but I disagree with the little bit of work. This would require CI systems registering with gearman which would mean security issues. The biggest problem with this though is that zuul would be stuck waiting from results from 3rd parties which often have very slow return times. Right, one of the other issues is the quality of the CI results varies as well. Agreed. After last summit, Anita, Jay and I decided to start gathering a team of 3rd party testers that have the goal of improving the quality of the third party systems. We are starting with gathering global unwritten requirements, improving documentation and reaching out to new projects that will have third party testing needs. We are still in the early stages but now have weekly meetings to discuss what needs to be done and track progress. https://wiki.openstack.org/wiki/Meetings/ThirdParty I think one of the test result burn out issues right now is based on the fact that they are too rolled up as it is. For instance, a docs only change gets Tempest results, which humans know are irrelevant, but Jenkins insists they aren't. I think that if we rolled up more information, and waited longer, we'd be in a worse state. Maybe it could promptly timeout and then report the system that did not complete? That would also have the benefit of enforcing a time limit on reporting results. Kurt Taylor (krtaylor) OpenStack Development Lead - PowerKVM CI IBM Linux Technology Center ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Neutron] VMware networking
On Mon, Jun 30, 2014 at 10:18 AM, Armando M. arma...@gmail.com wrote: Hi Gary, Thanks for sending this out, comments inline. Indeed, thanks Gary! On 29 June 2014 00:15, Gary Kotton gkot...@vmware.com wrote: Hi, At the moment there are a number of different BP’s that are proposed to enable different VMware network management solutions. The following specs are in review: VMware NSX-vSphere plugin: https://review.openstack.org/102720 Neutron mechanism driver for VMWare vCenter DVS network creation:https://review.openstack.org/#/c/101124/ VMware dvSwitch/vSphere API support for Neutron ML2: https://review.openstack.org/#/c/100810/ I've commented in these reviews about combining efforts here, I'm glad you're taking the lead to make this happen Gary. This is much appreciated! In addition to this there is also talk about HP proposing some for of VMware network management. I believe this is blueprint [1]. This was proposed a while ago, but now it needs to go through the new BP review process. [1] - https://blueprints.launchpad.net/neutron/+spec/ovsvapp-esxi-vxlan Each of the above has specific use case and will enable existing vSphere users to adopt and make use of Neutron. Items #2 and #3 offer a use case where the user is able to leverage and manage VMware DVS networks. This support will have the following limitations: Only VLANs are supported (there is no VXLAN support) No security groups #3 – the spec indicates that it will make use of pyvmomi (https://github.com/vmware/pyvmomi). There are a number of disclaimers here: This is currently blocked regarding the integration into the requirements project (https://review.openstack.org/#/c/69964/) The idea was to have oslo.vmware leverage this in the future (https://github.com/openstack/oslo.vmware) Item #1 will offer support for all of the existing Neutron API’s and there functionality. This solution will require a additional component called NSX (https://www.vmware.com/support/pubs/nsx_pubs.html). It's great to see this breakdown, it's very useful in order to identify the potential gaps and overlaps amongst the various efforts around ESX and Neutron. This will also ensure a path towards a coherent code contribution. It would be great if we could all align our efforts and have some clear development items for the community. In order to do this I’d like suggest that we meet to sync and discuss all efforts. Please let me know if the following sounds ok for an initial meeting to discuss how we can move forwards: - Tuesday 15:00 UTC - IRC channel #openstack-vmware I am available to join. We can discuss the following: Different proposals Combining efforts Setting a formal time for meetings and follow ups Looking forwards to working on this stuff with the community and providing a gateway to using Neutron and further enabling the adaption of OpenStack. I think code contribution is only one aspect of this story; my other concern is that from a usability standpoint we would need to provide a clear framework for users to understand what these solutions can do for them and which one to choose. Going forward I think it would be useful if we produced an overarching blueprint that outlines all the ESX options being proposed for OpenStack Networking (and the existing ones, like NSX - formerly known as NVP, or nova-network), their benefits and drawbacks, their technical dependencies, system requirements, API supported etc. so that a user can make an informed decision when looking at ESX deployments in OpenStack. Thanks Gary Cheers, Armando ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
2014-06-30 19:17 GMT+04:00 Kurt Taylor krtay...@us.ibm.com: Dan Smith d...@danplanet.com wrote on 06/27/2014 12:33:48 PM: If it really does show up right in Gerrit as if it were integrated, then that would be fine with me. I think the biggest problem we have right now is that a lot of the CI systems are very inconsistent in their reporting and we often don't realize when one of them *hasn't* voted. If the new thing could fill out the chart based on everything we expect to see a vote from, so that it's very clear that one is missing, then that's a net win right there. There is a similar old bug for that, with a good suggestion for how it could possibly be done: https://bugs.launchpad.net/openstack-ci/+bug/1251758 What about to have report like this: http://stackalytics.com/report/ci/neutron/7 ? Thanks, Ilya ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all] 3rd Party CI vs. Gerrit
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 There is a similar old bug for that, with a good suggestion for how it could possibly be done: https://bugs.launchpad.net/openstack-ci/+bug/1251758 This isn't what I'm talking about. What we need is, for each new patchset on a given change, an empty table listing all the answers we expect to see (i.e. one for each of the usual suspect nova CI systems). The above bug (AFAICT) is simply for tracking last-status of each, so that if one stops reporting entirely (as minesweeper often does), we get some indication that the *system* is broken. - --Dan -BEGIN PGP SIGNATURE- Version: GnuPG v1 Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/ iQEcBAEBAgAGBQJTsYc+AAoJEBeZxaMESjNVcQYH+wayg4T9QFe4tTvGn24PisCf 5cEaeSkwXl+Adiae5cCfCGSTjlErK4lpFtzFapKukcM0+eEp464toskl7vNC0izp UWCpcg2gbON6Ef/AMa1+PT8uXR9OYAo+/eU8NUJNM01ajeZqqe3H3jnltgoUau0O fq3O3+Wa2PxBTAVVGi3HXJl4SWpVdEuYDZYOBOtkDXwhIS/hvBdIRuJwt0CygxHx 78WatFsQ09tBHaQCJbs2E+Oar0rD4sF93qjG8jAFiVB/0SJ6wV7AsLColVId2hbe Qfua3Q6CufJBO2WHV7JORX2fBSOTUmcPcOM1IE4/lgXGiyu3aw5ataL9e4qxudQ= =VoHH -END PGP SIGNATURE- ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][i18n] why isn't the client translated?
forwarding I18n Team. -- About.me 부터: Matt Riedemann mrie...@linux.vnet.ibm.com 답장: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org 날짜: 2014년 7월 1일 at 오전 12:02:41 에게: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.org 제목: [openstack-dev] [nova][i18n] why isn't the client translated? I noticed that there is no locale directory or setup.cfg entry for babel, which surprises me. The v1_1 shell in python-novaclient has a lot of messages marked for translation using the _() function but the v3 shell doesn't, presumably because someone figured out we don't translate the client messages anyway. I'm just wondering why we don't translate the client? -- Thanks, Matt Riedemann ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron][third-party] Simple and robust CI script?
On 30 June 2014 17:34, Kyle Mestery mest...@noironetworks.com wrote: It would be great to get you to join the 3rd Party meeting [1] in #openstack-meeting at 1800UTC to discuss this. Can you make it today Luke? Yes, I'll be there. Currently I'm looking into the simplest 3rd party CI that could possibly work which would be less sophisticated than devstack-gate but easier for the operator to understand and be responsible for. (Or: CI in 100 lines of shell with no Jenkins or other large pieces of software to install.) I'm on #openstack-neutron if anybody wants to kick around ideas or share scripts that I can borrow ideas from (thanks ODL gang for doing this already). Cheers, -Luke ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [infra][oslo][neutron] Need help getting oslo.messaging 1.4.0.0a2 in global requirements
I have out for review 103536 to add this version to global requirements, so that Neutron has an oslo fix (review 102909) for encoding failure, which affects some gate runs. This review for global requirements is failing requirements check (http://logs.openstack.org/36/103536/1/check/check-requirements-integration-dsvm/6d9581c/console.html#_2014-06-30_12_34_56_921). I did a recheck bug 1334898, but see the same error, with the release not found, even though it is in PyPI. Infra folks say this is a known issue with pushing out pre-releases. Do we have a work-around? Any proposed solution to try? Thanks! PCM (Paul Michali) MAIL …..…. p...@cisco.com IRC ……..… pcm_ (irc.freenode.com) TW ………... @pmichali GPG Key … 4525ECC253E31A83 Fingerprint .. 307A 96BB 1A4C D2C7 931D 8D2D 4525 ECC2 53E3 1A83 signature.asc Description: Message signed with OpenPGP using GPGMail ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [oslo] Openstack and SQLAlchemy
Hi all - For those who don't know me, I'm Mike Bayer, creator/maintainer of SQLAlchemy, Alembic migrations and Dogpile caching. In the past month I've become a full time Openstack developer working for Red Hat, given the task of carrying Openstack's database integration story forward. To that extent I am focused on the oslo.db project which going forward will serve as the basis for database patterns used by other Openstack applications. I've summarized what I've learned from the community over the past month in a wiki entry at: https://wiki.openstack.org/wiki/Openstack_and_SQLAlchemy The page also refers to an ORM performance proof of concept which you can see at https://github.com/zzzeek/nova_poc. The goal of this wiki page is to publish to the community what's come up for me so far, to get additional information and comments, and finally to help me narrow down the areas in which the community would most benefit by my contributions. I'd like to get a discussion going here, on the wiki, on IRC (where I am on freenode with the nickname zzzeek) with the goal of solidifying the blueprints, issues, and SQLAlchemy / Alembic features I'll be focusing on as well as recruiting contributors to help in all those areas. I would welcome contributors on the SQLAlchemy / Alembic projects directly as well, as we have many areas that are directly applicable to Openstack. I'd like to thank Red Hat and the Openstack community for welcoming me on board and I'm looking forward to digging in more deeply in the coming months! - mike ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] [Heat] Reminder: Mid-cycle Meetup - Attendance Confirmation
On 25/06/14 14:32 -0400, Jordan OMara wrote: On 25/06/14 18:20 +, Carlino, Chuck (OpenStack TripleO, Neutron) wrote: Is $179/day the expected rate? Thanks, Chuck Yes, that's the best rate available from both of the downtown (walkable) hotels. Just an update that we only have a few rooms left in our block at the Marriott. Please book ASAP if you haven't -- Jordan O'Mara jomara at redhat.com Red Hat Engineering, Raleigh pgplWdMerEsnR.pgp Description: PGP signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo] Openstack and SQLAlchemy
woot! From: Mike Bayer [mba...@redhat.com] Sent: Monday, June 30, 2014 1:56 PM To: OpenStack Development Mailing List (not for usage questions) Subject: [openstack-dev] [oslo] Openstack and SQLAlchemy Hi all - For those who don't know me, I'm Mike Bayer, creator/maintainer of SQLAlchemy, Alembic migrations and Dogpile caching. In the past month I've become a full time Openstack developer working for Red Hat, given the task of carrying Openstack's database integration story forward. To that extent I am focused on the oslo.db project which going forward will serve as the basis for database patterns used by other Openstack applications. I've summarized what I've learned from the community over the past month in a wiki entry at: https://wiki.openstack.org/wiki/Openstack_and_SQLAlchemy The page also refers to an ORM performance proof of concept which you can see at https://github.com/zzzeek/nova_poc. The goal of this wiki page is to publish to the community what's come up for me so far, to get additional information and comments, and finally to help me narrow down the areas in which the community would most benefit by my contributions. I'd like to get a discussion going here, on the wiki, on IRC (where I am on freenode with the nickname zzzeek) with the goal of solidifying the blueprints, issues, and SQLAlchemy / Alembic features I'll be focusing on as well as recruiting contributors to help in all those areas. I would welcome contributors on the SQLAlchemy / Alembic projects directly as well, as we have many areas that are directly applicable to Openstack. I'd like to thank Red Hat and the Openstack community for welcoming me on board and I'm looking forward to digging in more deeply in the coming months! - mike ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron][third-party] Simple and robust CI script?
Not sure if these are “minimalist” but at least they setup automagically, so you don’t need to do it from scratch: You can check out these repos which automate the 3rd party ci setup: Jay Pipe’s solution: https://github.com/jaypipes/os-ext-testing https://github.com/jaypipes/os-ext-testing-data My WIP fork of Jay Pipe’s that also includes Nodepool: https://github.com/rasselin/os-ext-testing https://github.com/rasselin/os-ext-testing-data Ramy From: Luke Gorrie [mailto:l...@tail-f.com] Sent: Monday, June 30, 2014 9:26 AM To: OpenStack Development Mailing List (not for usage questions) Subject: Re: [openstack-dev] [neutron][third-party] Simple and robust CI script? On 30 June 2014 17:34, Kyle Mestery mest...@noironetworks.commailto:mest...@noironetworks.com wrote: It would be great to get you to join the 3rd Party meeting [1] in #openstack-meeting at 1800UTC to discuss this. Can you make it today Luke? Yes, I'll be there. Currently I'm looking into the simplest 3rd party CI that could possibly work which would be less sophisticated than devstack-gate but easier for the operator to understand and be responsible for. (Or: CI in 100 lines of shell with no Jenkins or other large pieces of software to install.) I'm on #openstack-neutron if anybody wants to kick around ideas or share scripts that I can borrow ideas from (thanks ODL gang for doing this already). Cheers, -Luke ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [Nova] Meeting this week?
On Jun 29, 2014, at 17:01, Michael Still mi...@stillhq.com wrote: Hi. The meeting this week would be on the 3rd of July, which I assume means that many people will be out of the office. Do people think its worth running the meeting or shall we give this week a miss? July 3 is a company holiday where I am, so I'm also +1 on skipping this week. signature.asc Description: Message signed with OpenPGP using GPGMail ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron][third-party] Simple and robust CI script?
On 30 June 2014 19:37, Asselin, Ramy ramy.asse...@hp.com wrote: Not sure if these are “minimalist” but at least they setup automagically, so you don’t need to do it from scratch: I'm aiming to do exactly the opposite of this i.e. no automagic. My experience has been that the really heavy-duty CI setups are too difficult to understand and troubleshoot for end-users like me. I spent a week working on installing Jay's tools, with very generous help from Jay himself, and for me it was a dead end [*]. I now consider these to be tools for openstack-infra insiders rather than guys like me who are maintaining tiny drivers. I don't know if other people have had a different experience from me. It does seem like 3rd party CI is broadly a lot more problematic than advertised. I have seen only relatively few frank experience reports from people operating CIs and I suspect there is a rich and interesting untold story there :-). (Maybe it would be interesting to do a survey some day.) [*] http://openstack-dev-openstack.git.net/2014-March/030022.html other posts in that thread. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron][third-party] Simple and robust CI script?
I have a really early sketch of this project on Github now. shellci - OpenStack 3rd party CI in 100 lines of shell https://github.com/SnabbCo/shellci This is not finished yet but I'll try to use it for the new Neutron mech driver that I want to contribute to Juno. Ideas and encouragement welcome :-). Cheers, -Luke ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [OSSG][OSSN] Cinder SSH Pool will auto-accept SSH host signatures by default
-BEGIN PGP SIGNED MESSAGE- Hash: SHA1 Cinder SSH Pool will auto-accept SSH host signatures by default - --- ### Summary### In OpenStack releases prior to Juno, the SSH connection pool used by Cinder drivers to control SAN hosts will silently auto-accept SSH host fingerprints. This potentially allows for a man in the middle attack through the impersonation of a legitimate storage host. ### Affected Services / Software ### Cinder, Icehouse, Havana, Grizzly, Folsom ### Discussion ### Cinder drivers for controlling SAN hardware communicate with storage hosts over SSH. To facilitate creation of these drivers, Cinder provides a utility mechanism to manage pooled SSH connections. This connection pool is using a policy that will silently accept the SSH fingerprint of any unknown host when it first connects. However, it is not properly maintaing the list of known hosts and will thus permit connections to a host regardless of the SSH fingerprint presented. This impacts all drivers built using the utility. At the time of writing these drivers include, but may not be limited to: - - Solaris ISCSI driver - - HP LeftHand SAN ISCSI driver - - Huawei OceanStor T series and Dorado series storage arrays - - Dell EqualLogic Storage - - IBM Storwize SVC In the event that a malicious adversary has a point of presence on the storage network, they could undermine network communications between Cinder and the SAN host. Should an adversary manage to impersonate the storage host, Cinder will silently accept the newly presented fingerprint of the bogus host and allow the connection. This behaviour constitutes a typical Man in the Middle attack that could intercept and manipulate communications with the storage host, possibly leaking login credentials. If login credentials can be acquired, then direct interaction with the legitimate storage host becomes possible. This could result in Cinder volumes being accessed or modified to export compromised code and data to other services. The presence of this defect can be detected by initially connecting to a storage host and then re-generating that hosts local SSH details. Cinder will still allow connections to the host despite its now modified fingerprint. This is the default configuration. ### Recommended Actions ### Deployers should pay attention to the SSH interface between the Cinder driver and the SAN host and take appropriate measures to defend the storage network. These measures could include physical network isolation or placing an Intrusion Detection System on the network. The IDS should detect attacks such as ARP table poisoning, DHCP spoofing or DNS forgery that could be used to impersonate a SAN host and enact an Man in the Middle attack. ### Contacts / References ### This OSSN : https://wiki.openstack.org/wiki/OSSN/OSSN-0019 Original LaunchPad Bug : https://bugs.launchpad.net/cinder/+bug/1320056 OpenStack Security ML : openstack-secur...@lists.openstack.org OpenStack Security Group : https://launchpad.net/~openstack-ossg -BEGIN PGP SIGNATURE- Version: GnuPG v1 Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/ iQEcBAEBAgAGBQJTsbC4AAoJEJa+6E7Ri+EVK44IAKmfcak6QIBtd9QT4bC013/8 083WqUa6rnhX7jGtRkwm6lELVDw5Vk8jUpNYqnu7W7X+7+q24S4R/52UrxJBE8f7 dkxIcTS6Nx9qxGeoVVWFa4QLEuuG82K0PYhyEasbn7m8e672QeqLVHxUzAH7L1Yg hyXyZvxpN3bz38PpOKjf2Sj4lG3g1DZkZTL1cW2HIla9ZFiqZ9IMa5f2FItrgLEJ epLtsEhkhM/M/Nk9Qqbvvn0Ir3WTFN0l43hGJP4iF+frEsSewZqDXwNafVXl8k9v 4He6I1gpR2bpmYGIv4Bd+9jnjuiujFUfIIZKQg4LvNpH0FB+DqvCGUS5A0D1WjU= =SGiN -END PGP SIGNATURE- ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [TripleO] Use MariaDB by default on Fedora
Excerpts from Michael Kerrin's message of 2014-06-30 02:16:07 -0700: I am trying to finish off https://review.openstack.org/#/c/90134 - percona xtradb cluster for debian based system. I have read into this thread that I can error out on Redhat systems when trying to install percona and tell them to use mariadb instead, percona isn't support here. Is this correct? Probably. But if CI for Fedora breaks as a result you'll need a solution first. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [third party] - minimum response time for 3rd party CI responses
Hello all, The subject of 3rd party CI voting responses came up in the 3rd-party IRC meeting today.[1] We would like to get feedback from the larger dev community on what acceptable response times are for third party CI systems. As a maintainer of a small CI system that tends to get backed up during milestone rush hours, it would be nice if we were allowed up to 12 hours. However, as a developer this seems like too long to have to wait for the results of a patch. One option might be to peg the response time to the main CI response assuming that it gets backed up in a similar manner. What would be acceptable to the community? 1. http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-30-18.01.html -- Kevin Benton ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard
On 06/29/2014 07:59 PM, Anita Kuno wrote: On 06/29/2014 07:43 PM, Anita Kuno wrote: On 06/29/2014 03:25 PM, Ilya Shakhat wrote: Hi! During last couple weeks there is an increasing demand on tracking 3rd-party CI statuses. We at Stackalytics decided to be in trend and (with some inspiration from Salvatore's proposal) implemented report that shows summary on external CI status. The initial version is available for Neutron - http://stackalytics.com/report/ci/neutron/7 The report shows summary of all CI jobs during specified period of time, including: * stats of runs on merged patch sets: - total number of runs - success rate (success to total ratio) - time of the latest run - last test result * stats for all patch sets (the same set as for merged) * last test results for every merged patch set grouped by days (useful to see how different CIs correlate with each other and how often they run) Under merged patch set report means the last patch set in the merged change request, thus it is almost the same as the trunk code. CI configuration is taken from DriverLog's default data https://git.openstack.org/cgit/stackforge/driverlog/tree/etc/default_data.json. Standard Stackalytics screen is also available for CIs - http://stackalytics.com/?metric=ci, including votes breakdown and activity log. Since this is the first version there are some open questions: * Currently report shows results per CI id, but there are CIs that run tests against multiple drivers and this case is not supported. What would be more useful: to get stats for a driver or for CI? * Most CIs run tests when patch set is posted. So even if change request is merged within selected time period corresponding CI results may be missing. * Patterns for non-voting CIs need to be verified. For example Cisco CI now runs 5 jobs, but DriverLog data still contains old data. Thanks, Ilya 2014-06-16 17:52 GMT+04:00 Salvatore Orlando sorla...@nicira.com: However, it would be great if we could start devising a solution for having health reports from the various CI systems. This report should report the following kind of information: - timestamp of last run - timestamp of last vote (a system might start job which then get aborted for CI infra problems) - % of success vs failures (not sure how important is that one but provides a metric of comparison with upstream jenkins) - % of disagreements with jenkins (this might allow us to quickly spot those CI systems which are randomly -1'ing patches) The percentage metrics might be taken over a 48 hours or 7 days interval, or both. Does this idea sound reasonable? ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Ilya: I look forward to hearing more about this dashboard and ensuring you or someone else associated with this dashboard are available for questions at the third party meeting tomorrow: https://wiki.openstack.org/wiki/Meetings/ThirdParty We missed you last week. Thanks Ilya, Anita. And one question I will have when we discuss this is regarding this statement: Green cell - tests ran successfully, Currently we don't have community consensus around the use of the statement tests ran successfully regarding third party ci systems. This is as statement, you recall, we had discussed at the third party meeting when we talked about driverlog. 18:46:02 krtaylor but what does CI tested really mean? just running tests? or tested to pass some level of requirements? http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-16-18.00.log.html Using a statement to convey success prior to having a definition from the community about what the statement means will create confusion in the community and frustration from folks, including myself, who are willing to have the conversation about what it means, who feel circumvented by this second use of a phrase which implies a decided meaning where none yet exists. Please participate in conversations around the definition of phrases of success and failure regarding third party systems and point to the logs where consensus is reached prior it its use in future. In addition to attending the third party meeting, please attend the infra meeting or the qa meeting, and hopefully meetings that include programs that have interactions with third party ci systems including nova, neutron, and cinder (if there are other programs interacting with third party ci systems please attend the third party meeting so I know about you). Thanks Ilya, I look forward to our future discussions, Anita. As an update, Ilya did attend the third party meeting today, thank you Ilya. I am disappointed to realize that Ilya (or stackalytics, I don't know where this is coming from) is unwilling to cease making up definitions of success
[openstack-dev] [neutron] [third-party] An update on third-party CI systems in Neutron
Hi folks: I wanted to give an update to people following the third-party CI conversation for Neutron. I sent an email out with a status update in June [1]. In that email, I had indicated that in-tree plugins and drivers needed to have functioning CI running by Juno-2. That is still the case, and I've seen some great work and efforts by many people to make this happen. But for systems which are not functional by Juno-2, we will look to remove code in Juno-3. The etherpad tracking this effort [3], while not perhaps the ideal place, is where we're doing this now. I encourage people to update this as their CI systems become functional. I encourage anyone operating a CI system for Neutron (or other OpenStack projects) to also join the weekly meeting to discuss issues, share ideas, and help each out out [2]. If you're having issues, this meeting, the ML, and #openstack-[infra,qa] are the best places to reach out for help. Thanks! Kyle [1] http://lists.openstack.org/pipermail/openstack-dev/2014-June/037665.html [2] https://wiki.openstack.org/wiki/Meetings/ThirdParty [3] https://etherpad.openstack.org/p/ZLp9Ow3tNq ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] ovs-neutron-agent wipes out all flows on startup
On Mon, Jun 30, 2014 at 2:11 PM, Paul Ward wpw...@linux.vnet.ibm.com wrote: The current design for ovs-neutron-agent is that it will wipe out all flows configured on the system when it starts up, recreating them for each neutron port it's aware of. This has a not-so-desirable side effects that there's a temporary hiccup in network connectivity for the VMs on the host. My questions to the list: Is there a reason it was designed this way (other than Everything on the system must be managed by OpenStack)? Is there ongoing work to address this or would it be a worthwhile contribution from our side? This was actually the result of a bug fix in Juno-1 [1]. As reported by the TripleO folks, having the agent default to setting up a NORMAL flow added may have allowed for VMs to talk to each other, but it was also a huge security hole. I'm curious what ideas you have around this, though. Thanks, Kyle [1] https://bugs.launchpad.net/tripleo/+bug/1290486 and https://bugs.launchpad.net/neutron/+bug/1324703 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard
On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote: I am disappointed to realize that Ilya (or stackalytics, I don't know where this is coming from) is unwilling to cease making up definitions of success for third party ci systems to allow the openstack community to arrive at its own definition. There is indeed a risk that the new dashboards won't give a meaningful view of whether a 3rd party CI is voting correctly or not. However, there is an elephant in the room and a much more important problem: To measure how accurately a CI is voting says much more about a driver author's Gerrit-fu ability to operate a CI system than it does about whether the code they have contributed to OpenStack actually works, and the latter is what is actually important. To my mind the whole 3rd party testing discussion should refocus on helping developers maintain good working code and less on waving you will be kicked out of OpenStack if you don't keep your swarm of nebulous daemons running 24/7. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard
On 06/30/2014 03:27 PM, Luke Gorrie wrote: On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote: I am disappointed to realize that Ilya (or stackalytics, I don't know where this is coming from) is unwilling to cease making up definitions of success for third party ci systems to allow the openstack community to arrive at its own definition. There is indeed a risk that the new dashboards won't give a meaningful view of whether a 3rd party CI is voting correctly or not. However, there is an elephant in the room and a much more important problem: To measure how accurately a CI is voting says much more about a driver author's Gerrit-fu ability to operate a CI system than it does about whether the code they have contributed to OpenStack actually works, and the latter is what is actually important. To my mind the whole 3rd party testing discussion should refocus on helping developers maintain good working code and less on waving you will be kicked out of OpenStack if you don't keep your swarm of nebulous daemons running 24/7. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Well I am glad you have reached that conclusion, Luke. Given that I gave you months of my time with that exact approach and you squandered it away, claiming I wasn't doing enough to help you, I am glad you have finally reached a productive place. I hope your direction produces results. Thanks Luke, Anita. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] ovs-neutron-agent wipes out all flows on startup
Indeed a blueprint has been filed (on Launchpad, not neutron-specs) on this already[0], but there has been no work on this as far as I can tell. I think it would be worthwhile contribution. Amir [0] https://blueprints.launchpad.net/neutron/+spec/neutron-agent-soft-restart From: Paul Ward [wpw...@linux.vnet.ibm.com] Sent: Monday, June 30, 2014 2:11 PM To: OpenStack Development Mailing List Subject: [openstack-dev] [neutron] ovs-neutron-agent wipes out all flows on startup The current design for ovs-neutron-agent is that it will wipe out all flows configured on the system when it starts up, recreating them for each neutron port it's aware of. This has a not-so-desirable side effects that there's a temporary hiccup in network connectivity for the VMs on the host. My questions to the list: Is there a reason it was designed this way (other than Everything on the system must be managed by OpenStack)? Is there ongoing work to address this or would it be a worthwhile contribution from our side? ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard
Well, Luke, this is collaborative effort by everybody. By having these CI systems in place ensures that one person's code does not break other person's code and vice versa. Therefore, having these CI systems operational and voting 24x7 is a critical step in achieving this goal. However, the details as to how and what should be tested is definitely debatable and the team has done excellent job in converging on that. Now, as to the issue at hand which Anita is describing, I attended the meeting this morning and was very pleased with the debate that took place and the definition as to Sucess On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie l...@tail-f.com wrote: On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote: I am disappointed to realize that Ilya (or stackalytics, I don't know where this is coming from) is unwilling to cease making up definitions of success for third party ci systems to allow the openstack community to arrive at its own definition. There is indeed a risk that the new dashboards won't give a meaningful view of whether a 3rd party CI is voting correctly or not. However, there is an elephant in the room and a much more important problem: To measure how accurately a CI is voting says much more about a driver author's Gerrit-fu ability to operate a CI system than it does about whether the code they have contributed to OpenStack actually works, and the latter is what is actually important. To my mind the whole 3rd party testing discussion should refocus on helping developers maintain good working code and less on waving you will be kicked out of OpenStack if you don't keep your swarm of nebulous daemons running 24/7. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [qa] checking success response in clients
We approved https://github.com/openstack/qa-specs/blob/master/specs/client-checks-success.rst which recommends that checking of correct success codes be moved to the tempest clients. This has been done for the image tests but not others yet. But new client/test code coming in should definitely be doing the checks in the client rather then the test bodies. Here is the image change for reference: https://review.openstack.org/#/c/101310/ -David ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [third-party-ci][neutron] What is Success exactly?
Hi Stackers, Some recent ML threads [1] and a hot IRC meeting today [2] brought up some legitimate questions around how a newly-proposed Stackalytics report page for Neutron External CI systems [2] represented the results of an external CI system as successful or not. First, I want to say that Ilya and all those involved in the Stackalytics program simply want to provide the most accurate information to developers in a format that is easily consumed. While there need to be some changes in how data is shown (and the wording of things like Tests Succeeded), I hope that the community knows there isn't any ill intent on the part of Mirantis or anyone who works on Stackalytics. OK, so let's keep the conversation civil -- we're all working towards the same goals of transparency and accuracy. :) Alright, now, Anita and Kurt Taylor were asking a very poignant question: But what does CI tested really mean? just running tests? or tested to pass some level of requirements? In this nascent world of external CI systems, we have a set of issues that we need to resolve: 1) All of the CI systems are different. Some run Bash scripts. Some run Jenkins slaves and devstack-gate scripts. Others run custom Python code that spawns VMs and publishes logs to some public domain. As a community, we need to decide whether it is worth putting in the effort to create a single, unified, installable and runnable CI system, so that we can legitimately say all of the external systems are identical, with the exception of the driver code for vendor X being substituted in the Neutron codebase. If the goal of the external CI systems is to produce reliable, consistent results, I feel the answer to the above is yes, but I'm interested to hear what others think. Frankly, in the world of benchmarks, it would be unthinkable to say go ahead and everyone run your own benchmark suite, because you would get wildly different results. A similar problem has emerged here. 2) There is no mediation or verification that the external CI system is actually testing anything at all As a community, we need to decide whether the current system of self-policing should continue. If it should, then language on reports like [3] should be very clear that any numbers derived from such systems should be taken with a grain of salt. Use of the word Success should be avoided, as it has connotations (in English, at least) that the result has been verified, which is simply not the case as long as no verification or mediation occurs for any external CI system. 3) There is no clear indication of what tests are being run, and therefore there is no clear indication of what success is I think we can all agree that a test has three possible outcomes: pass, fail, and skip. The results of a test suite run therefore is nothing more than the aggregation of which tests passed, which failed, and which were skipped. As a community, we must document, for each project, what are expected set of tests that must be run for each merged patch into the project's source tree. This documentation should be discoverable so that reports like [3] can be crystal-clear on what the data shown actually means. The report is simply displaying the data it receives from Gerrit. The community needs to be proactive in saying this is what is expected to be tested. This alone would allow the report to give information such as External CI system ABC performed the expected tests. X tests passed. Y tests failed. Z tests were skipped. Likewise, it would also make it possible for the report to give information such as External CI system DEF did not perform the expected tests., which is excellent information in and of itself. === In thinking about the likely answers to the above questions, I believe it would be prudent to change the Stackalytics report in question [3] in the following ways: a. Change the Success % column header to % Reported +1 Votes b. Change the phrase Green cell - tests ran successfully, red cell - tests failed to Green cell - System voted +1, red cell - System voted -1 and then, when we have more and better data (for example, # tests passed, failed, skipped, etc), we can provide more detailed information than just reported +1 or not. Thoughts? Best, -jay [1] http://lists.openstack.org/pipermail/openstack-dev/2014-June/038933.html [2] http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-30-18.01.log.html [3] http://stackalytics.com/report/ci/neutron/7 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [all] Latest revision of log guidelines
The latest revisions of the logging guidelines is available in the nova-specs repo - https://review.openstack.org/#/c/91446/ I tried to integrate the comments from the last go around into that. I do think we're at a state where if we believe this is a set of first pass guidelines that we're good with, we take them as they are, and move forward with fixes during this cycle. I suspect things such as the AUDIT removal are actually doable on most projects, and would simplify things. I've started trying to fix a few of these in nova based on the assumption we're going to move forward here. https://review.openstack.org/#/c/103535/ - which removes one of the secret decoder rings merged today. http://logs.openstack.org/35/103535/3/gate/gate-tempest-dsvm-full/20e8a0e/logs/screen-n-cpu.txt.gz#_2014-06-30_16_56_32_723 -Sean -- Sean Dague http://dague.net signature.asc Description: OpenPGP digital signature ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [neutron] [third-party] Neutron 3rd Party CI status dashboard
Sorry, accidentally hit the wrong key and message went out. Was making a mention about the definition of Success. I thought the debate in the meeting was very productive - when a CI posts a +1 that is success, and when a CI posts a -1 (or no vote with comment) is also a success - as this reflects that the CI is doing what it is suppose to do. So, when it comes to stackalytics, it is more critical to show if a given CI is operational or not - and for how long? Another thing we can debate is how to present the +1/-1 votes by a given CI - unless we have some benchmark, it will be hard to consider success/failure. So, I am of the opinion that, initially, we only report on the operational status and duration of the CIs, and a counter of +1 and -1 votes over a period of time. For example, looking at Arista CI, it has casted 7,958 votes so far and it has been operational for past 6 months. This information is not available anywhere - hence, presenting this kind of information on a dashboard created by Ilya would be very useful to the community as well to the vendors.. thoughts? -Sukhdev On Mon, Jun 30, 2014 at 12:49 PM, Sukhdev Kapur sukhdevka...@gmail.com wrote: Well, Luke, this is collaborative effort by everybody. By having these CI systems in place ensures that one person's code does not break other person's code and vice versa. Therefore, having these CI systems operational and voting 24x7 is a critical step in achieving this goal. However, the details as to how and what should be tested is definitely debatable and the team has done excellent job in converging on that. Now, as to the issue at hand which Anita is describing, I attended the meeting this morning and was very pleased with the debate that took place and the definition as to Sucess On Mon, Jun 30, 2014 at 12:27 PM, Luke Gorrie l...@tail-f.com wrote: On 30 June 2014 21:08, Anita Kuno ante...@anteaya.info wrote: I am disappointed to realize that Ilya (or stackalytics, I don't know where this is coming from) is unwilling to cease making up definitions of success for third party ci systems to allow the openstack community to arrive at its own definition. There is indeed a risk that the new dashboards won't give a meaningful view of whether a 3rd party CI is voting correctly or not. However, there is an elephant in the room and a much more important problem: To measure how accurately a CI is voting says much more about a driver author's Gerrit-fu ability to operate a CI system than it does about whether the code they have contributed to OpenStack actually works, and the latter is what is actually important. To my mind the whole 3rd party testing discussion should refocus on helping developers maintain good working code and less on waving you will be kicked out of OpenStack if you don't keep your swarm of nebulous daemons running 24/7. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [third-party-ci][neutron] What is Success exactly?
Hi Jay, Couple of points. I support the fact that we need to define what is success is. I believe that the metrics that should be used are Voted +1 and Skipped. But to certain valid case, I would say that the Voted -1 is really mostly a metric of bad health of a CI. Most of the -1 are due to environment issue, configuration problem, etc... In my case, the -1 are done manually since I want to avoid giving some extra work to the developer. That are some possible solutions ? On the Jenkins, I think we could develop a script that will parse the result html file. Jenkins will then vote (+1, 0, -1) on the behalf of the 3rd party CI. - It would prevent the abusive +1 - If the result HTML is empty, it would indicate the CI health is bad - if all the result are failing, it would also indicate that CI health is bad Franck Franck Franck On Mon, Jun 30, 2014 at 1:22 PM, Jay Pipes jaypi...@gmail.com wrote: Hi Stackers, Some recent ML threads [1] and a hot IRC meeting today [2] brought up some legitimate questions around how a newly-proposed Stackalytics report page for Neutron External CI systems [2] represented the results of an external CI system as successful or not. First, I want to say that Ilya and all those involved in the Stackalytics program simply want to provide the most accurate information to developers in a format that is easily consumed. While there need to be some changes in how data is shown (and the wording of things like Tests Succeeded), I hope that the community knows there isn't any ill intent on the part of Mirantis or anyone who works on Stackalytics. OK, so let's keep the conversation civil -- we're all working towards the same goals of transparency and accuracy. :) Alright, now, Anita and Kurt Taylor were asking a very poignant question: But what does CI tested really mean? just running tests? or tested to pass some level of requirements? In this nascent world of external CI systems, we have a set of issues that we need to resolve: 1) All of the CI systems are different. Some run Bash scripts. Some run Jenkins slaves and devstack-gate scripts. Others run custom Python code that spawns VMs and publishes logs to some public domain. As a community, we need to decide whether it is worth putting in the effort to create a single, unified, installable and runnable CI system, so that we can legitimately say all of the external systems are identical, with the exception of the driver code for vendor X being substituted in the Neutron codebase. If the goal of the external CI systems is to produce reliable, consistent results, I feel the answer to the above is yes, but I'm interested to hear what others think. Frankly, in the world of benchmarks, it would be unthinkable to say go ahead and everyone run your own benchmark suite, because you would get wildly different results. A similar problem has emerged here. 2) There is no mediation or verification that the external CI system is actually testing anything at all As a community, we need to decide whether the current system of self-policing should continue. If it should, then language on reports like [3] should be very clear that any numbers derived from such systems should be taken with a grain of salt. Use of the word Success should be avoided, as it has connotations (in English, at least) that the result has been verified, which is simply not the case as long as no verification or mediation occurs for any external CI system. 3) There is no clear indication of what tests are being run, and therefore there is no clear indication of what success is I think we can all agree that a test has three possible outcomes: pass, fail, and skip. The results of a test suite run therefore is nothing more than the aggregation of which tests passed, which failed, and which were skipped. As a community, we must document, for each project, what are expected set of tests that must be run for each merged patch into the project's source tree. This documentation should be discoverable so that reports like [3] can be crystal-clear on what the data shown actually means. The report is simply displaying the data it receives from Gerrit. The community needs to be proactive in saying this is what is expected to be tested. This alone would allow the report to give information such as External CI system ABC performed the expected tests. X tests passed. Y tests failed. Z tests were skipped. Likewise, it would also make it possible for the report to give information such as External CI system DEF did not perform the expected tests., which is excellent information in and of itself. === In thinking about the likely answers to the above questions, I believe it would be prudent to change the Stackalytics report in question [3] in the following ways: a. Change the Success % column header to % Reported +1 Votes b. Change the phrase Green cell - tests ran successfully, red cell - tests failed
[openstack-dev] [all][specs] Please stop doing specs for any changes in projects
Hi all, Specs are interesting idea, that may be really useful, when you need to discuss large topics: 1) work on API 2) Large refactoring 3) Large features 4) Performance, scale, ha, security issues that requires big changes And I really dislike idea of adding spec for every patch. Especially when changes (features) are small, don't affect too much, and they are optional. It really kills OpenStack. And will drastically slow down process of contribution and reduce amount of contributors. Thanks. Best regards, Boris Pavlovic ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [oslo][messaging] Further improvements and refactoring
I'm far from an oslo.messaging expert, but a few general thoughts below. On 06/30/2014 02:34 PM, Alexei Kornienko wrote: Hello, My understanding is that your analysis is mostly based on running a profiler against the code. Network operations can be bottlenecked in other places. You compare 'simple script using kombu' with 'script using oslo.messaging'. You don't compare script using oslo.messaging before refactoring and 'after that. The latter would show whether refactoring was worth the effort. Your test shows that oslo.messaging performance sucks, but it's not definite that hotspots you've revealed, once fixed, will show huge boost. My concern is that it may turn out that once all the effort to refactor the code is done, we won't see major difference. So we need base numbers, and performance tests would be a great helper here. It's really sad for me to see so little faith in what I'm saying. The test I've done using plain kombu driver was needed exactly to check that network is not the bottleneck for messaging performance. If you don't believe in my performance analysis we could ask someone else to do their own research and provide results. The problem is that extremely simple test cases are often not representative of overall performance, so comparing a purpose-built test doing a single thing as fast as possible to a full library that has to be able to handle all of OpenStack's messaging against every supported back end isn't sufficient on its own to convince me that there is a rewrite all the things issue here. Problem with refactoring that I'm planning is that it's not a minor refactoring that can be applied in one patch but it's the whole library rewritten from scratch. Which is exactly why we want to make sure it's something that needs to be done before heading down that path. I know I've wasted more time than I'd like to admit optimizing the wrong code paths, only to find that my changes made a .1% difference because I was mistaken about what the bottleneck was. Add to that the fact that we're _just_ completing the migration to oslo.messaging in the first place and I hope you can understand why no one wants to undertake another massive, possibly compatibility breaking, refactoring unless we're absolutely certain it's the only way to address the performance limitations of the existing code. Existing messaging code was written long long time ago (in a galaxy far far away maybe?) and it was copy-pasted directly from nova. It was not built as a library and it was never intended to be used outside of nova. This isn't really true anymore. The oslo.messaging code underwent significant changes in the move from the incubator rpc module to the oslo.messaging library. One of the major points of emphasis in all Oslo graduations is to make sure the new lib has a proper API and isn't just a naive copy-paste of the existing code. Some parts of it cannot even work normally cause it was not designed to work with drivers like zeromq (matchmaker stuff). The reason I've raised this question on the mailing list was to get some agreement about future plans of oslo.messaging development and start working on it in coordination with community. For now I don't see any actions plan emerging from it. I would like to see us bringing more constructive ideas about what should be done. If you think that first action should be profiling lets discuss how it should be implemented (cause it works for me just fine on my local PC). I guess we'll need to define some basic scenarios that would show us overall performance of the library. There are a lot of questions that should be answered to implement this: Where such tests would run (jenking, local PC, devstack VM)? How such scenarios should look like? How do we measure performance (cProfile, etc.)? How do we collect results? How do we analyze results to find bottlenecks? etc. Another option would be to spend some of my free time implementing mentioned refactoring (as I see it) and show you the results of performance testing compared with existing code. The only problem with such approach is that my code won't be oslo.messaging and it won't be accepted by community. It may be drop in base for v2.0 but I'm afraid this won't be acceptable either. Regards, Alexei Kornienko 2014-06-30 17:51 GMT+03:00 Gordon Sim g...@redhat.com: On 06/30/2014 12:22 PM, Ihar Hrachyshka wrote: Alexei Kornienko wrote: Some performance tests may be introduced but they would be more like functional tests since they require setup of actual messaging server (rabbit, etc.). Yes. I think we already have some. F.e. tests/drivers/test_impl_qpid.py attempts to use local Qpid server (backing up to fake server if it's not available). I always get failures when there is a real qpidd service listening on the expected port. Does anyone else see this? ___ OpenStack-dev
Re: [openstack-dev] [infra][oslo][neutron] Need help getting oslo.messaging 1.4.0.0a2 in global requirements
On Mon, 2014-06-30 at 16:52 +, Paul Michali (pcm) wrote: I have out for review 103536 to add this version to global requirements, so that Neutron has an oslo fix (review 102909) for encoding failure, which affects some gate runs. This review for global requirements is failing requirements check (http://logs.openstack.org/36/103536/1/check/check-requirements-integration-dsvm/6d9581c/console.html#_2014-06-30_12_34_56_921). I did a recheck bug 1334898, but see the same error, with the release not found, even though it is in PyPI. Infra folks say this is a known issue with pushing out pre-releases. Do we have a work-around? Any proposed solution to try? That makes two oslo alpha releases which are failing openstack/requirements checks: https://review.openstack.org/103256 https://review.openstack.org/103536 and an issue with the py27 stable/icehouse test jobs seemingly pulling in oslo.messaging 1.4.0.0a2: http://lists.openstack.org/pipermail/openstack-dev/2014-June/039021.html and these comments on IRC: http://eavesdrop.openstack.org/irclogs/%23openstack-infra/%23openstack-infra.2014-06-30.log 2014-06-30T15:27:33 pcm__ hi. Need help with getting latest oslo.messaging release added to global requirements. Can someone advise on the issues I see. 2014-06-30T15:28:06 mordred pcm__: there are issues adding oslo pre-releases to the mirror right now - we're working on a solution ... so you're not alone at least :) 2014-06-30T15:29:02 pcm__ mordred: Jenkins failed saying that it could not find the release, but it is available. 2014-06-30T15:29:31 bknudson pcm__: mordred: is the fix to remove the check for --no-use-wheel in the check-requirements-integration-dsvm ? 2014-06-30T15:29:55 mordred bknudson: nope. it's to completely change our mirroring infrastructure :) Presumably there's more information somewhere on what solution infra are working on, but that's all I got ... We knew this pre-release-with-wheels stuff was going to be a little rocky, so this isn't surprising. Hopefully it'll get sorted out soon. Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all][specs] Please stop doing specs for any changes in projects
Thanks boris for starting this topic, There is a balance here that needs to be worked out and I've seen specs start to turn into requirements for every single patch (even if the patch is pretty small). I hope we can rework the 'balance in the force' to avoid being so strict that every little thing requires a spec. This will not end well for us as a community. How have others thought the spec process has worked out so far? To much overhead, to little…? I personally am of the opinion that specs should be used for large topics (defining large is of course arbitrary); and I hope we find the right balance to avoid scaring everyone away from working with openstack. Maybe all of this is part of openstack maturing, I'm not sure, but it'd be great if we could have some guidelines around when is a spec needed and when isn't it and take it into consideration when requesting a spec that the person you have requested may get frustrated and just leave the community (and we must not have this happen) if you ask for it without explaining why and how clearly. From: Boris Pavlovic bpavlo...@mirantis.commailto:bpavlo...@mirantis.com Reply-To: OpenStack Development Mailing List (not for usage questions) openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org Date: Monday, June 30, 2014 at 2:11 PM To: OpenStack Development Mailing List openstack-dev@lists.openstack.orgmailto:openstack-dev@lists.openstack.org Subject: [openstack-dev] [all][specs] Please stop doing specs for any changes in projects Hi all, Specs are interesting idea, that may be really useful, when you need to discuss large topics: 1) work on API 2) Large refactoring 3) Large features 4) Performance, scale, ha, security issues that requires big changes And I really dislike idea of adding spec for every patch. Especially when changes (features) are small, don't affect too much, and they are optional. It really kills OpenStack. And will drastically slow down process of contribution and reduce amount of contributors. Thanks. Best regards, Boris Pavlovic ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [OpenStack-Dev] OSLO messaging update and icehouse config files
On Mon, Jun 30, 2014 at 3:17 PM, Mark McLoughlin mar...@redhat.com wrote: On Mon, 2014-06-30 at 12:04 -0600, John Griffith wrote: Hey Everyone, So I sent a note out yesterday asking about config changes brought in to Icehouse due to the OSLO Messaging update that went out over the week-end here. My initial email prior to realizing the update that caused the problem was OSLO Messaging update here [1]. (Periodic reminder that Oslo is not an acronym) In the meantime I tried updating the cinder.conf sample in Cinder's stable Icehouse branch, but noticed that py26 doesn't seem to pick up the changes when running the oslo conf generation tools against oslo.messaging. I haven't spent any time digging into this yet, was hoping that perhaps somebody from the OSLO team or somewhere else maybe had some insight as to what's going on here. Here's the patch I submitted that shows the failure on py26 and success on py27 [2]. I'll get around to this eventually if nobody else knows anything off the top of their head. Thanks, John [1]: http://lists.openstack.org/pipermail/openstack-dev/2014-June/038926.html [2]: https://review.openstack.org/#/c/103426/ Ok, that new cinder.conf.sample is showing changes caused by these oslo.messaging changes: https://review.openstack.org/101583 https://review.openstack.org/99291 Both of those changes were first released in 1.4.0.0a1 which is an alpha version targeting Juno and are not available in the 1.3.0 Icehouse version - i.e. 1.4.0.0a1 should not be used with stable/icehouse Cinder. It seems 1.3.0 *is* being used: http://logs.openstack.org/26/103426/1/check/gate-cinder-python26/5c6c1dd/console.html 2014-06-29 19:17:50.154 | oslo.messaging==1.3.0 and the output is just confusing: 2014-06-29 19:17:49.900 | --- /tmp/cinder.UtGHjm/cinder.conf.sample 2014-06-29 19:17:50.270071741 + 2014-06-29 19:17:49.900 | +++ etc/cinder/cinder.conf.sample 2014-06-29 19:10:48.396072037 + ... 2014-06-29 19:17:49.903 | +[matchmaker_redis] 2014-06-29 19:17:49.903 | + i.e. it's showing that the file you proposed was generated with 1.4.0.0a1 and the file generated during the test job was generated with 1.3.0. Which is what I'd expect - the update you proposed is not appropriate for stable/icehouse. So why is the py27 job passing? http://logs.openstack.org/26/103426/1/check/gate-cinder-python27/7844c61/console.html 2014-06-29 19:21:12.875 | oslo.messaging==1.4.0.0a2 That's the problem right there - 1.4.0.0a2 should not be getting installed on the stable/icehouse branch. I'm not sure why it is. Someone on #openstack-infra could probably help figure it out. Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Thanks Mark... so the problem is Oslo messaging in requirements is = in stable/icehouse [1](please note I used Oslo not OSLO). [1]: https://github.com/openstack/requirements/blob/stable/icehouse/global-requirements.txt#L49 Thanks for pointing me in the right direction. John ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [all][specs] Please stop doing specs for any changes in projects
Excerpts from Boris Pavlovic's message of 2014-06-30 14:11:08 -0700: Hi all, Specs are interesting idea, that may be really useful, when you need to discuss large topics: 1) work on API 2) Large refactoring 3) Large features 4) Performance, scale, ha, security issues that requires big changes And I really dislike idea of adding spec for every patch. Especially when changes (features) are small, don't affect too much, and they are optional. It really kills OpenStack. And will drastically slow down process of contribution and reduce amount of contributors. Who says there needs to be a spec for every patch? I agree with your items above. Any other change is likely just fixing a bug. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [nova][scheduler] Proposal: FairShareScheduler.
Hi Eric, definitely... In my view a FairShareScheduler could be a very interesting option for private clouds that support scientific communities. Basically this is the model used by batch systems in order to fully use the available resources. I'm very curious about the work that you are doing. Is it available in github? Belmiro -- Belmiro Moreira CERN Email: belmiro.more...@cern.ch IRC: belmoreira On Mon, Jun 30, 2014 at 4:05 PM, Eric Frizziero eric.frizzi...@pd.infn.it wrote: Hi All, we have analyzed the nova-scheduler component (FilterScheduler) in our Openstack installation used by some scientific teams. In our scenario, the cloud resources need to be distributed among the teams by considering the predefined share (e.g. quota) assigned to each team, the portion of the resources currently used and the resources they have already consumed. We have observed that: 1) User requests are sequentially processed (FIFO scheduling), i.e. FilterScheduler doesn't provide any dynamic priority algorithm; 2) User requests that cannot be satisfied (e.g. if resources are not available) fail and will be lost, i.e. on that scenario nova-scheduler doesn't provide any queuing of the requests; 3) OpenStack simply provides a static partitioning of resources among various projects / teams (use of quotas). If project/team 1 in a period is systematically underutilizing its quota and the project/team 2 instead is systematically saturating its quota, the only solution to give more resource to project/team 2 is a manual change (to be done by the admin) to the related quotas. The need to find a better approach to enable a more effective scheduling in Openstack becomes more and more evident when the number of the user requests to be handled increases significantly. This is a well known problem which has already been solved in the past for the Batch Systems. In order to solve those issues in our usage scenario of Openstack, we have developed a prototype of a pluggable scheduler, named FairShareScheduler, with the objective to extend the existing OpenStack scheduler (FilterScheduler) by integrating a (batch like) dynamic priority algorithm. The architecture of the FairShareScheduler is explicitly designed to provide a high scalability level. To all user requests will be assigned a priority value calculated by considering the share allocated to the user by the administrator and the evaluation of the effective resource usage consumed in the recent past. All requests will be inserted in a priority queue, and processed in parallel by a configurable pool of workers without interfering with the priority order. Moreover all significant information (e.g. priority queue) will be stored in a persistence layer in order to provide a fault tolerance mechanism while a proper logging system will annotate all relevant events, useful for auditing processing. In more detail, some features of the FairshareScheduler are: a) It assigns dynamically the proper priority to every new user requests; b) The priority of the queued requests will be recalculated periodically using the fairshare algorithm. This feature guarantees the usage of the cloud resources is distributed among users and groups by considering the portion of the cloud resources allocated to them (i.e. share) and the resources already consumed; c) all user requests will be inserted in a (persistent) priority queue and then processed asynchronously by the dedicated process (filtering + weighting phase) when compute resources are available; d) From the client point of view the queued requests remain in “Scheduling” state till the compute resources are available. No new states added: this prevents any possible interaction issue with the Openstack clients; e) User requests are dequeued by a pool of WorkerThreads (configurable), i.e. no sequential processing of the requests; f) The failed requests at filtering + weighting phase may be inserted again in the queue for n-times (configurable). We have integrated the FairShareScheduler in our Openstack installation (release HAVANA). We're now working to adapt the FairShareScheduler to the new release IceHouse. Does anyone have experiences in those issues found in our cloud scenario? Could the FairShareScheduler be useful for the Openstack community? In that case, we'll be happy to share our work. Any feedback/comment is welcome! Cheers, Eric. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [OpenStack-Dev] OSLO messaging update and icehouse config files
On Mon, 2014-06-30 at 15:35 -0600, John Griffith wrote: On Mon, Jun 30, 2014 at 3:17 PM, Mark McLoughlin mar...@redhat.com wrote: On Mon, 2014-06-30 at 12:04 -0600, John Griffith wrote: Hey Everyone, So I sent a note out yesterday asking about config changes brought in to Icehouse due to the OSLO Messaging update that went out over the week-end here. My initial email prior to realizing the update that caused the problem was OSLO Messaging update here [1]. (Periodic reminder that Oslo is not an acronym) In the meantime I tried updating the cinder.conf sample in Cinder's stable Icehouse branch, but noticed that py26 doesn't seem to pick up the changes when running the oslo conf generation tools against oslo.messaging. I haven't spent any time digging into this yet, was hoping that perhaps somebody from the OSLO team or somewhere else maybe had some insight as to what's going on here. Here's the patch I submitted that shows the failure on py26 and success on py27 [2]. I'll get around to this eventually if nobody else knows anything off the top of their head. Thanks, John [1]: http://lists.openstack.org/pipermail/openstack-dev/2014-June/038926.html [2]: https://review.openstack.org/#/c/103426/ Ok, that new cinder.conf.sample is showing changes caused by these oslo.messaging changes: https://review.openstack.org/101583 https://review.openstack.org/99291 Both of those changes were first released in 1.4.0.0a1 which is an alpha version targeting Juno and are not available in the 1.3.0 Icehouse version - i.e. 1.4.0.0a1 should not be used with stable/icehouse Cinder. It seems 1.3.0 *is* being used: http://logs.openstack.org/26/103426/1/check/gate-cinder-python26/5c6c1dd/console.html 2014-06-29 19:17:50.154 | oslo.messaging==1.3.0 and the output is just confusing: 2014-06-29 19:17:49.900 | --- /tmp/cinder.UtGHjm/cinder.conf.sample 2014-06-29 19:17:50.270071741 + 2014-06-29 19:17:49.900 | +++ etc/cinder/cinder.conf.sample 2014-06-29 19:10:48.396072037 + ... 2014-06-29 19:17:49.903 | +[matchmaker_redis] 2014-06-29 19:17:49.903 | + i.e. it's showing that the file you proposed was generated with 1.4.0.0a1 and the file generated during the test job was generated with 1.3.0. Which is what I'd expect - the update you proposed is not appropriate for stable/icehouse. So why is the py27 job passing? http://logs.openstack.org/26/103426/1/check/gate-cinder-python27/7844c61/console.html 2014-06-29 19:21:12.875 | oslo.messaging==1.4.0.0a2 That's the problem right there - 1.4.0.0a2 should not be getting installed on the stable/icehouse branch. I'm not sure why it is. Someone on #openstack-infra could probably help figure it out. Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Thanks Mark... so the problem is Oslo messaging in requirements is = in stable/icehouse [1](please note I used Oslo not OSLO). [1]: https://github.com/openstack/requirements/blob/stable/icehouse/global-requirements.txt#L49 Thanks for pointing me in the right direction. Ah, yes! This is the problem: oslo.messaging=1.3.0a9 This essentially allows *any* alpha release of oslo.messaging to be used. We should change stable/icehouse to simply be: oslo.messaging=1.3.0 I'm happy to do that tomorrow, but I suspect you'll get there first Thanks, Mark. ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Neutron] DVR demo and how-to
Hi folks, The DVR team is working really hard to complete this important task for Juno and Neutron. In order to help see this feature in action, a video has been made available and link can be found in [2]. There is still some work to do, however I wanted to remind you that all of the relevant information is available on the wiki [1, 2] and Gerrit [3]. [1] - https://wiki.openstack.org/wiki/Meetings/Neutron-L3-Subteam [2] - https://wiki.openstack.org/wiki/Neutron/DVR/HowTo [3] - https://review.openstack.org/#/q/topic:bp/neutron-ovs-dvr,n,z More to follow! Cheers, Armando ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Octavia] Temporary project leadership team
Howdy folks, Given the feedback from people on this list last week to my suggestion about holding elections for PTL and core reviewers for the Octavia project, it's clear that since we've only recently been added as a stackforge project, we don't have any code or review history on which people can base their opinions for voting on project leadership. The problem is that if we want to be able to use the gerrit system for reviewing code, specs, and documentation then somebody needs the ability to +2 proposed changes and workflow so they get merged into the source tree (ie. if we want to follow the usual OpenStack process on this, which we do.) So after discussing this with the group of people who have been working together on the design for this project for the last several months, the consensus was to elect from among ourselves an interim PTL and interim core reviewers until we have enough official history on this project to be able to actually do meaningful public elections. We are anticipating holding these elections around the time of the Juno summit. Obviously, please feel free to disagree with our process and/or decision for doing this, but I would ask that if you do, please also do us the courtesy of explaining how we're supposed to merge code using the gerrit system without having a PTL or core reviewers. Anyway, this is the list of folks we agreed upon: Interim PTL: Stephen Balukoff Interim core reviewers: Vivek Jain, Brandon Logan, German Eichberger Thanks, Stephen -- Stephen Balukoff Blue Box Group, LLC (800)613-4305 x807 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Ironic] Juno priorities and spec review timeline
Hi all! We're roughly at the midway point between summit and release, and I feel that's a good time to take a look at our progress compared to the goals we set out at the design summit. To that end, I re-opened my summit notes about what features we had prioritized in Atlanta, and engaged many the core reviewers in a discussion last friday to estimate what we'll have time to review and land in the remainder of this cycle. Based on that, I've created this spreadsheet to represent those expectations and our current progress towards what we think we can achieve this cycle: https://docs.google.com/spreadsheets/d/1Hxyfy60hN_Fit0b-plsPzK6yW3ePQC5IfwuzJwltlbo Aside from several cleanup- and test-related tasks, these goals correlate to spec reviews that have already been proposed. I've crossed off ones which we discussed at the summit, but for which no proposal has yet been submitted. The spec-review team and I will be referring to this to help us prioritize specs reviews. While I am not yet formally blocking proposals which do not fit within this list of priorities, the review team is working with a large back-log and probably won't have time to review anything else this cycle. If you're concerned that you won't be able to land your favorite feature in Juno, the best thing you can do is to participate in reviewing other people's code, join the core team, and help us accelerate the development process of K. Borrowing a little from Nova's timeline, I have proposed the following timeline for Ironic. Note that dates listed are Thursdays, and numbers in parentheses are weeks until feature freeze. You may also note that I'll be offline for two weeks immediately prior to the Juno-3 milestone, which is another reason why I'd like the core review team to have a solid plan (read: approved specs) in place by Aug 14. July 3 (-9): spec review day on Wednesday (July 2) focus on landing specs for our priorities: https://docs.google.com/spreadsheets/d/1Hxyfy60hN_Fit0b-plsPzK6yW3ePQC5IfwuzJwltlbo Jul 24 (-6): Juno-2 milestone tagged new spec proposal freeze Jul 31 (-5): midcycle meetup (July 27-30) https://wiki.openstack.org/wiki/Sprints/BeavertonJunoSprint Aug 14 (-3): last spec review day on Wednesday (Aug 13) Aug 21 (-2): PTL offline all week Aug 28 (-1): PTL offline all week Sep 4 ( 0): Juno-3 milestone tagged Feature freeze K opens for spec proposals Unmerged J spec proposals must rebase on K Merged J specs with no code proposed are deleted and may be re-proposed for K Merged J specs with code proposed need to be reviewed for feature-freeze-exception Sep 25 (+3): RC 1 build expected K spec reviews start Oct 16 (+6): Release! Oct 30 (+8): K summit spec proposal freeze K summit sessions should have corresponding spec proposal Nov 6 (+9): K design summit Thanks! Devananda ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Fuel] Triaging bugs: milestones vs release series
When you create a bug against a project (in our case, fuel) in Launchpad, it is always initially targeted at the default release series (currently, 5.1.x). On the bug summary, that isn't explicitly stated and shows as being targeted to the project in general (Fuel for OpenStack). As you add more release series to a bug, these will be listed under release series name (e.g. 5.0.x). Unfortunately, Launchpad doesn't limit the list of milestones you can target to the targeted release series, so it will happily allow you to target a bug at 4.1.x release series and set milestone in that series to 5.1. A less obvious inconsistency is when a bug is found in a stable release series like 5.0.x: it seems natural to target it to milestone like 5.0.1 and be done with it. The problem with that approach is that there's no way to reflect whether this bug is relevant for current release series (5.1.x) and if it is, to track status of the fix separately in current and stable release series. Therefore, when triaging new bugs in stable versions of Fuel or Mirantis OpenStack, please set the milestone to the next release in the current release focus (5.1.x), and target to the series it was found in separately. If there are more recent stable release series, target those as well. Example: a bug is found in 4.1.1. Set primary milestone to 5.1 (as long as current release focus is 5.1.x and 5.1 is the next milestone in that series), target 2 more release series: 4.1.x and 5.0.x, set milestones for those to 4.1.2 and 5.0.1 respectively. If there is reason to believe that the bug does not apply to some of the targeted release series, explain that in the commit and mark the bug Invalid for that release series. If the bug is present in a series but cannot be addressed there (e.g. priority is not high enough to do a backport), mark it Won't Fix for that series. If there are no objections to this approach, I'll put it in Fuel wiki. Thanks, -DmitryB -- Dmitry Borodaenko ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [Oslo][TaskFlow] Proposal for new core reviewer (pranesh pandurangan)
Greetings all stackers, I propose that we add Pranesh Pandurangan[1] to the taskflow-core team[2]. Pranesh has been actively contributing to taskflow for a while now, both in helping develop code and helping with the review load. He has provided quality reviews and is doing an awesome job with the various taskflow concepts and helping make taskflow the best library it can be! Overall I think he would make a great addition to the core review team. Please respond with +1/-1. Thanks much! -- Joshua Harlow It's openstack, relax... | harlo...@yahoo-inc.com [1] https://launchpad.net/~praneshp [2] https://wiki.openstack.org/wiki/TaskFlow/CoreTeam ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [ceilometer] How to filter out meters whose resources have been deleted?
anyone knows this? 2014-6-30 PM6:35于 Ke Xia xiake9...@gmail.com写道: Hi, As time goes on, meters will be a huge list, and some meters whose resources have been deleted may be useless for me, can I filter them out from meter list? Thanks ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
Re: [openstack-dev] [third-party-ci][neutron] What is Success exactly?
On 06/30/2014 07:08 PM, Anita Kuno wrote: On 06/30/2014 04:22 PM, Jay Pipes wrote: Hi Stackers, Some recent ML threads [1] and a hot IRC meeting today [2] brought up some legitimate questions around how a newly-proposed Stackalytics report page for Neutron External CI systems [2] represented the results of an external CI system as successful or not. First, I want to say that Ilya and all those involved in the Stackalytics program simply want to provide the most accurate information to developers in a format that is easily consumed. While there need to be some changes in how data is shown (and the wording of things like Tests Succeeded), I hope that the community knows there isn't any ill intent on the part of Mirantis or anyone who works on Stackalytics. OK, so let's keep the conversation civil -- we're all working towards the same goals of transparency and accuracy. :) Alright, now, Anita and Kurt Taylor were asking a very poignant question: But what does CI tested really mean? just running tests? or tested to pass some level of requirements? In this nascent world of external CI systems, we have a set of issues that we need to resolve: 1) All of the CI systems are different. Some run Bash scripts. Some run Jenkins slaves and devstack-gate scripts. Others run custom Python code that spawns VMs and publishes logs to some public domain. As a community, we need to decide whether it is worth putting in the effort to create a single, unified, installable and runnable CI system, so that we can legitimately say all of the external systems are identical, with the exception of the driver code for vendor X being substituted in the Neutron codebase. If the goal of the external CI systems is to produce reliable, consistent results, I feel the answer to the above is yes, but I'm interested to hear what others think. Frankly, in the world of benchmarks, it would be unthinkable to say go ahead and everyone run your own benchmark suite, because you would get wildly different results. A similar problem has emerged here. 2) There is no mediation or verification that the external CI system is actually testing anything at all As a community, we need to decide whether the current system of self-policing should continue. If it should, then language on reports like [3] should be very clear that any numbers derived from such systems should be taken with a grain of salt. Use of the word Success should be avoided, as it has connotations (in English, at least) that the result has been verified, which is simply not the case as long as no verification or mediation occurs for any external CI system. 3) There is no clear indication of what tests are being run, and therefore there is no clear indication of what success is I think we can all agree that a test has three possible outcomes: pass, fail, and skip. The results of a test suite run therefore is nothing more than the aggregation of which tests passed, which failed, and which were skipped. As a community, we must document, for each project, what are expected set of tests that must be run for each merged patch into the project's source tree. This documentation should be discoverable so that reports like [3] can be crystal-clear on what the data shown actually means. The report is simply displaying the data it receives from Gerrit. The community needs to be proactive in saying this is what is expected to be tested. This alone would allow the report to give information such as External CI system ABC performed the expected tests. X tests passed. Y tests failed. Z tests were skipped. Likewise, it would also make it possible for the report to give information such as External CI system DEF did not perform the expected tests., which is excellent information in and of itself. === In thinking about the likely answers to the above questions, I believe it would be prudent to change the Stackalytics report in question [3] in the following ways: a. Change the Success % column header to % Reported +1 Votes b. Change the phrase Green cell - tests ran successfully, red cell - tests failed to Green cell - System voted +1, red cell - System voted -1 and then, when we have more and better data (for example, # tests passed, failed, skipped, etc), we can provide more detailed information than just reported +1 or not. Thoughts? Best, -jay [1] http://lists.openstack.org/pipermail/openstack-dev/2014-June/038933.html [2] http://eavesdrop.openstack.org/meetings/third_party/2014/third_party.2014-06-30-18.01.log.html [3] http://stackalytics.com/report/ci/neutron/7 ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev Hi Jay: Thanks for starting this thread. You raise some interesting questions. The question I had identified as needing definition is what algorithm do we use to assess fitness of a third party ci system.
Re: [openstack-dev] [Oslo][TaskFlow] Proposal for new core reviewer (pranesh pandurangan)
+1 Thanks Changbin On Mon, Jun 30, 2014 at 8:05 PM, Joshua Harlow harlo...@yahoo-inc.com wrote: Greetings all stackers, I propose that we add Pranesh Pandurangan[1] to the taskflow-core team[2]. Pranesh has been actively contributing to taskflow for a while now, both in helping develop code and helping with the review load. He has provided quality reviews and is doing an awesome job with the various taskflow concepts and helping make taskflow the best library it can be! Overall I think he would make a great addition to the core review team. Please respond with +1/-1. Thanks much! -- Joshua Harlow It's openstack, relax... | harlo...@yahoo-inc.com [1] https://launchpad.net/~praneshp [2] https://wiki.openstack.org/wiki/TaskFlow/CoreTeam ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev
[openstack-dev] [nova][sriov] weekly meeting for july 1st and july 8th
Hi, I will be on PTO from Tuesday, and come back to office on July 9th Wednesday. Therefore, I won’t be present in the next two SR-IOV weekly meetings. Regarding the sr-iov development status, I finally fixed all the failures in the existing unit tests. Rob and I are still working on adding new unit test cases in the PCI and libvirt driver area. Once that’s done, we should be able to push another two patches up. Thanks, Robert ___ OpenStack-dev mailing list OpenStack-dev@lists.openstack.org http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev