Re: Coverity submission broken?

2015-12-09 Thread Nadav Goldin
Hi, Contacted them. will update.

Nadav.


On Wed, Dec 9, 2015 at 10:35 AM, Allon Mureinik  wrote:

> Can wait till next week, thanks.
>
> On Wed, Dec 9, 2015 at 10:20 AM, Eyal Edri  wrote:
>
>> adding nadav who has been with contact with them,
>> they might have blocked us again.
>>
>> Can this wait till next week?
>> if not, sagi can look at it.
>>
>> e.
>>
>> On Tue, Dec 8, 2015 at 7:13 PM, Allon Mureinik 
>> wrote:
>>
>>> Hi Infra,
>>>
>>> Coverity doesn't seem to be updating. The last couple of jenkins jobs
>>> have failed (e.g., [1]):
>>>
>>> + curl --form project=ovirt-engine --form token=YqBYeyDp2jPuC_uVz0Hdog 
>>> --form email=ih...@redhat.com --form file=@ovirt-engine-cov.tgz 
>>> http://scan5.coverity.com/cgi-bin/upload.py
>>>   % Total% Received % Xferd  Average Speed   TimeTime Time  
>>> Current
>>>  Dload  Upload   Total   SpentLeft  
>>> Speed
>>>
>>>   0 00 00 0  0  0 --:--:-- --:--:-- --:--:--
>>>  0curl: (7) Failed connect to scan5.coverity.com:80; Connection refused
>>> Build step 'Execute shell' marked build as failure
>>> Finished: FAILURE
>>>
>>>
>>>
>>> Can someone take a look at this please?
>>>
>>> [If you need my help, of course, don't hesitate to ask]
>>>
>>>
>>>
>>> Thanks,
>>>
>>> Allon
>>>
>>>
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
>>>
>>
>>
>> --
>> Eyal Edri
>> Supervisor, RHEV CI
>> EMEA ENG Virtualization R&D
>> Red Hat Israel
>>
>> phone: +972-9-7692018
>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


ng another test for email tickets

2015-12-09 Thread Nadav Goldin

___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Coverity submission broken?

2015-12-14 Thread Nadav Goldin
update: been ping-ponging with their support for a few days,
scan-ad...@coverity.com(which is in charge of the scan service) isn't
replying and supp...@coverity.com says they are not responsible. will keep
updating.




On Wed, Dec 9, 2015 at 12:51 PM, Nadav Goldin  wrote:

> Hi, Contacted them. will update.
>
> Nadav.
>
>
> On Wed, Dec 9, 2015 at 10:35 AM, Allon Mureinik 
> wrote:
>
>> Can wait till next week, thanks.
>>
>> On Wed, Dec 9, 2015 at 10:20 AM, Eyal Edri  wrote:
>>
>>> adding nadav who has been with contact with them,
>>> they might have blocked us again.
>>>
>>> Can this wait till next week?
>>> if not, sagi can look at it.
>>>
>>> e.
>>>
>>> On Tue, Dec 8, 2015 at 7:13 PM, Allon Mureinik 
>>> wrote:
>>>
>>>> Hi Infra,
>>>>
>>>> Coverity doesn't seem to be updating. The last couple of jenkins jobs
>>>> have failed (e.g., [1]):
>>>>
>>>> + curl --form project=ovirt-engine --form token=YqBYeyDp2jPuC_uVz0Hdog 
>>>> --form email=ih...@redhat.com --form file=@ovirt-engine-cov.tgz 
>>>> http://scan5.coverity.com/cgi-bin/upload.py
>>>>   % Total% Received % Xferd  Average Speed   TimeTime Time  
>>>> Current
>>>>  Dload  Upload   Total   SpentLeft  
>>>> Speed
>>>>
>>>>   0 00 00 0  0  0 --:--:-- --:--:-- --:--:--   
>>>>   0curl: (7) Failed connect to scan5.coverity.com:80; Connection refused
>>>> Build step 'Execute shell' marked build as failure
>>>> Finished: FAILURE
>>>>
>>>>
>>>>
>>>> Can someone take a look at this please?
>>>>
>>>> [If you need my help, of course, don't hesitate to ask]
>>>>
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Allon
>>>>
>>>>
>>>> ___
>>>> Infra mailing list
>>>> Infra@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>>
>>>>
>>>
>>>
>>> --
>>> Eyal Edri
>>> Supervisor, RHEV CI
>>> EMEA ENG Virtualization R&D
>>> Red Hat Israel
>>>
>>> phone: +972-9-7692018
>>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>>
>>
>>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Request to mirror the open source project oVirt.org

2015-12-28 Thread Nadav Goldin
Hi Lior,
I've added your key to our mirror user, so you should be able to perform
rsync as described in http://www.ovirt.org/Repository_mirrors.
Please check it out and tell us everything is working. if it's all good
send us the http/ftp links and we will update our mirror and wiki pages.

Thanks for contributing to oVirt,

Nadav.




On Mon, Dec 28, 2015 at 10:17 AM, Eyal Edri  wrote:

>
> -- Forwarded message --
> From: Lior Kaplan 
> Date: Thu, Dec 10, 2015 at 7:18 PM
> Subject: Re: Request to mirror the open source project oVirt.org
> To: Anton Marchukov 
> Cc: mirrormaster-isoc Israel , "Fediuck, Doron"
> , Eyal Edri 
>
>
> Hi,
>
> I'm sorry for the long delay. The SSH public key is:
>
> ssh-rsa
> B3NzaC1yc2EDAQABAAABAQDO1w2Hq1Kls2NhNfmv09SwD6bYbpaEaywoqqafQ/37g9K6iT0jtuS3BhnSSV8p5oWC+7VWmf64jD0+n2bKsMaBY2XK77RbWwsErmHqAEzaY4/coMPTmoIMV67Y34tVgzAgbqV6XUQpnrsFbla6N8DNYoE6LPHprF3/rP66UJOuyYC/gd3Zyx5r7TQQLoReC0NYjCmq0bjc6hLfV9cyslkFGJXZGTIZrCFenuAsa/XXdN7NKGoSwh+47sSEKhR5qR7mtgXh4e7Pxo1OoQQ/gF9QzJT+8esxGntTRBP5by5amPt2w7WFNUE2daLkUUWAtrfOMPXwEwsM7WHdNpaFbRmp
>
> Kaplan
>
> On Mon, Nov 2, 2015 at 4:33 PM, Anton Marchukov 
> wrote:
>
>> Hello Lior.
>>
>> Please check the info on the following page:
>>
>> http://www.ovirt.org/Repository_mirrors
>>
>> Basically you need to send us your ssh public key and configure rsync
>> command on your side as specified. Mirroring is done by rsync over ssh with
>> key authentication.
>>
>> Anton.
>>
>> On Mon, Nov 2, 2015 at 8:27 AM, Eyal Edri  wrote:
>>
>>> Hi Lior,
>>>
>>> Sorry if i was misunderstood, when i mentioned the developers it was
>>> because the terms of conditions on the site, for mirroring,
>>> being an israeli project and all, the mirrors of course are used by
>>> users, developer seldom uses rpms, most of the time they run local builds
>>> and use "developer mode", so i expect that most of the usage will be by
>>> developers.
>>>
>>> I'm adding Anton, one of the engineers from the oVirt infra team who
>>> handles mirrors among other tasks in the team.
>>> Anton - can you please provide lior with the info needed to mirror ovirt
>>> to their site?
>>>
>>> thanks,
>>>
>>> eyal.
>>>
>>> On Sun, Nov 1, 2015 at 7:38 PM, Lior Kaplan 
>>> wrote:
>>>
 Hi Eyal,

 The mirrors are mostly intended for users, less for developers. But we
 can add a mirror.

 Can you arrange the ssh key your mirror pages mentions?

 Kaplan

 On Sun, Nov 1, 2015 at 7:33 PM, Eyal Edri  wrote:

> Hi,
>
> I noticed that you're offering mirror for israeli open source
> projects, which is great!
> I'd like to ask if it's possible to mirror the oVirt [1] project as
> well.
> oVirt is an open source virtualization product for enterprise vm
> management, similar to vSphere
> (but much better :)
>
> Most of the development of the project is done in Israel, by israeli
> engineers, but there are also other engineers around the world,
> contributing to the project.
>
> Our project's rpms reside at http://resources.ovirt.org/pub/
> and takes approx 20-30G of space to host all existing rpms for the
> latest 3 versions.
> You can see existing sites which are already mirroring oVirt here [2]
>
> Would it be possible to mirror it to http://mirror.isoc.org.il/ as
> well?
>
> Please let me know if you need more info on the project!
> Looking forward to hear from you
>
>
> [1] www.ovirt.org
> [2] http://www.ovirt.org/Repository_mirrors
>
>
> --
> Eyal Edri
> Supervisor, RHEV CI
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>


>>>
>>>
>>> --
>>> Eyal Edri
>>> Supervisor, RHEV CI
>>> EMEA ENG Virtualization R&D
>>> Red Hat Israel
>>>
>>> phone: +972-9-7692018
>>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>>
>>
>>
>>
>> --
>> Anton Marchukov
>> Senior Software Engineer - RHEV CI - Red Hat
>>
>>
>
>
>
> --
> Eyal Edri
> Associate Manager
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


oVirt mirror @ rit.edu

2016-01-03 Thread Nadav Goldin
Hey,
Nadav from oVirt infra team here.
I see the mirror hasn't been synced for a while, can you check it?

Thanks
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Fwd: Request to mirror the open source project oVirt.org

2016-01-10 Thread Nadav Goldin
Hey Lior,
can you please try again without the z flag:
rsync -rltHvvP mir...@resources.ovirt.org:/var/www/html mirror/ovirt

Nadav.


On Wed, Jan 6, 2016 at 3:55 PM, Sagi Shnaidman  wrote:

> Adding rhev-ci,
> please help him
>
>
> On 01/06/2016 03:53 PM, Lior Kaplan wrote:
>
> $ rsync -e /usr/bin/ssh -rltHvvzP
> mir...@resources.ovirt.org:/var/www/html/pub mirror/ovirt/
> opening connection using: /usr/bin/ssh -l mirror resources.ovirt.org
> rsync --server --sender -vvlHtrze.iLsf . /var/www/html/pub
> receiving file list ...
> [Receiver] expand file_list pointer array to 262144 bytes, did move
> 64162 files to consider
> delta-transmission enabled
> ovirt-node-base-stable is uptodate
> ovirt-3.3/rpm/el6Server is uptodate
> ovirt-3.4/rpm/el6Server is uptodate
> ovirt-3.4/rpm/el7Server is uptodate
> ovirt-3.5-snapshot-static/rpm/el6.6 is uptodate
> ovirt-3.5-snapshot-static/rpm/el6Server is uptodate
> ovirt-3.5-snapshot-static/rpm/el6Workstation is uptodate
> ovirt-3.5-snapshot-static/rpm/el7Server is uptodate
> ovirt-3.5-snapshot-static/rpm/el7Workstation is uptodate
> ovirt-3.5-snapshot/rpm/el6.6 is uptodate
> ovirt-3.5-snapshot/rpm/el6Server is uptodate
> ovirt-3.5-snapshot/rpm/el6Workstation is uptodate
> ovirt-3.5-snapshot/rpm/el7Server is uptodate
> ovirt-3.5-snapshot/rpm/el7Workstation is uptodate
> ovirt-3.5/rpm/el6.6 is uptodate
> ovirt-3.5/rpm/el6Server is uptodate
> ovirt-3.5/rpm/el6Workstation is uptodate
> ovirt-3.5/rpm/el7Server is uptodate
> ovirt-3.5/rpm/el7Workstation is uptodate
> ./
> ovirt-3.6-pre/rpm/el6.7 is uptodate
> ovirt-3.6-pre/rpm/el6Server is uptodate
> ovirt-3.6-pre/rpm/el6Workstation is uptodate
> ovirt-3.6-pre/rpm/el7Server is uptodate
> ovirt-3.6-pre/rpm/el7Workstation is uptodate
> keys/
> keys/RPM-GPG-ovirt
>0   0%0.00kB/s0:00:00
> ovirt-3.6-snapshot-static/rpm/el6.6 is uptodate
> ovirt-3.6-snapshot-static/rpm/el6Server is uptodate
> ovirt-3.6-snapshot-static/rpm/el6Workstation is uptodate
> ovirt-3.6-snapshot-static/rpm/el7Server is uptodate
> ovirt-3.6-snapshot-static/rpm/el7Workstation is uptodate
>
> inflate returned -3 (0 bytes)
> rsync error: error in rsync protocol data stream (code 12) at token.c(548)
> [receiver=3.0.9]
> rsync: connection unexpectedly closed (3651558 bytes received so far)
> [generator]
> rsync error: error in rsync protocol data stream (code 12) at io.c(605)
> [generator=3.0.9]
>
>
> On Wed, Jan 6, 2016 at 3:44 PM, Sagi Shnaidman 
> wrote:
>
>> Hi,
>>
>> try please adding  "-e /usr/bin/ssh" to rsync options.
>> rsync -e /usr/bin/ssh -rltHvvzP  ...
>>
>> "/usr/bin/ssh" should be your SSH path.
>>
>> tell me please if it succeeds.
>>
>> thanks
>>
>> On 01/06/2016 11:33 AM, Lior Kaplan wrote:
>>
>> *rsync error: error in rsync protocol data stream*
>>
>>
>>
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Fwd: Request to mirror the open source project oVirt.org

2016-01-13 Thread Nadav Goldin
Great :)
I'll track it for a few days to see its all good and update our wiki
pages/mirror list afterwards.

thanks,

Nadav.

On Tue, Jan 12, 2016 at 11:04 AM, Lior Kaplan  wrote:

> Bingo (:
>
> http://mirror.isoc.org.il/pub/ovirt/
>
>
> I'll set a daily update.
>
> Kaplan
>
> On Mon, Jan 11, 2016 at 4:13 PM, Nadav Goldin  wrote:
>
>> Hey Lior, sorry sent this to the wrong address:
>> can you please try again without the z flag:
>> rsync -rltHvvP mir...@resources.ovirt.org:/var/www/html mirror/ovirt
>>
>>
>> On Mon, Jan 11, 2016 at 4:09 PM, Lior Kaplan 
>> wrote:
>>
>>> Hi Guys,
>>>
>>> Any news?
>>>
>>> On Wed, Jan 6, 2016 at 3:53 PM, Lior Kaplan 
>>> wrote:
>>>
>>>> $ rsync -e /usr/bin/ssh -rltHvvzP   
>>>> mir...@resources.ovirt.org:/var/www/html/pub
>>>> mirror/ovirt/
>>>> opening connection using: /usr/bin/ssh -l mirror resources.ovirt.org
>>>> rsync --server --sender -vvlHtrze.iLsf . /var/www/html/pub
>>>> receiving file list ...
>>>> [Receiver] expand file_list pointer array to 262144 bytes, did move
>>>> 64162 files to consider
>>>> delta-transmission enabled
>>>> ovirt-node-base-stable is uptodate
>>>> ovirt-3.3/rpm/el6Server is uptodate
>>>> ovirt-3.4/rpm/el6Server is uptodate
>>>> ovirt-3.4/rpm/el7Server is uptodate
>>>> ovirt-3.5-snapshot-static/rpm/el6.6 is uptodate
>>>> ovirt-3.5-snapshot-static/rpm/el6Server is uptodate
>>>> ovirt-3.5-snapshot-static/rpm/el6Workstation is uptodate
>>>> ovirt-3.5-snapshot-static/rpm/el7Server is uptodate
>>>> ovirt-3.5-snapshot-static/rpm/el7Workstation is uptodate
>>>> ovirt-3.5-snapshot/rpm/el6.6 is uptodate
>>>> ovirt-3.5-snapshot/rpm/el6Server is uptodate
>>>> ovirt-3.5-snapshot/rpm/el6Workstation is uptodate
>>>> ovirt-3.5-snapshot/rpm/el7Server is uptodate
>>>> ovirt-3.5-snapshot/rpm/el7Workstation is uptodate
>>>> ovirt-3.5/rpm/el6.6 is uptodate
>>>> ovirt-3.5/rpm/el6Server is uptodate
>>>> ovirt-3.5/rpm/el6Workstation is uptodate
>>>> ovirt-3.5/rpm/el7Server is uptodate
>>>> ovirt-3.5/rpm/el7Workstation is uptodate
>>>> ./
>>>> ovirt-3.6-pre/rpm/el6.7 is uptodate
>>>> ovirt-3.6-pre/rpm/el6Server is uptodate
>>>> ovirt-3.6-pre/rpm/el6Workstation is uptodate
>>>> ovirt-3.6-pre/rpm/el7Server is uptodate
>>>> ovirt-3.6-pre/rpm/el7Workstation is uptodate
>>>> keys/
>>>> keys/RPM-GPG-ovirt
>>>>0   0%0.00kB/s0:00:00
>>>> ovirt-3.6-snapshot-static/rpm/el6.6 is uptodate
>>>> ovirt-3.6-snapshot-static/rpm/el6Server is uptodate
>>>> ovirt-3.6-snapshot-static/rpm/el6Workstation is uptodate
>>>> ovirt-3.6-snapshot-static/rpm/el7Server is uptodate
>>>> ovirt-3.6-snapshot-static/rpm/el7Workstation is uptodate
>>>>
>>>> inflate returned -3 (0 bytes)
>>>> rsync error: error in rsync protocol data stream (code 12) at
>>>> token.c(548) [receiver=3.0.9]
>>>> rsync: connection unexpectedly closed (3651558 bytes received so far)
>>>> [generator]
>>>> rsync error: error in rsync protocol data stream (code 12) at io.c(605)
>>>> [generator=3.0.9]
>>>>
>>>>
>>>> On Wed, Jan 6, 2016 at 3:44 PM, Sagi Shnaidman 
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> try please adding  "-e /usr/bin/ssh" to rsync options.
>>>>> rsync -e /usr/bin/ssh -rltHvvzP  ...
>>>>>
>>>>> "/usr/bin/ssh" should be your SSH path.
>>>>>
>>>>> tell me please if it succeeds.
>>>>>
>>>>> thanks
>>>>>
>>>>> On 01/06/2016 11:33 AM, Lior Kaplan wrote:
>>>>>
>>>>> *rsync error: error in rsync protocol data stream*
>>>>>
>>>>>
>>>>>
>>>>
>>>
>>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Fwd: Request to mirror the open source project oVirt.org

2016-01-18 Thread Nadav Goldin
Hi Lior,
we usually add an email/compay name in the mirror's wiki page, you can have
a look at: http://www.ovirt.org/Repository_mirrors
which details do you prefer I add?


Nadav.



On Wed, Jan 13, 2016 at 2:13 PM, Nadav Goldin  wrote:

> Great :)
> I'll track it for a few days to see its all good and update our wiki
> pages/mirror list afterwards.
>
> thanks,
>
> Nadav.
>
> On Tue, Jan 12, 2016 at 11:04 AM, Lior Kaplan 
> wrote:
>
>> Bingo (:
>>
>> http://mirror.isoc.org.il/pub/ovirt/
>>
>>
>> I'll set a daily update.
>>
>> Kaplan
>>
>> On Mon, Jan 11, 2016 at 4:13 PM, Nadav Goldin  wrote:
>>
>>> Hey Lior, sorry sent this to the wrong address:
>>> can you please try again without the z flag:
>>> rsync -rltHvvP mir...@resources.ovirt.org:/var/www/html mirror/ovirt
>>>
>>>
>>> On Mon, Jan 11, 2016 at 4:09 PM, Lior Kaplan 
>>> wrote:
>>>
>>>> Hi Guys,
>>>>
>>>> Any news?
>>>>
>>>> On Wed, Jan 6, 2016 at 3:53 PM, Lior Kaplan 
>>>> wrote:
>>>>
>>>>> $ rsync -e /usr/bin/ssh -rltHvvzP   
>>>>> mir...@resources.ovirt.org:/var/www/html/pub
>>>>> mirror/ovirt/
>>>>> opening connection using: /usr/bin/ssh -l mirror resources.ovirt.org
>>>>> rsync --server --sender -vvlHtrze.iLsf . /var/www/html/pub
>>>>> receiving file list ...
>>>>> [Receiver] expand file_list pointer array to 262144 bytes, did move
>>>>> 64162 files to consider
>>>>> delta-transmission enabled
>>>>> ovirt-node-base-stable is uptodate
>>>>> ovirt-3.3/rpm/el6Server is uptodate
>>>>> ovirt-3.4/rpm/el6Server is uptodate
>>>>> ovirt-3.4/rpm/el7Server is uptodate
>>>>> ovirt-3.5-snapshot-static/rpm/el6.6 is uptodate
>>>>> ovirt-3.5-snapshot-static/rpm/el6Server is uptodate
>>>>> ovirt-3.5-snapshot-static/rpm/el6Workstation is uptodate
>>>>> ovirt-3.5-snapshot-static/rpm/el7Server is uptodate
>>>>> ovirt-3.5-snapshot-static/rpm/el7Workstation is uptodate
>>>>> ovirt-3.5-snapshot/rpm/el6.6 is uptodate
>>>>> ovirt-3.5-snapshot/rpm/el6Server is uptodate
>>>>> ovirt-3.5-snapshot/rpm/el6Workstation is uptodate
>>>>> ovirt-3.5-snapshot/rpm/el7Server is uptodate
>>>>> ovirt-3.5-snapshot/rpm/el7Workstation is uptodate
>>>>> ovirt-3.5/rpm/el6.6 is uptodate
>>>>> ovirt-3.5/rpm/el6Server is uptodate
>>>>> ovirt-3.5/rpm/el6Workstation is uptodate
>>>>> ovirt-3.5/rpm/el7Server is uptodate
>>>>> ovirt-3.5/rpm/el7Workstation is uptodate
>>>>> ./
>>>>> ovirt-3.6-pre/rpm/el6.7 is uptodate
>>>>> ovirt-3.6-pre/rpm/el6Server is uptodate
>>>>> ovirt-3.6-pre/rpm/el6Workstation is uptodate
>>>>> ovirt-3.6-pre/rpm/el7Server is uptodate
>>>>> ovirt-3.6-pre/rpm/el7Workstation is uptodate
>>>>> keys/
>>>>> keys/RPM-GPG-ovirt
>>>>>0   0%0.00kB/s0:00:00
>>>>> ovirt-3.6-snapshot-static/rpm/el6.6 is uptodate
>>>>> ovirt-3.6-snapshot-static/rpm/el6Server is uptodate
>>>>> ovirt-3.6-snapshot-static/rpm/el6Workstation is uptodate
>>>>> ovirt-3.6-snapshot-static/rpm/el7Server is uptodate
>>>>> ovirt-3.6-snapshot-static/rpm/el7Workstation is uptodate
>>>>>
>>>>> inflate returned -3 (0 bytes)
>>>>> rsync error: error in rsync protocol data stream (code 12) at
>>>>> token.c(548) [receiver=3.0.9]
>>>>> rsync: connection unexpectedly closed (3651558 bytes received so far)
>>>>> [generator]
>>>>> rsync error: error in rsync protocol data stream (code 12) at
>>>>> io.c(605) [generator=3.0.9]
>>>>>
>>>>>
>>>>> On Wed, Jan 6, 2016 at 3:44 PM, Sagi Shnaidman 
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> try please adding  "-e /usr/bin/ssh" to rsync options.
>>>>>> rsync -e /usr/bin/ssh -rltHvvzP  ...
>>>>>>
>>>>>> "/usr/bin/ssh" should be your SSH path.
>>>>>>
>>>>>> tell me please if it succeeds.
>>>>>>
>>>>>> thanks
>>>>>>
>>>>>> On 01/06/2016 11:33 AM, Lior Kaplan wrote:
>>>>>>
>>>>>> *rsync error: error in rsync protocol data stream*
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Updating http://www.ovirt.org/Download

2016-01-18 Thread Nadav Goldin
Hi Brian,
Can you please update http://www.ovirt.org/Download with the new
http://mirror.isoc.org.il/pub/ovirt/ mirror?

I've added its details in
http://www.ovirt.org/Repository_mirrors#Current_mirrors


Thanks

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Fwd: Request to mirror the open source project oVirt.org

2016-01-18 Thread Nadav Goldin
great - I've updated http://www.ovirt.org/Repository_mirrors and requested
an update on http://www.ovirt.org/Download.
once this <https://gerrit.ovirt.org/#/c/52384> is merged(probably next
couple of days) you should start getting requests.

Thanks,
Nadav.


On Mon, Jan 18, 2016 at 10:04 PM, Lior Kaplan  wrote:

> ISOC-IL
> mirrormas...@isoc.org.il
>
> Kaplan
>
> On Mon, Jan 18, 2016 at 8:02 PM, Nadav Goldin  wrote:
>
>> Hi Lior,
>> we usually add an email/compay name in the mirror's wiki page, you can
>> have a look at: http://www.ovirt.org/Repository_mirrors
>> which details do you prefer I add?
>>
>>
>> Nadav.
>>
>>
>>
>> On Wed, Jan 13, 2016 at 2:13 PM, Nadav Goldin  wrote:
>>
>>> Great :)
>>> I'll track it for a few days to see its all good and update our wiki
>>> pages/mirror list afterwards.
>>>
>>> thanks,
>>>
>>> Nadav.
>>>
>>> On Tue, Jan 12, 2016 at 11:04 AM, Lior Kaplan 
>>> wrote:
>>>
>>>> Bingo (:
>>>>
>>>> http://mirror.isoc.org.il/pub/ovirt/
>>>>
>>>>
>>>> I'll set a daily update.
>>>>
>>>> Kaplan
>>>>
>>>> On Mon, Jan 11, 2016 at 4:13 PM, Nadav Goldin 
>>>> wrote:
>>>>
>>>>> Hey Lior, sorry sent this to the wrong address:
>>>>> can you please try again without the z flag:
>>>>> rsync -rltHvvP mir...@resources.ovirt.org:/var/www/html mirror/ovirt
>>>>>
>>>>>
>>>>> On Mon, Jan 11, 2016 at 4:09 PM, Lior Kaplan 
>>>>> wrote:
>>>>>
>>>>>> Hi Guys,
>>>>>>
>>>>>> Any news?
>>>>>>
>>>>>> On Wed, Jan 6, 2016 at 3:53 PM, Lior Kaplan 
>>>>>> wrote:
>>>>>>
>>>>>>> $ rsync -e /usr/bin/ssh -rltHvvzP   
>>>>>>> mir...@resources.ovirt.org:/var/www/html/pub
>>>>>>> mirror/ovirt/
>>>>>>> opening connection using: /usr/bin/ssh -l mirror resources.ovirt.org
>>>>>>> rsync --server --sender -vvlHtrze.iLsf . /var/www/html/pub
>>>>>>> receiving file list ...
>>>>>>> [Receiver] expand file_list pointer array to 262144 bytes, did move
>>>>>>> 64162 files to consider
>>>>>>> delta-transmission enabled
>>>>>>> ovirt-node-base-stable is uptodate
>>>>>>> ovirt-3.3/rpm/el6Server is uptodate
>>>>>>> ovirt-3.4/rpm/el6Server is uptodate
>>>>>>> ovirt-3.4/rpm/el7Server is uptodate
>>>>>>> ovirt-3.5-snapshot-static/rpm/el6.6 is uptodate
>>>>>>> ovirt-3.5-snapshot-static/rpm/el6Server is uptodate
>>>>>>> ovirt-3.5-snapshot-static/rpm/el6Workstation is uptodate
>>>>>>> ovirt-3.5-snapshot-static/rpm/el7Server is uptodate
>>>>>>> ovirt-3.5-snapshot-static/rpm/el7Workstation is uptodate
>>>>>>> ovirt-3.5-snapshot/rpm/el6.6 is uptodate
>>>>>>> ovirt-3.5-snapshot/rpm/el6Server is uptodate
>>>>>>> ovirt-3.5-snapshot/rpm/el6Workstation is uptodate
>>>>>>> ovirt-3.5-snapshot/rpm/el7Server is uptodate
>>>>>>> ovirt-3.5-snapshot/rpm/el7Workstation is uptodate
>>>>>>> ovirt-3.5/rpm/el6.6 is uptodate
>>>>>>> ovirt-3.5/rpm/el6Server is uptodate
>>>>>>> ovirt-3.5/rpm/el6Workstation is uptodate
>>>>>>> ovirt-3.5/rpm/el7Server is uptodate
>>>>>>> ovirt-3.5/rpm/el7Workstation is uptodate
>>>>>>> ./
>>>>>>> ovirt-3.6-pre/rpm/el6.7 is uptodate
>>>>>>> ovirt-3.6-pre/rpm/el6Server is uptodate
>>>>>>> ovirt-3.6-pre/rpm/el6Workstation is uptodate
>>>>>>> ovirt-3.6-pre/rpm/el7Server is uptodate
>>>>>>> ovirt-3.6-pre/rpm/el7Workstation is uptodate
>>>>>>> keys/
>>>>>>> keys/RPM-GPG-ovirt
>>>>>>>0   0%0.00kB/s0:00:00
>>>>>>> ovirt-3.6-snapshot-static/rpm/el6.6 is uptodate
>>>>>>> ovirt-3.6-snapshot-static/rpm/el6Server is uptodate
>>>>>>> ovirt-3.6-snapshot-static/rpm/el6Workstation is uptodate
>>>>>>> ovirt-3.6-snapshot-static/rpm/el7Server is uptodate
>>>>>>> ovirt-3.6-snapshot-static/rpm/el7Workstation is uptodate
>>>>>>>
>>>>>>> inflate returned -3 (0 bytes)
>>>>>>> rsync error: error in rsync protocol data stream (code 12) at
>>>>>>> token.c(548) [receiver=3.0.9]
>>>>>>> rsync: connection unexpectedly closed (3651558 bytes received so
>>>>>>> far) [generator]
>>>>>>> rsync error: error in rsync protocol data stream (code 12) at
>>>>>>> io.c(605) [generator=3.0.9]
>>>>>>>
>>>>>>>
>>>>>>> On Wed, Jan 6, 2016 at 3:44 PM, Sagi Shnaidman 
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> try please adding  "-e /usr/bin/ssh" to rsync options.
>>>>>>>> rsync -e /usr/bin/ssh -rltHvvzP  ...
>>>>>>>>
>>>>>>>> "/usr/bin/ssh" should be your SSH path.
>>>>>>>>
>>>>>>>> tell me please if it succeeds.
>>>>>>>>
>>>>>>>> thanks
>>>>>>>>
>>>>>>>> On 01/06/2016 11:33 AM, Lior Kaplan wrote:
>>>>>>>>
>>>>>>>> *rsync error: error in rsync protocol data stream*
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


new Jenkins VM in PHX lab

2016-02-09 Thread Nadav Goldin
Hi all,
I installed yesterday a fresh Jenkins VM in the PHX datacenter at
jenkins.phx.ovirt.org running centos 7.2, few details:
1. It is puppetized, but still needs some verification[1], I was able to
auto-generate the plugins list from jenkins.ovirt.org into the puppet
manifest, so it has exactly the same plugins installed. the concept might
also work for future migrations we do between the instances.
2. the jenkins data directory is 500gb configured with lvm and xfs. I was
able to do "live" storage incrase by adding a new volume to the LVM group
in the following procedure:

   -  created new virtio volume in the VM from phx-engine
   -  fdisk /dev/vdd then: n -> p -> 1 -> enter -> enter -> t -> 8e -> w
   -  vgextend jenkins_lvm /dev/vdd1
   -  lvextend /dev/mapper/jenkins_lvm-data -L28G
   -  xfs_growfs /dev/mapper/jenkins_lvm-data

I was unable to increase the volume size from the engine, and then increase
the partition size(only create a new partition on the same volume with the
new increased space in the volume), not sure if that is possible.

3. authentication - same as jenkins.ovirt.org for now: self enrol and
adding permissions(ping me)

4. there is much more configuration that needs to be done to make it
functional, but i guess we can start by testing it :)
[1] https://gerrit.ovirt.org/#/c/53309/
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: new Jenkins VM in PHX lab

2016-02-10 Thread Nadav Goldin
>
> You forgot the 1st thing we told you - "don't create partitions on the
> disk"...
> Instead the whole disk '/dev/vdX' should be formatted as a PV, then you can
> grow it from the engine and then 'pvresize' followed by 'lvresize'.


I didn't forget, I came across [1] which quotes [2] and [3] saying
best-practice is to create a partition on the PV:

> *Not Recommended*
>
> Using the whole disk as a PV (as opposed to a partition spanning the whole
> disk) is not recommended because of the management issues it can create.
> Any other OS that looks at the disk will not recognize the LVM metadata and
> display the disk as being free, so it is likely it will be overwritten. LVM
> itself will work fine with whole disk PVs.
>
although I am no expert in lvm so if we agree its ok, no problem, I'll
change it.

[1]
http://unix.stackexchange.com/questions/76588/what-is-the-best-practice-for-adding-disks-in-lvm
[2] http://tldp.org/HOWTO/LVM-HOWTO/initdisks.html
[3]
https://access.redhat.com/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Logical_Volume_Manager_Administration/LVM_components.html#multiple_partitions


On Wed, Feb 10, 2016 at 11:02 AM, Barak Korren  wrote:

> > able to do "live" storage incrase by adding a new volume to the LVM
> group in
> > the following procedure:
> >
> >  created new virtio volume in the VM from phx-engine
> >  fdisk /dev/vdd then: n -> p -> 1 -> enter -> enter -> t -> 8e -> w
> >  vgextend jenkins_lvm /dev/vdd1
> >  lvextend /dev/mapper/jenkins_lvm-data -L28G
> >  xfs_growfs /dev/mapper/jenkins_lvm-data
> >
> > I was unable to increase the volume size from the engine, and then
> increase
> > the partition size(only create a new partition on the same volume with
> the
> > new increased space in the volume), not sure if that is possible.
> >
> You forgot the 1st thing we told you - "don't create partitions on the
> disk"...
> Instead the whole disk '/dev/vdX' should be formatted as a PV, then you can
> grow it from the engine and then 'pvresize' followed by 'lvresize'.
>
>
> --
> Barak Korren
> bkor...@redhat.com
> RHEV-CI Team
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: new Jenkins VM in PHX lab

2016-02-10 Thread Nadav Goldin
re-installed and configured it without a partition table on the block
device. the updated procedure for increasing the volume size:
if there is still available space in the lvm group jenkins_lvm (can be seen
using vgdisplay):
1. lvextend /dev/mapper/jenkins_lvm-data -L410G
2. xfs_growfs /dev/mapper/jenkins_lvm-data

else:
1. increase volume size in engine-ui
2. pvresize /dev/vdb
3. lvextend /dev/mapper/jenkins_lvm-data -L410G
4. xfs_growfs /dev/mapper/jenkins_lvm-data



On Wed, Feb 10, 2016 at 11:54 AM, Barak Korren  wrote:

> On 10 February 2016 at 11:22, Anton Marchukov  wrote:
> > Hello All.
> >
> > Why do we need LVM at all there? It is good when you cannot resize the
> > underlying disk and have to combine it from several hardware ones into
> one
> > virtual. But here we have "cloud" and disks are already resizeable.
> >
> Becasue LVM lets you do snapshots you can mount and copy somewhere
> else (e.g. to do atomic backups). You cannot do that easily with oVirt
> disk snapshots ATM.
>
>
>
> --
> Barak Korren
> bkor...@redhat.com
> RHEV-CI Team
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


SSL certificates

2016-03-06 Thread Nadav Goldin
Hi David & Anton,
Have we had any changes in the SSL certificates for ovirt.org recently?
I suspect that is what failing the infra-supp...@ovirt.org emailing right
now.

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: SSL certificates

2016-03-06 Thread Nadav Goldin
switched it to use unsecured imap until we I fix the certificates issue(or
something else, though I'm pretty sure that is the issue).
apologizes for the tickets flood.


On Sun, Mar 6, 2016 at 11:12 AM, Eyal Edri  wrote:

> Adding Mikey,
> The site was moved to github pages, so its probably related to the
> migration.
>
> e/
>
> On Sun, Mar 6, 2016 at 9:09 AM, Nadav Goldin  wrote:
>
>> Hi David & Anton,
>> Have we had any changes in the SSL certificates for ovirt.org recently?
>> I suspect that is what failing the infra-supp...@ovirt.org emailing
>> right now.
>>
>> Nadav.
>>
>>
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>
>
> --
> Eyal Edri
> Associate Manager
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: SSL certificates

2016-03-07 Thread Nadav Goldin
ok, thanks.
yep the star* one is what dovecot in lists.ovirt.org is using. probably
something else then.


On Mon, Mar 7, 2016 at 10:24 AM, David Caro Estevez 
wrote:

> On 03/06 11:09, Nadav Goldin wrote:
> > Hi David & Anton,
> > Have we had any changes in the SSL certificates for ovirt.org recently?
> > I suspect that is what failing the infra-supp...@ovirt.org emailing
> right
> > now.
>
> We added ssl support to ovirt.org http service, but afaik that's not
> related to
> any other (imap/smtp...) as they use different services and everything.
>
> The certificate we used for ovirt.org is the star certificate we use for
> all
> the other ovirt services (it's valid, and publicly verifiable).
>
> >
> > Nadav.
>
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
>
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


gerrit.ovirt.org restart

2016-03-26 Thread Nadav Goldin
Hi,
gerrit started throwing weird errors this morning, such as:

> internal server error: com.google.gerrit.server.git.UpdateException:
> java.io.IOException: java.util.concurrent.ExecutionException:
> org.apache.lucene.store.AlreadyClosedException: this IndexWriter is closed)
>

restarting it now.


Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Supporting Ansible as another tool for ovirt infra mgmt

2016-04-03 Thread Nadav Goldin
I think another point for consideration is the Puppet+Foreman support,
foreman doesn't support puppet 4 yet[1], but f23 runs only with puppet >4
agents, if that doesn't get fixed soon and we can't upgrade puppet to 4, we
will hit a big problem when we need to migrate more slaves to f23.

[1] http://projects.theforeman.org/issues/8447


On Sun, Apr 3, 2016 at 10:42 AM, Barak Korren  wrote:

> On 3 April 2016 at 10:21, Eyal Edri  wrote:
> > I'd like to ask the team what do you think on $subject, in terms of pros
> &
> > cons.
> >
> > As you all know we have been using puppet to manage our production infra
> > (user access, server configuration,etc... ).
> >
> > Recently we started looking into migrating our mailman instance into new
> > hyper-kitty instance to run on the oVirt DC in PHX.
> > It seems that there is no true puppet classes available to install/manage
> > mailman3 but there is an Ansible playbook used / written by fedora to
> deploy
> > their instance.
> > So the question is should we start using/supporting Ansible as another
> tool
> > to manage our infra and leverage existing playbooks out there to reduce
> work
> > on migration of new services?
> > I'm not suggesting to migrate all puppet code into Ansible, but to allow
> > using Ansible when it make sense.
> >
> > Here is what I had in mind with regards to pro/cons:
> > Pros
> >
> > Saving time writing puppet classes for services (if Ansible playbook
> exists)
> > Bringing in new infra members which are interested in Ansbile (maybe
> publish
> > the team in the relevant communities)
> >
> >
> > Cons:
> >
> > Another tool to support/maintain
> > No real support to manage in foreman as we do for puppet (for sure not in
> > our old version)
> >
> >
> >
> > I'd love to hear your thoughts about it.
> >
>
> As I've already stated elsewhere Ansible is interesting for a number
> of reasons, but a dual-tool scenario will not be welcome IMO.
>
> There is also a lage question of the possibility of replacing Puppet
> with Ansible. Puppet is a continues configuration-management system
> that monitors servers for configuration drift and repairs it
> (deploying missing components in the process), to do that it supports
> a declarative language and a master/slave-agent architecture.
> The common Ansible usage scenario OTOH seems to be AFAIK a
> developer/op launching a deployment task from his laptop. For that
> Ansible supports a more imperative syntax and an SSH-based agent-less
> architecture.
>
> IMO, for long-running on-premise infrastructure (Not ad-hoc in "the
> cloud") which is what oVirt has and what what it targets, the drift
> monitoring approach is more suitable.
>
> Now, I've hared that that Ansible could also be deployed with agents
> and a central server (Tower? Foreman? something else?), but I'm not
> sure how mature that deployment scenario is right now, nor wither
> existing Ansible code fits that scenario.
>
>
> --
> Barak Korren
> bkor...@redhat.com
> RHEV-CI Team
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Lago jobs migration to jenkins.phx.ovirt.org

2016-04-03 Thread Nadav Goldin
Hey David,
as part of the migration to jenkins.phx.ovirt.org,I want to advance with
the Lago jobs. I already migrated
infra-puppet/infra-docs/ovirt-node/appliance/imgbased jobs, and so far they
seem all to work. As far as I understand the Lago jobs are pretty
independent so it should be rather simple. Currently there are 3 slaves
configured (fc23, el7, fc21).

At the fist stage(until we finish the migration)
jenkins_master_deploy-configs_merged is not running, so we could control
which jobs get migrated. So if a patch to the jenkins yaml will be
introduced during the migration process it will have to be re-run manually.

After migrating I'll disable the lago jobs in jenkins.ovirt.org, so even if
JJB runs we will have only one jenkins running the CI checks.

One more question is if there are any other jobs which are dependent on the
Lago jobs(like the publishers which are dependent on all build_artifacts on
ovirt-node/appliance/node)

As far as I understand the only thing needed for migration is updating the
github api tokens and running JJB with *lago*.

What do you think?


Thanks

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Lago jobs migration to jenkins.phx.ovirt.org

2016-04-03 Thread Nadav Goldin
Hi Sandro,
currently you need to sign up and I'll add you the permissions(you can self
enrol in the main page)



On Mon, Apr 4, 2016 at 9:23 AM, Sandro Bonazzola 
wrote:

>
>
> On Sun, Apr 3, 2016 at 7:27 PM, Nadav Goldin  wrote:
>
>> Hey David,
>> as part of the migration to jenkins.phx.ovirt.org,I want to advance with
>> the Lago jobs. I already migrated
>> infra-puppet/infra-docs/ovirt-node/appliance/imgbased jobs, and so far they
>> seem all to work. As far as I understand the Lago jobs are pretty
>> independent so it should be rather simple. Currently there are 3 slaves
>> configured (fc23, el7, fc21).
>>
>> At the fist stage(until we finish the migration)
>> jenkins_master_deploy-configs_merged is not running, so we could control
>> which jobs get migrated. So if a patch to the jenkins yaml will be
>> introduced during the migration process it will have to be re-run manually.
>>
>> After migrating I'll disable the lago jobs in jenkins.ovirt.org, so even
>> if JJB runs we will have only one jenkins running the CI checks.
>>
>> One more question is if there are any other jobs which are dependent on
>> the Lago jobs(like the publishers which are dependent on all
>> build_artifacts on ovirt-node/appliance/node)
>>
>> As far as I understand the only thing needed for migration is updating
>> the github api tokens and running JJB with *lago*.
>>
>> What do you think?
>>
>
> Can you migrate user accounts as well? Can't login into
> jenkins.phx.ovirt.org
>
>
>>
>>
>> Thanks
>>
>> Nadav.
>>
>>
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>
>
> --
> Sandro Bonazzola
> Better technology. Faster innovation. Powered by community collaboration.
> See how it works at redhat.com
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Lago jobs migration to jenkins.phx.ovirt.org

2016-04-04 Thread Nadav Goldin
done, try now.


On Mon, Apr 4, 2016 at 9:57 AM, Sandro Bonazzola 
wrote:

>
>
> On Mon, Apr 4, 2016 at 8:35 AM, Nadav Goldin  wrote:
>
>> Hi Sandro,
>> currently you need to sign up and I'll add you the permissions(you can
>> self enrol in the main page)
>>
>
> User created:
> Access Denied
>
> sbonazzo is missing the Overall/Read permission
>
>
>
>
>>
>>
>>
>> On Mon, Apr 4, 2016 at 9:23 AM, Sandro Bonazzola 
>> wrote:
>>
>>>
>>>
>>> On Sun, Apr 3, 2016 at 7:27 PM, Nadav Goldin  wrote:
>>>
>>>> Hey David,
>>>> as part of the migration to jenkins.phx.ovirt.org,I want to advance
>>>> with the Lago jobs. I already migrated
>>>> infra-puppet/infra-docs/ovirt-node/appliance/imgbased jobs, and so far they
>>>> seem all to work. As far as I understand the Lago jobs are pretty
>>>> independent so it should be rather simple. Currently there are 3 slaves
>>>> configured (fc23, el7, fc21).
>>>>
>>>> At the fist stage(until we finish the migration)
>>>> jenkins_master_deploy-configs_merged is not running, so we could control
>>>> which jobs get migrated. So if a patch to the jenkins yaml will be
>>>> introduced during the migration process it will have to be re-run manually.
>>>>
>>>> After migrating I'll disable the lago jobs in jenkins.ovirt.org, so
>>>> even if JJB runs we will have only one jenkins running the CI checks.
>>>>
>>>> One more question is if there are any other jobs which are dependent on
>>>> the Lago jobs(like the publishers which are dependent on all
>>>> build_artifacts on ovirt-node/appliance/node)
>>>>
>>>> As far as I understand the only thing needed for migration is updating
>>>> the github api tokens and running JJB with *lago*.
>>>>
>>>> What do you think?
>>>>
>>>
>>> Can you migrate user accounts as well? Can't login into
>>> jenkins.phx.ovirt.org
>>>
>>>
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Nadav.
>>>>
>>>>
>>>> ___
>>>> Infra mailing list
>>>> Infra@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>>
>>>>
>>>
>>>
>>> --
>>> Sandro Bonazzola
>>> Better technology. Faster innovation. Powered by community collaboration.
>>> See how it works at redhat.com
>>>
>>
>>
>
>
> --
> Sandro Bonazzola
> Better technology. Faster innovation. Powered by community collaboration.
> See how it works at redhat.com
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Lago jobs migration to jenkins.phx.ovirt.org

2016-04-04 Thread Nadav Goldin
Just to make sure, don't confuse lago jobs with ovirt-system tests, the
> ovirt-system tests use bare metal slaves, those can't be created from
> templates
> so those can only be installed with pxe (or physically on site, as the
> virtual
> media of the ilo does not work so good)
>
sure, only lago* jobs, either way we can have a final check before I add
the github credentials.

> We already have a template for fc23, it's just creating a new slave from
> that
> > template (can be done from foreman too, faster than installing from
> scratch).

thought we could migrate all lago jobs with the slaves at once, but because
of [1]
NGN needs them too, so either way we need to have fc23 slaves on both
jenkins's until
full migration.

+1, lets try to install a few new f23 slaves from template then.
>
working on it, will update.



[1] https://ovirt-jira.atlassian.net/browse/OVIRT-461

On Mon, Apr 4, 2016 at 12:10 PM, Eyal Edri  wrote:

>
>
> On Mon, Apr 4, 2016 at 11:52 AM, David Caro Estevez 
> wrote:
>
>> On 04/04 11:49, Eyal Edri wrote:
>> > On Mon, Apr 4, 2016 at 10:38 AM, David Caro Estevez 
>> > wrote:
>> >
>> > > On 04/03 20:27, Nadav Goldin wrote:
>> > > > Hey David,
>> > > > as part of the migration to jenkins.phx.ovirt.org,I want to
>> advance with
>> > > > the Lago jobs. I already migrated
>> > > > infra-puppet/infra-docs/ovirt-node/appliance/imgbased jobs, and so
>> far
>> > > they
>> > > > seem all to work. As far as I understand the Lago jobs are pretty
>> > > > independent so it should be rather simple. Currently there are 3
>> slaves
>> > > > configured (fc23, el7, fc21).
>> > >
>> > >
>> > > There are only 3 fc23 slaves, having one less, duplicates the check
>> run
>> > > time,
>> > > and having only one, triplicates it, can you create new slaves
>> instead of
>> > > moving them from the old jenkins? (lago is not the only one using
>> them, so
>> > > migrating all of them is not an option)
>> > >
>> >
>> > Is it possible to add new slaves with the current state of pxe not
>> working?
>> > The ideal will be to have all new servers installed with the pxe fixed
>> so
>> > we can deploy many more slaves.
>> > This way we can just add lots of slaves to the new jenkins.
>>
>> We already have a template for fc23, it's just creating a new slave from
>> that
>> template (can be done from foreman too, faster than installing from
>> scratch).
>>
>
> +1, lets try to install a few new f23 slaves from template then.
>
>
>>
>> >
>> >
>> > >
>> > >
>> > > >
>> > > > At the fist stage(until we finish the migration)
>> > > > jenkins_master_deploy-configs_merged is not running, so we could
>> control
>> > > > which jobs get migrated. So if a patch to the jenkins yaml will be
>> > > > introduced during the migration process it will have to be re-run
>> > > manually.
>> > > >
>> > > > After migrating I'll disable the lago jobs in jenkins.ovirt.org, so
>> > > even if
>> > > > JJB runs we will have only one jenkins running the CI checks.
>> > >
>> > > Don't allow both to run anything at the same time, that will lead to
>> > > confusion
>> > > and branches being deleted at strange times on the github repo, if
>> they
>> > > run on
>> > > one jenkins master, run them there only.
>> > >
>> > > >
>> > > > One more question is if there are any other jobs which are
>> dependent on
>> > > the
>> > > > Lago jobs(like the publishers which are dependent on all
>> build_artifacts
>> > > on
>> > > > ovirt-node/appliance/node)
>> > >
>> > > Lago is self-contained, anything lago needs (check-build-deploy) is
>> tagged
>> > > as
>> > > lago*, any other job that uses lago, get's it from the repos.
>> > >
>> > > >
>> > > > As far as I understand the only thing needed for migration is
>> updating
>> > > the
>> > > > github api tokens and running JJB with *lago*.
>> > >
>> > > And disabling the jobs on the other jenkins.
>> > > The github configuration is not trivial though, the api token is
>> valid only
>&g

Re: Lago jobs migration to jenkins.phx.ovirt.org

2016-04-04 Thread Nadav Goldin
updates:
- lago jobs were migrated, triggering pull request was tested and working,
lago-bot commenting and check_merged still needs to be tested.
- 4 new fc23 VMs were added to the new Jenkins instance(fc23-vm04-07)
- 1 new el7 VM was added (el7-vm25)

I've given admin permissions to all infra members who already enrolled, in
case anyone needs access.



On Mon, Apr 4, 2016 at 12:16 PM, Nadav Goldin  wrote:

>
>
> Just to make sure, don't confuse lago jobs with ovirt-system tests, the
>> ovirt-system tests use bare metal slaves, those can't be created from
>> templates
>> so those can only be installed with pxe (or physically on site, as the
>> virtual
>> media of the ilo does not work so good)
>>
> sure, only lago* jobs, either way we can have a final check before I add
> the github credentials.
>
> > We already have a template for fc23, it's just creating a new slave
>> from that
>> > template (can be done from foreman too, faster than installing from
>> scratch).
>
> thought we could migrate all lago jobs with the slaves at once, but
> because of [1]
> NGN needs them too, so either way we need to have fc23 slaves on both
> jenkins's until
> full migration.
>
> +1, lets try to install a few new f23 slaves from template then.
>>
> working on it, will update.
>
>
>
> [1] https://ovirt-jira.atlassian.net/browse/OVIRT-461
>
> On Mon, Apr 4, 2016 at 12:10 PM, Eyal Edri  wrote:
>
>>
>>
>> On Mon, Apr 4, 2016 at 11:52 AM, David Caro Estevez 
>> wrote:
>>
>>> On 04/04 11:49, Eyal Edri wrote:
>>> > On Mon, Apr 4, 2016 at 10:38 AM, David Caro Estevez 
>>> > wrote:
>>> >
>>> > > On 04/03 20:27, Nadav Goldin wrote:
>>> > > > Hey David,
>>> > > > as part of the migration to jenkins.phx.ovirt.org,I want to
>>> advance with
>>> > > > the Lago jobs. I already migrated
>>> > > > infra-puppet/infra-docs/ovirt-node/appliance/imgbased jobs, and so
>>> far
>>> > > they
>>> > > > seem all to work. As far as I understand the Lago jobs are pretty
>>> > > > independent so it should be rather simple. Currently there are 3
>>> slaves
>>> > > > configured (fc23, el7, fc21).
>>> > >
>>> > >
>>> > > There are only 3 fc23 slaves, having one less, duplicates the check
>>> run
>>> > > time,
>>> > > and having only one, triplicates it, can you create new slaves
>>> instead of
>>> > > moving them from the old jenkins? (lago is not the only one using
>>> them, so
>>> > > migrating all of them is not an option)
>>> > >
>>> >
>>> > Is it possible to add new slaves with the current state of pxe not
>>> working?
>>> > The ideal will be to have all new servers installed with the pxe fixed
>>> so
>>> > we can deploy many more slaves.
>>> > This way we can just add lots of slaves to the new jenkins.
>>>
>>> We already have a template for fc23, it's just creating a new slave from
>>> that
>>> template (can be done from foreman too, faster than installing from
>>> scratch).
>>>
>>
>> +1, lets try to install a few new f23 slaves from template then.
>>
>>
>>>
>>> >
>>> >
>>> > >
>>> > >
>>> > > >
>>> > > > At the fist stage(until we finish the migration)
>>> > > > jenkins_master_deploy-configs_merged is not running, so we could
>>> control
>>> > > > which jobs get migrated. So if a patch to the jenkins yaml will be
>>> > > > introduced during the migration process it will have to be re-run
>>> > > manually.
>>> > > >
>>> > > > After migrating I'll disable the lago jobs in jenkins.ovirt.org,
>>> so
>>> > > even if
>>> > > > JJB runs we will have only one jenkins running the CI checks.
>>> > >
>>> > > Don't allow both to run anything at the same time, that will lead to
>>> > > confusion
>>> > > and branches being deleted at strange times on the github repo, if
>>> they
>>> > > run on
>>> > > one jenkins master, run them there only.
>>> > >
>>> > > >
>>> > > > One more question is if there are any other jobs which are
&g

Re: Lago jobs migration to jenkins.phx.ovirt.org

2016-04-06 Thread Nadav Goldin
On Wed, Apr 6, 2016 at 1:57 PM, David Caro Estevez  wrote:

> hings that I had to change so far:
>
> * Allow non-logged in users to read jobs (not managed by puppet)
>
> * Add a new credential for the lago deploy job (ssh to resources.ovirt.org
> as
>   lago-deploy-snapshot user, not in puppet, using private key)
>
> * Upgrade the ssh-agent plugin and restart jenkins, as it was not pulling
>   correctly the upgraded plugin just by 'reloading'
>   https://gerrit.ovirt.org/#/c/55722/
>
commented on the patch, did the ssh-agent upgrade solve it?

>
>
> btw. the jenkins host is currently failing to run puppet (it's using
> testing
> env), so I was unable to actually verify any patches, as I did not want to
> mess
> up any on-going tests
>
sure I'll cherry-pick and test.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Dropping support for none Standard-CI packages on Jenkins' slaves

2016-04-11 Thread Nadav Goldin
Hi,
As most jobs were migrated to standard CI, we still have prior puppet code
that installs various packages on all slaves. These packages cause 2
problems:
1) mask possible bugs by using different packages than the ones intended in
the standard CI files.
2) overhead and unneeded complication in puppet and the VM templates.

to ensure they are indeed not needed any more, I want to start removing the
packages gradually,
if no one objects I'll start by removing the following packages:

jasperreports-server
> postgresql-jdbc
> libnl
> log4j
> chrpath
> sos
>
mailcap
>




Thanks
Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: out of sync hosts in forman upstream

2016-04-11 Thread Nadav Goldin
>
> artifactory.ovirt.org
>  - tried to run
> puppet agent -t. It completed successfully but the host
>  is still not synced
>

what is the error?



> deb81-vm01.phx.ovirt.org
>  - Cannot login
>
> not sure who set that up and when, our puppet code doesn't support debian
anyways afaik.

> grafana.phx.ovirt.org
>  - no ping
>
this machine is shutdown(under testing), I'll disable the alerts.

>
> graphite.phx.ovirt.org
>  - Cannot login
>
I'll fix.

>
> monitoring.phx.ovirt.org
>  - no ping
>
same as grafana.


On Mon, Apr 11, 2016 at 6:43 PM, Gil Shinar  wrote:

> Hi,
>
> Here is a list of out of sync hosts in the upstream forman:
> artifactory.ovirt.org
>  - tried to run
> puppet agent -t. It completed successfully but the host
>  is still not synced
>
> deb81-vm01.phx.ovirt.org
>  - Cannot login
>
> grafana.phx.ovirt.org
>  - no ping
>
> graphite.phx.ovirt.org
>  - Cannot login
>
> monitoring.phx.ovirt.org
>  - no ping
>
> How should I handle these hosts?
>
> Thanks
> Gil
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Experimental Jenkins monitoring

2016-04-14 Thread Nadav Goldin
Hi,
I've created an experimental dashboard for Jenkins at our Grafana instance:
http://graphite.phx.ovirt.org/dashboard/db/jenkins-monitoring
(if you don't have an account, you can enrol with github/google)

currently it collects the following metrics:
1) How many jobs in the Build Queue are waiting per slaves' label:

for instance: if there are 4 builds of a job that is restricted to 'el7'
and 2 builds of another job
which is restricted to 'el7' in the build queue we will see 6 for 'el7' in
the first graph.
'No label' sums jobs which are waiting but are unrestricted.

2) How many slaves are idle per label.
note that the slave's labels are contained in the job's labels, but not
vice versa, as
we allow regex expressions such as (fc21 || fc22 ). right now it treats
them as simple
strings.

3) Total number of online/offline/idle slaves

besides the normal monitoring, it can help us:
1) minimize the difference between 'idle' slaves per label and jobs waiting
in the build queue per label.
this might be caused by unnecessary restrictions on the label, or maybe by
the
'Throttle Concurrent Builds' plugin.
2) decide how many VMs and which OS to install on the new hosts.
3) in the future, once we have the 'slave pools' implemented, we could
implement
auto-scaling based on thresholds or some other function.


'experimental' - as it still needs to be tested for stability(it is based
on python-jenkins
and graphite-send) and also more metrics can be added(maybe avg running time
per job? builds per hour? ) - will be happy to hear.

I plan later to pack it all into independent fabric tasks(i.e. fab
do.jenkins.slaves.show)


Nadav
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Experimental Jenkins monitoring

2016-04-17 Thread Nadav Goldin
>
> I think that will change a lot per-project basis, if we can get that info
> per
> job, with grafana then we can aggregate and create secondary stats (like
> bilds
> per hour as you say).
> So I'd say just to collect the 'bare' data, like job built event, job
> ended,
> duration and such.

agree. will need to improve that, right now it 'pulls' each X seconds via
the CLI,
instead of Jenkins sending the events, so it is limited to what the CLI can
provide and not that efficient. I plan to install [1] and do the opposite
(Jenkins will send a POST request with the data on each build
event and then it would be sent to graphite)

Have you checked the current ds fabric checks?
> There are already a bunch of fabric tasks that monitor jenkins, if we
> install
> the nagiosgraph (see ds for details) to send the nagios performance data
> into
> graphite, we can use them as is to also start alarms and such
>
Icinga2 has integrated graphite support, so after the upgrade we will
get all of our alarms data sent to graphite 'out-of-the-box'.

>
> dcaro@akhos$ fab -l | grep nagi
> do.jenkins.nagios.check_build_load  Checks if the
> bui...
> do.jenkins.nagios.check_executors   Checks if the
> exe...
> do.jenkins.nagios.check_queue   Check if the
> buil...
> do.provision.nagios_check   Show a summary
> of...
>
> Though those will not give you the bare data (were designed with nagios in
> mind, not graphite so they are just checks, the stats were added later)
>
> There's also a bunch of helpers functions to create nagios checks too.
>

cool, wasn't aware of those fabric checks.
I think for simple metrics(loads and such) we could use that(i.e. query
Jenkins from fabric)
but for more complicated queries we'd need to query graphite itself,
with this[2] I could create scripts that query graphite and trigger Icinga
alerts.
such as: calculate the 'expected' slaves load for the next hour(in graphite)
and then:
Icinga queries graphite -> triggers another Icinga alert -> triggers custom
script(such as
fab task to create slaves)

for now, added two more metrics: top 10 jobs in past X time, and
avg number of builds running / builds waiting in queue in the past X time.
some metrics might 'glitch' from time to time as there is not a lot of data
yet
and it mainly counts integer values while graphite is oriented towards
floats, so the data has to be smoothed(usually with movingAverage())



[1]
https://wiki.jenkins-ci.org/display/JENKINS/Statistics+Notification+Plugin
[2] https://github.com/klen/graphite-beacon

On Fri, Apr 15, 2016 at 9:39 AM, David Caro  wrote:

> On 04/15 01:24, Nadav Goldin wrote:
> > Hi,
> > I've created an experimental dashboard for Jenkins at our Grafana
> instance:
> > http://graphite.phx.ovirt.org/dashboard/db/jenkins-monitoring
> > (if you don't have an account, you can enrol with github/google)
>
> Nice! \o/
>
> >
> > currently it collects the following metrics:
> > 1) How many jobs in the Build Queue are waiting per slaves' label:
> >
> > for instance: if there are 4 builds of a job that is restricted to 'el7'
> > and 2 builds of another job
> > which is restricted to 'el7' in the build queue we will see 6 for 'el7'
> in
> > the first graph.
> > 'No label' sums jobs which are waiting but are unrestricted.
> >
> > 2) How many slaves are idle per label.
> > note that the slave's labels are contained in the job's labels, but not
> > vice versa, as
> > we allow regex expressions such as (fc21 || fc22 ). right now it treats
> > them as simple
> > strings.
> >
> > 3) Total number of online/offline/idle slaves
> >
> > besides the normal monitoring, it can help us:
> > 1) minimize the difference between 'idle' slaves per label and jobs
> waiting
> > in the build queue per label.
> > this might be caused by unnecessary restrictions on the label, or maybe
> by
> > the
> > 'Throttle Concurrent Builds' plugin.
> > 2) decide how many VMs and which OS to install on the new hosts.
> > 3) in the future, once we have the 'slave pools' implemented, we could
> > implement
> > auto-scaling based on thresholds or some other function.
> >
> >
> > 'experimental' - as it still needs to be tested for stability(it is based
> > on python-jenkins
> > and graphite-send) and also more metrics can be added(maybe avg running
> time
> > per job? builds per hour? ) - will be happy to hear.
>
> I think 

Jenkins.ovirt.org Upgrade | 24/04/2016(Sunday) - 18:00 TLV

2016-04-21 Thread Nadav Goldin
Hey all,
On Sunday, 24.04.2015 - 18:00 TLV (17:00 CET) we plan to migrate Jenkins(
http://jenkins.ovirt.org) to a new VM in the PHX datacenter, this will
increase the instance storage and allow better connectivity with the slaves.

1. The expected downtime is 2 hours, during that time no patches will be
checked and you could not login to Jenkins. Patches sent to gerrit during
the downtime, might get checked after the downtime.

2.What will be migrated:
All yamlized jobs, global configuration and most of the existing slaves.

3. The old Jenkins instance will still be available under
jenkins-old.ovirt.org, with a minimum number of slaves. It will be kept at
least in the following months for backup and for the none-yamlized jobs(but
with no gerrit triggers)

4. User authentication: all users will have to enrol again, this can be
done already this week via http://jenkins.phx.ovirt.org

Another reminder will be sent ~ 2 hours before the migration.

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


[ATTENTION NEEDED] Jenkins upgrade | Today(24/04/2016) - 18:00 TLV

2016-04-24 Thread Nadav Goldin
Hey all,

Reminder: http://jenkins.ovirt.org will be down today between 18:00-20:00
TLV(17:00-19:00 CET),
details below.

Thanks,
Nadav.

-- Forwarded message --
From: Nadav Goldin 
Date: Thu, Apr 21, 2016 at 10:59 AM
Subject: [ovirt-devel] Jenkins.ovirt.org Upgrade | 24/04/2016(Sunday) -
18:00 TLV
To: devel , infra 


Hey all,
On Sunday, 24.04.2015 - 18:00 TLV (17:00 CET) we plan to migrate Jenkins(
http://jenkins.ovirt.org) to a new VM in the PHX datacenter, this will
increase the instance storage and allow better connectivity with the slaves.

1. The expected downtime is 2 hours, during that time no patches will be
checked and you could not login to Jenkins. Patches sent to gerrit during
the downtime, might get checked after the downtime.

2.What will be migrated:
All yamlized jobs, global configuration and most of the existing slaves.

3. The old Jenkins instance will still be available under
jenkins-old.ovirt.org, with a minimum number of slaves. It will be kept at
least in the following months for backup and for the none-yamlized jobs(but
with no gerrit triggers)

4. User authentication: all users will have to enrol again, this can be
done already this week via http://jenkins.phx.ovirt.org

Another reminder will be sent ~ 2 hours before the migration.

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Jenkins.ovirt.org Upgrade | 24/04/2016(Sunday) - 18:00 TLV

2016-04-24 Thread Nadav Goldin
The migration was completed, if you encounter any problems please email
infra-supp...@ovirt.org to track the issue.
The old instance, with the none-yamlized jobs can be found in
http://jenkins-old.ovirt.org,
if you need any of the none-yamlized jobs feel free to enable them in
http://jenkins-old.ovirt.org(only yamlized jobs were migrated)





On Thu, Apr 21, 2016 at 10:59 AM, Nadav Goldin  wrote:

> Hey all,
> On Sunday, 24.04.2015 - 18:00 TLV (17:00 CET) we plan to migrate Jenkins(
> http://jenkins.ovirt.org) to a new VM in the PHX datacenter, this will
> increase the instance storage and allow better connectivity with the slaves.
>
> 1. The expected downtime is 2 hours, during that time no patches will be
> checked and you could not login to Jenkins. Patches sent to gerrit during
> the downtime, might get checked after the downtime.
>
> 2.What will be migrated:
> All yamlized jobs, global configuration and most of the existing slaves.
>
> 3. The old Jenkins instance will still be available under
> jenkins-old.ovirt.org, with a minimum number of slaves. It will be kept
> at least in the following months for backup and for the none-yamlized
> jobs(but with no gerrit triggers)
>
> 4. User authentication: all users will have to enrol again, this can be
> done already this week via http://jenkins.phx.ovirt.org
>
> Another reminder will be sent ~ 2 hours before the migration.
>
> Thanks,
>
> Nadav.
>
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


none-yamlized jobs

2016-04-24 Thread Nadav Goldin
Hey Sandro,
[1] is a list of all the none-yamlized jobs in jenkins-old.ovirt.org, can
you help us map which jobs still need to be enabled? we already mapped dao
and find_bugs, we want to minimize the number of jobs that are not yamlized
yet and must be enabled in jenkins-old.ovirt.org


Thanks,

Nadav.



[1] https://paste.fedoraproject.org/359265/
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


User authentication on http://jenkins.ovirt.org

2016-04-25 Thread Nadav Goldin
Hi all,
I note again that personal user accounts were not migrated to the new
Jenkins instance,
so if you need other than read-only access please enrol again(press sign up
in the welcome page).

If you require permissions to trigger jobs manually:
email infra-supp...@ovirt.org 'please put my Jenkins $username, in dev role
group'

for other permissions(create job, etc) - if it is a 'testing' job feel free
to use http://jenkins-old.ovirt.org, as we are trying to minimize the
number of none-yamlized jobs in the new instance. If that is not suitable
please email what you need.


Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


vdsm-jsonrpc-java_{master,3.6}_build-artifacts-fc23 Failure

2016-04-25 Thread Nadav Goldin
Hey Piotr and Oved,
Can you have a look at:
http://jenkins.ovirt.org/job/vdsm-jsonrpc-java_master_build-artifacts-fc23-x86_64/3/console
http://jenkins.ovirt.org/job/vdsm-jsonrpc-java_3.6_build-artifacts-fc23-x86_64/1/console

>> 07:02:35 Tests run: 3, Failures: 0, Errors: 3, Skipped: 0, Time elapsed:
> 560.634 sec <<< FAILURE! - in
> org.ovirt.vdsm.jsonrpc.client.reactors.stomp.SSLStompClientTestCase
> >> 07:02:35
> testLongMessage(org.ovirt.vdsm.jsonrpc.client.reactors.stomp.SSLStompClientTestCase)
>  Time elapsed: 180.56 sec  <<< ERROR!
> >> 07:02:35 org.ovirt.vdsm.jsonrpc.client.ClientConnectionException:
> Connection timeout


to ensure this isn't related to the Jenkins migration I re-triggered it on
the old instance with the same results(
http://jenkins-old.ovirt.org/job/vdsm-jsonrpc-java_master_build-artifacts-fc23-x86_64/14/
)

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


httpcomponets-cliet_master/core_master create-rpms

2016-04-25 Thread Nadav Goldin
Hey Juan,

>
> http://jenkins-old.ovirt.org/job/httpcomponents-client_master_create-rpms_merged/
>
> http://jenkins-old.ovirt.org/job/httpcomponents-core_master_create-rpms_merged/
>
do you happened to know if these 2 jobs are still needed? they were last
built
on 06/2014, but they are in the 3.6 publisher(which is failing right now as
we didn't
migrate the past builds, only the yamlized jobs)

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


ovirt-image-uploader_3.6_build-artifacts-el6-x86_64 failure

2016-04-25 Thread Nadav Goldin
Hey Sandro,
http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-el6-x86_64/2/
is failing because:

> DEBUG util.py:417:
> http://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el6/repodata/repomd.xml:
> [Errno 14] HTTP Error 404 - Not Found
>
but there  is no /rpm/el6 for ovirt-master-snapshot already.
should it be ovirt-3.6-snapshot?

Thanks

Nadav
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Change in ovirt-engine[master]: engine : Change SSO classes to CamelCase

2016-04-25 Thread Nadav Goldin
Hi, should be ok now, can you check again?

On Mon, Apr 25, 2016 at 6:17 PM, Martin Perina  wrote:

> Hi,
>
> could you please take a look upgrade jobs? Following error seems to me
> like a CI issue:
>
>   No JDK named ?java-1.8.0-openjdk? found
>
> Thanks
>
> Martin
>
>
>
> On Mon, Apr 25, 2016 at 5:15 PM, Jenkins CI 
> wrote:
>
>> Jenkins CI has posted comments on this change.
>>
>> Change subject: engine : Change SSO classes to CamelCase
>> ..
>>
>>
>> Patch Set 6:
>>
>> Build Failed
>>
>>
>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_upgrade-from-master_el7_merged/6/
>> : FAILURE
>>
>>
>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_upgrade-from-3.6_el7_merged/6/
>> : FAILURE
>>
>>
>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_check-merged-el7-x86_64/6/
>> : SUCCESS
>>
>>
>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_check-merged-fc23-x86_64/6/
>> : SUCCESS
>>
>> --
>> To view, visit https://gerrit.ovirt.org/56410
>> To unsubscribe, visit https://gerrit.ovirt.org/settings
>>
>> Gerrit-MessageType: comment
>> Gerrit-Change-Id: Iaa6b23079549330646701f3b2322db77934e7453
>> Gerrit-PatchSet: 6
>> Gerrit-Project: ovirt-engine
>> Gerrit-Branch: master
>> Gerrit-Owner: Ravi Nori 
>> Gerrit-Reviewer: Jenkins CI
>> Gerrit-Reviewer: Martin Peřina 
>> Gerrit-Reviewer: Ravi Nori 
>> Gerrit-Reviewer: gerrit-hooks 
>> Gerrit-HasComments: No
>>
>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Jenkins login fails

2016-04-25 Thread Nadav Goldin
Hi Piotr,
updated your permissions(user account: pkliczewski)


On Mon, Apr 25, 2016 at 9:48 PM, Piotr Kliczewski <
piotr.kliczew...@gmail.com> wrote:

> On Mon, Apr 25, 2016 at 8:33 PM, Fabian Deutsch 
> wrote:
> > I had to re-register after the migration.
> >
>
> :/ I re-registered but I see that I need the same permissions that I had.
> Not able to retrigger jobs anymore.
>
> Thanks,
> Piotr
>
> > - fabian
> >
> > On Mon, Apr 25, 2016 at 7:04 PM, Piotr Kliczewski
> >  wrote:
> >> Hi,
> >>
> >> I am not able to login to:
> >>
> >> http://jenkins.ovirt.org/
> >>
> >> Is the credentials are different for:
> >>
> >> http://jenkins.phx.ovirt.org/
> >>
> >> Thanks,
> >> Piotr
> >> ___
> >> Infra mailing list
> >> Infra@ovirt.org
> >> http://lists.ovirt.org/mailman/listinfo/infra
> >
> >
> >
> > --
> > Fabian Deutsch 
> > RHEV Hypervisor
> > Red Hat
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ovirt-image-uploader_3.6_build-artifacts-el6-x86_64 failure

2016-04-25 Thread Nadav Goldin
looks like there are 4 jobs failing for that reason:
http://jenkins.ovirt.org/job/ovirt-scheduler-proxy_master_build-artifacts-el6-x86_64
http://jenkins.ovirt.org/job/ovirt-scheduler-proxy_master_build-artifacts-fc22-x86_64/
http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-el6-x86_64
http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-fc22-x86_64

On Mon, Apr 25, 2016 at 6:11 PM, Sandro Bonazzola 
wrote:

>
> Il 25/Apr/2016 16:45, "Nadav Goldin"  ha scritto:
> >
> > Hey Sandro,
> >
> http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-el6-x86_64/2/
> is failing because:
> >>
> >> DEBUG util.py:417:
> http://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el6/repodata/repomd.xml:
> [Errno 14] HTTP Error 404 - Not Found
> >
> > but there  is no /rpm/el6 for ovirt-master-snapshot already.
> > should it be ovirt-3.6-snapshot?
> >
>
> Yes it should
>
> > Thanks
> >
> > Nadav
> >
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ovirt-image-uploader_3.6_build-artifacts-el6-x86_64 failure

2016-04-26 Thread Nadav Goldin
nope, the error is:

> DEBUG util.py:417:  failure: repodata/repomd.xml from
> ovirt-master-snapshot: [Errno 256] No more mirrors to try.
> DEBUG util.py:417:
> http://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el6/repodata/repomd.xml:
> [Errno 14] HTTP Error 404 - Not Found
>
and the reason is that automation/build-artifacts.repo has
ovirt-master-snapshot instead of ovirt-3.6-snapshot,
the reason it wasn't caught until now is that it always took the last
successful build(which could have been a while ago)



On Tue, Apr 26, 2016 at 9:41 AM, Piotr Kliczewski <
piotr.kliczew...@gmail.com> wrote:

> Is it related to java-1.8.0-openjdk-devel-1.8.0.60 issue that I
> reported on this list?
>
> On Tue, Apr 26, 2016 at 12:11 AM, Nadav Goldin  wrote:
> > looks like there are 4 jobs failing for that reason:
> >
> http://jenkins.ovirt.org/job/ovirt-scheduler-proxy_master_build-artifacts-el6-x86_64
> >
> http://jenkins.ovirt.org/job/ovirt-scheduler-proxy_master_build-artifacts-fc22-x86_64/
> >
> http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-el6-x86_64
> >
> http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-fc22-x86_64
> >
> > On Mon, Apr 25, 2016 at 6:11 PM, Sandro Bonazzola 
> > wrote:
> >>
> >>
> >> Il 25/Apr/2016 16:45, "Nadav Goldin"  ha scritto:
> >> >
> >> > Hey Sandro,
> >> >
> >> >
> http://jenkins.ovirt.org/job/ovirt-image-uploader_3.6_build-artifacts-el6-x86_64/2/
> >> > is failing because:
> >> >>
> >> >> DEBUG util.py:417:
> >> >>
> http://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/el6/repodata/repomd.xml
> :
> >> >> [Errno 14] HTTP Error 404 - Not Found
> >> >
> >> > but there  is no /rpm/el6 for ovirt-master-snapshot already.
> >> > should it be ovirt-3.6-snapshot?
> >> >
> >>
> >> Yes it should
> >>
> >> > Thanks
> >> >
> >> > Nadav
> >> >
> >
> >
> >
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
> >
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Change in ovirt-engine[master]: engine : Change SSO classes to CamelCase

2016-04-26 Thread Nadav Goldin
actually it could be re-triggered, though not recommended usually. for
example:
http://jenkins.phx.ovirt.org/job/ovirt-engine_master_upgrade-from-master_el7_merged/10/gerrit-trigger-retrigger-this
either way, it is working now(there were 2 unrelated problems, one with the
openjdk and a missing artifactory xml)

thanks,
Nadav.


On Mon, Apr 25, 2016 at 10:49 PM, Martin Perina  wrote:

> Hi,
>
> how could I do that? Those upgrade jobs are started automatically for
> merged patches only ...
>
> Thanks
>
> Martin
>
> On Mon, Apr 25, 2016 at 6:03 PM, Nadav Goldin  wrote:
>
>> Hi, should be ok now, can you check again?
>>
>> On Mon, Apr 25, 2016 at 6:17 PM, Martin Perina 
>> wrote:
>>
>>> Hi,
>>>
>>> could you please take a look upgrade jobs? Following error seems to me
>>> like a CI issue:
>>>
>>>   No JDK named ?java-1.8.0-openjdk? found
>>>
>>> Thanks
>>>
>>> Martin
>>>
>>>
>>>
>>> On Mon, Apr 25, 2016 at 5:15 PM, Jenkins CI 
>>> wrote:
>>>
>>>> Jenkins CI has posted comments on this change.
>>>>
>>>> Change subject: engine : Change SSO classes to CamelCase
>>>> ..
>>>>
>>>>
>>>> Patch Set 6:
>>>>
>>>> Build Failed
>>>>
>>>>
>>>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_upgrade-from-master_el7_merged/6/
>>>> : FAILURE
>>>>
>>>>
>>>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_upgrade-from-3.6_el7_merged/6/
>>>> : FAILURE
>>>>
>>>>
>>>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_check-merged-el7-x86_64/6/
>>>> : SUCCESS
>>>>
>>>>
>>>> http://jenkins.phx.ovirt.org/job/ovirt-engine_master_check-merged-fc23-x86_64/6/
>>>> : SUCCESS
>>>>
>>>> --
>>>> To view, visit https://gerrit.ovirt.org/56410
>>>> To unsubscribe, visit https://gerrit.ovirt.org/settings
>>>>
>>>> Gerrit-MessageType: comment
>>>> Gerrit-Change-Id: Iaa6b23079549330646701f3b2322db77934e7453
>>>> Gerrit-PatchSet: 6
>>>> Gerrit-Project: ovirt-engine
>>>> Gerrit-Branch: master
>>>> Gerrit-Owner: Ravi Nori 
>>>> Gerrit-Reviewer: Jenkins CI
>>>> Gerrit-Reviewer: Martin Peřina 
>>>> Gerrit-Reviewer: Ravi Nori 
>>>> Gerrit-Reviewer: gerrit-hooks 
>>>> Gerrit-HasComments: No
>>>>
>>>
>>>
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
>>>
>>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: mirrors management

2016-04-26 Thread Nadav Goldin
https://ovirt-jira.atlassian.net/browse/OVIRT-257 is the right think I
think.

Last time we discussed it it was summarized to whether we want
a tool that manages also the mirror lists for yum, or just a simple
monitoring tool such as
http://www.staff.science.uu.nl/~penni101/mirmon/
which I think could be setup quite easily.
either way we didn't reach any decision, I agree with Sandro that
this is quite important, as last time I went over the mirror sites
some were not updated for more than a month(which was fixed
then.. but some time passed)






On Tue, Apr 26, 2016 at 6:30 PM, Eyal Edri  wrote:

> Nadav,
> Can you check we have a ticket on this to track it?
> Sandro, what issues are you seeing now that will require us to prioritize
> this task?
>
> e.
>
> On Tue, Apr 26, 2016 at 6:28 PM, Sandro Bonazzola 
> wrote:
>
>>
>>
>> On Tue, Apr 26, 2016 at 5:24 PM, Eyal Edri  wrote:
>>
>>> You mean this ticket? [1]
>>> I don't think we started working on it yet, If you have suggestions,
>>> please add them to the doc.
>>> What issues are we seeing as a result of that?
>>>
>>>
>> See this thread:
>> http://lists.ovirt.org/pipermail/infra/2015-October/011486.html
>>
>>
>>
>>
>>>
>>>
>>> [1] https://ovirt-jira.atlassian.net/browse/OVIRT-257
>>>
>>> On Tue, Apr 26, 2016 at 6:11 PM, Sandro Bonazzola 
>>> wrote:
>>>
 Hi,
 it has been raised several times now that we have our mirrors not
 synchronized with our main repo.
 I remember that a review of mirror monitoring solution has been done in
 the past.
 Do we have a design for making sure our users miss updates due to
 obsolete mirrors?

 Thanks,
 --
 Sandro Bonazzola
 Better technology. Faster innovation. Powered by community
 collaboration.
 See how it works at redhat.com

 ___
 Infra mailing list
 Infra@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/infra


>>>
>>>
>>> --
>>> Eyal Edri
>>> Associate Manager
>>> RHEV DevOps
>>> EMEA ENG Virtualization R&D
>>> Red Hat Israel
>>>
>>> phone: +972-9-7692018
>>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>>
>>
>>
>>
>> --
>> Sandro Bonazzola
>> Better technology. Faster innovation. Powered by community collaboration.
>> See how it works at redhat.com
>>
>
>
>
> --
> Eyal Edri
> Associate Manager
> RHEV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: mirrors management

2016-04-26 Thread Nadav Goldin
Maybe half a day or a day with automatic updates to the mirrorlist file
(remove unsynced mirrors, currently its a static file served by puppet)


On Tue, Apr 26, 2016 at 6:58 PM, Eyal Edri  wrote:

>
>
> On Tue, Apr 26, 2016 at 6:56 PM, Sandro Bonazzola 
> wrote:
>
>>
>>
>> On Tue, Apr 26, 2016 at 5:30 PM, Eyal Edri  wrote:
>>
>>> Nadav,
>>> Can you check we have a ticket on this to track it?
>>> Sandro, what issues are you seeing now that will require us to
>>> prioritize this task?
>>>
>>
>> We released 3.6.5 last week and users are getting 3.6.3 from outdated
>> mirrors...
>> And it's 2 months old...
>>
>>
> OK,
> Nadav, how much time you do think is needed to setup something initially?
>
>
>>
>>
>>>
>>> e.
>>>
>>> On Tue, Apr 26, 2016 at 6:28 PM, Sandro Bonazzola 
>>> wrote:
>>>


 On Tue, Apr 26, 2016 at 5:24 PM, Eyal Edri  wrote:

> You mean this ticket? [1]
> I don't think we started working on it yet, If you have suggestions,
> please add them to the doc.
> What issues are we seeing as a result of that?
>
>
 See this thread:
 http://lists.ovirt.org/pipermail/infra/2015-October/011486.html




>
>
> [1] https://ovirt-jira.atlassian.net/browse/OVIRT-257
>
> On Tue, Apr 26, 2016 at 6:11 PM, Sandro Bonazzola  > wrote:
>
>> Hi,
>> it has been raised several times now that we have our mirrors not
>> synchronized with our main repo.
>> I remember that a review of mirror monitoring solution has been done
>> in the past.
>> Do we have a design for making sure our users miss updates due to
>> obsolete mirrors?
>>
>> Thanks,
>> --
>> Sandro Bonazzola
>> Better technology. Faster innovation. Powered by community
>> collaboration.
>> See how it works at redhat.com
>>
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>
>
> --
> Eyal Edri
> Associate Manager
> RHEV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>



 --
 Sandro Bonazzola
 Better technology. Faster innovation. Powered by community
 collaboration.
 See how it works at redhat.com

>>>
>>>
>>>
>>> --
>>> Eyal Edri
>>> Associate Manager
>>> RHEV DevOps
>>> EMEA ENG Virtualization R&D
>>> Red Hat Israel
>>>
>>> phone: +972-9-7692018
>>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>>
>>
>>
>>
>> --
>> Sandro Bonazzola
>> Better technology. Faster innovation. Powered by community collaboration.
>> See how it works at redhat.com
>>
>
>
>
> --
> Eyal Edri
> Associate Manager
> RHEV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: [Gluster-infra] [ovirt-users] [Attention needed] GlusterFS repository down - affects CI / Installations

2016-04-27 Thread Nadav Goldin
adding infra

-- Forwarded message --
From: Niels de Vos 
Date: Wed, Apr 27, 2016 at 12:09 PM
Subject: Re: [Gluster-infra] [ovirt-users] [Attention needed] GlusterFS
repository down - affects CI / Installations
To: Ravishankar N 
Cc: devel , gluster-infra ,
Nadav Goldin , "gluster-us...@gluster.org List" <
gluster-us...@gluster.org>, us...@ovirt.org


On Wed, Apr 27, 2016 at 02:30:57PM +0530, Ravishankar N wrote:
> @gluster infra  - FYI.
>
> On 04/27/2016 02:20 PM, Nadav Goldin wrote:
> >Hi,
> >The GlusterFS repository became unavailable this morning, as a result all
> >Jenkins jobs that use the repository will fail, the common error would
be:
> >
> >
http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-7/noarch/repodata/repomd.xml
:
> >[Errno 14] HTTP Error 403 - Forbidden
> >
> >
> >Also, installations of oVirt will fail.

I thought oVirt moved to using the packages from the CentOS Storage SIG?
In any case, automated tests should probably use those instead of the
packages on download.gluster.org. We're trying to minimize the work
packagers need to do, and get the glusterfs and other components in the
repositories that are provided by different distributions.

For more details, see the quickstart for the Storage SIG here:
  https://wiki.centos.org/SpecialInterestGroup/Storage/gluster-Quickstart

HTH,
Niels


signature.asc
Description: PGP signature
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: none-yamlized jobs

2016-04-30 Thread Nadav Goldin
thanks everyone for putting their input.
attached is a list of all none-yamlized needed jobs and the ones not mapped
yet,
I've ensured all needed jobs in jenkins-old are enabled, they are going to
be kept there until yamlized. as far as I know there is no technical
difficulty
in having 2 Jenkins instances listening to the same repos, only thing
that needs care before merging is when one instance might give CI '+1' and
the
other one '-1'.

still unmapped are the following jobs:
ovirt-appliance_for-testing_build-artifacts-el7-x86_64
ovirt-engine_master_build-artifacts-el6-x86_64_no_spm_testing
ovirt-engine_master_build-artifacts-el7-x86_64_no_spm_testing
ovirt-engine_master_check-merged-el7-x86_64-testing-clone-github
ovirt_integration-test-poc_created
ovirt-node_master_check-local_gerrit
ovirt-node-plugin-hosted-engine_master_create-rpms_merged
ovirt-scheduler-proxy_compile_gerrit
ovirt-setup-lib_unit-tests
qemu-kvm-rhev_create-rpms_el6
validate_open_sources_licenses_ovirt

On Tue, Apr 26, 2016 at 3:09 PM, David Caro  wrote:

> On 04/26 14:04, Sandro Bonazzola wrote:
> > On Sun, Apr 24, 2016 at 8:02 PM, Nadav Goldin 
> wrote:
> >
> > > Hey Sandro,
> > > [1] is a list of all the none-yamlized jobs in jenkins-old.ovirt.org,
> can
> > > you help us map which jobs still need to be enabled? we already mapped
> dao
> > > and find_bugs, we want to minimize the number of jobs that are not
> yamlized
> > > yet and must be enabled in jenkins-old.ovirt.org
> > >
> > >
> > > Thanks,
> > >
> > > Nadav.
> > >
> > >
> > >
> > > [1] https://paste.fedoraproject.org/359265/
> > >
> >
> >
> >
> > httpcomponents-client_master_create-rpms_merged
> > httpcomponents-core_master_create-rpms_merged
> > vhostmd_create-rpms_el6
> > vhostmd_create-rpms_el7
> > all can be dropped after publishing rpms in static repos
> >
> >
> > archive_jobs_removed_from_yaml_test_pdangur
> > mom_any_create-rpms_manual
> > otopi_any_create-rpms_manual
> > ovirt-dwh_any_create-rpms_manual
> > ovirt-engine-cli_any_create-rpms_manual
> > ovirt-engine-extension-aaa-ldap_any_create-rpms_manual
> > ovirt-engine-extension-aaa-misc_any_create-rpms_manual
> > ovirt-engine-extension-logger-log4j_any_create-rpms_manual
> > ovirt-hosted-engine-setup_any_create-rpms_manual
> > vdsm_any_create-rpms_manual
> > vdsm-jsonrpc-java_any_create-rpms_manual
> > ovirt-host-deploy_any_create-rpms_manual
> > ovirt-hosted-engine-ha_any_create-rpms_manual
> > ovirt-image-uploader_any_create-rpms_manual
> > ovirt-iso-uploader_any_create-rpms_manual
> > ovirt-live_3.6-create-iso
> > ovirt-live_master-create-iso
> > ovirt-log-collector_any_create-rpms_manual
> > ovirt-reports_any_create-rpms_manual
> > ovirt-setup-lib_any_create-rpms_manual
> > repos_3.6_check-closure_merged
> > repos_master_check-closure_merged
> > spagobi_repo_merged
> > All needed
> >
> >
> > ovirt-hosted-engine-ha_gerrit
> > ovirt-engine-jboss-as_master_create-rpms_merged
> > all can be dropped
> >
> > PatchMate-commit-hook
> > ovirt-appliance_for-testing_build-artifacts-el7-x86_64
> > ovirt-engine_3.6_dao-unit-tests_created
> > ovirt-engine_3.6_dao-unit-tests_merged
> > ovirt-engine_3.6_find-bugs_gerrit
> > ovirt-engine_3.6_find-bugs_merged
> > ovirt-engine_3.6_style_gerrit
> > ovirt-engine_3.6_unit-tests_gerrit
> > ovirt-engine_3.6_unit-tests_merged
> > ovirt-engine_master_build-artifacts-el6-x86_64_no_spm_testing
> > ovirt-engine_master_build-artifacts-el7-x86_64_no_spm_testing
> > ovirt-engine_master_check-merged-el7-x86_64-testing-clone-github
> > ovirt-engine_master_coverity-analysis_merged
> > ovirt-engine_master_dao-unit-tests_created
> > ovirt-engine_master_dao-unit-tests_merged
> > ovirt-engine_master_find-bugs_gerrit
> > ovirt-engine_master_find-bugs_gerrit_juan
> > ovirt-engine_master_find-bugs_merged
> > ovirt-engine_master_style_gerrit
> > ovirt-engine_master_unit-tests_gerrit
> > ovirt-engine_master_unit-tests_merged
> > ovirt-node-plugin-hosted-engine_master_create-rpms_merged
> > ovirt-node_master_check-local_gerrit
> > ovirt-optimizer_master_create-rpms_gerrit
> > ovirt-scheduler-proxy_compile_gerrit
> > ovirt-setup-lib_unit-tests
> > ovirt-vmconsole_any_create-rpms_manual
> > ovirt-vmconsole_master_create-rpms-el6-x86_64_created
> > ovirt-vmconsole_master_create-rpms-el6-x86_64_merged
> > ovirt_3.6_image-system-tests
> > ovirt_integration-test-poc_created
> > qemu-kvm-rh

Re: Change in ovirt-engine[master]: engine: Externalized default value in policy

2016-05-01 Thread Nadav Goldin
Sure, I enabled them again 2 days ago as thought they were needed.
so all of the above are not needed also right?
ovirt-engine_3.6_dao-unit-tests_created
ovirt-engine_3.6_dao-unit-tests_merged
ovirt-engine_master_dao-unit-tests_created
ovirt-engine_master_dao-unit-tests_merged
ovirt-engine_master_unit-tests_gerrit
ovirt-engine_master_unit-tests_merged
ovirt-engine_3.6_unit-tests_gerrit
ovirt-engine_3.6_unit-tests_merged
ovirt-setup-lib_unit-tests


On Sun, May 1, 2016 at 12:03 PM, Eyal Edri  wrote:

> Nadav,
> can you disable/remove this job?
>
> All unit tests were moved to standard CI for ovirt-engine.
>
> e.
>
> On Sun, May 1, 2016 at 11:27 AM, Tomer Saban  wrote:
>
>> Hi,
>>
>> Seems to be a problem with the CI tests. Which are not related to the
>> patch. See the following stacktrace:
>> """
>> 09:23:38 [ERROR] COMPILATION ERROR :
>> 09:23:38 [INFO]
>> -
>> 09:23:38 [ERROR]
>> /home/jenkins/workspace/ovirt-engine_master_unit-tests_gerrit/ovirt-engine/backend/manager/modules/vdsbroker/src/main/java/org/ovirt/engine/core/vdsbroker/jsonrpc/JsonRpcVdsServer.java:[471,41]
>> error: cannot find symbol
>> 09:23:38 [ERROR]  variable client of type JsonRpcClient
>> 09:23:38
>> /home/jenkins/workspace/ovirt-engine_master_unit-tests_gerrit/ovirt-engine/backend/manager/modules/vdsbroker/src/main/java/org/ovirt/engine/core/vdsbroker/jsonrpc/JsonRpcVdsServer.java:[474,47]
>> error: cannot find symbol
>> 09:23:38 [INFO] 2 errors
>> """
>>
>> Thanks,
>> Tomer
>>
>> - Forwarded Message -
>> From: "Jenkins CI" 
>> To: "Tomer Saban" , "Roy Golan" 
>> Sent: Sunday, May 1, 2016 11:24:00 AM
>> Subject: Change in ovirt-engine[master]: engine: Externalized default
>> value in policy
>>
>> Jenkins CI has posted comments on this change.
>>
>> Change subject: engine: Externalized default value in policy
>> .
>>
>>
>> Patch Set 8: Continuous-Integration-1
>>
>> Build Failed
>>
>>
>> http://jenkins-old.ovirt.org/job/ovirt-engine_master_unit-tests_gerrit/57498/
>> : FAILURE
>>
>> --
>> To view, visit https://gerrit.ovirt.org/38875
>> To unsubscribe, visit https://gerrit.ovirt.org/settings
>>
>> Gerrit-MessageType: comment
>> Gerrit-Change-Id: I244ac0cd9cb41833c7a5357264ddca99e489de94
>> Gerrit-PatchSet: 8
>> Gerrit-Project: ovirt-engine
>> Gerrit-Branch: master
>> Gerrit-Owner: Tomer Saban 
>> Gerrit-Reviewer: Jenkins CI
>> Gerrit-Reviewer: Martin Sivák 
>> Gerrit-Reviewer: Roy Golan 
>> Gerrit-Reviewer: Tomer Saban 
>> Gerrit-Reviewer: gerrit-hooks 
>> Gerrit-HasComments: No
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>>
>
>
> --
> Eyal Edri
> Associate Manager
> RHEV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Change in ovirt-engine[master]: engine: Externalized default value in policy

2016-05-01 Thread Nadav Goldin
done


On Sun, May 1, 2016 at 12:28 PM, Eyal Edri  wrote:

> Please keep dao tests for now.
> On May 1, 2016 12:16 PM, "Nadav Goldin"  wrote:
>
>> Sure, I enabled them again 2 days ago as thought they were needed.
>> so all of the above are not needed also right?
>> ovirt-engine_3.6_dao-unit-tests_created
>> ovirt-engine_3.6_dao-unit-tests_merged
>> ovirt-engine_master_dao-unit-tests_created
>> ovirt-engine_master_dao-unit-tests_merged
>> ovirt-engine_master_unit-tests_gerrit
>> ovirt-engine_master_unit-tests_merged
>> ovirt-engine_3.6_unit-tests_gerrit
>> ovirt-engine_3.6_unit-tests_merged
>> ovirt-setup-lib_unit-tests
>>
>>
>> On Sun, May 1, 2016 at 12:03 PM, Eyal Edri  wrote:
>>
>>> Nadav,
>>> can you disable/remove this job?
>>>
>>> All unit tests were moved to standard CI for ovirt-engine.
>>>
>>> e.
>>>
>>> On Sun, May 1, 2016 at 11:27 AM, Tomer Saban  wrote:
>>>
>>>> Hi,
>>>>
>>>> Seems to be a problem with the CI tests. Which are not related to the
>>>> patch. See the following stacktrace:
>>>> """
>>>> 09:23:38 [ERROR] COMPILATION ERROR :
>>>> 09:23:38 [INFO]
>>>> -
>>>> 09:23:38 [ERROR]
>>>> /home/jenkins/workspace/ovirt-engine_master_unit-tests_gerrit/ovirt-engine/backend/manager/modules/vdsbroker/src/main/java/org/ovirt/engine/core/vdsbroker/jsonrpc/JsonRpcVdsServer.java:[471,41]
>>>> error: cannot find symbol
>>>> 09:23:38 [ERROR]  variable client of type JsonRpcClient
>>>> 09:23:38
>>>> /home/jenkins/workspace/ovirt-engine_master_unit-tests_gerrit/ovirt-engine/backend/manager/modules/vdsbroker/src/main/java/org/ovirt/engine/core/vdsbroker/jsonrpc/JsonRpcVdsServer.java:[474,47]
>>>> error: cannot find symbol
>>>> 09:23:38 [INFO] 2 errors
>>>> """
>>>>
>>>> Thanks,
>>>> Tomer
>>>>
>>>> - Forwarded Message -
>>>> From: "Jenkins CI" 
>>>> To: "Tomer Saban" , "Roy Golan" 
>>>> Sent: Sunday, May 1, 2016 11:24:00 AM
>>>> Subject: Change in ovirt-engine[master]: engine: Externalized default
>>>> value in policy
>>>>
>>>> Jenkins CI has posted comments on this change.
>>>>
>>>> Change subject: engine: Externalized default value in policy
>>>> .
>>>>
>>>>
>>>> Patch Set 8: Continuous-Integration-1
>>>>
>>>> Build Failed
>>>>
>>>>
>>>> http://jenkins-old.ovirt.org/job/ovirt-engine_master_unit-tests_gerrit/57498/
>>>> : FAILURE
>>>>
>>>> --
>>>> To view, visit https://gerrit.ovirt.org/38875
>>>> To unsubscribe, visit https://gerrit.ovirt.org/settings
>>>>
>>>> Gerrit-MessageType: comment
>>>> Gerrit-Change-Id: I244ac0cd9cb41833c7a5357264ddca99e489de94
>>>> Gerrit-PatchSet: 8
>>>> Gerrit-Project: ovirt-engine
>>>> Gerrit-Branch: master
>>>> Gerrit-Owner: Tomer Saban 
>>>> Gerrit-Reviewer: Jenkins CI
>>>> Gerrit-Reviewer: Martin Sivák 
>>>> Gerrit-Reviewer: Roy Golan 
>>>> Gerrit-Reviewer: Tomer Saban 
>>>> Gerrit-Reviewer: gerrit-hooks 
>>>> Gerrit-HasComments: No
>>>> ___
>>>> Infra mailing list
>>>> Infra@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Eyal Edri
>>> Associate Manager
>>> RHEV DevOps
>>> EMEA ENG Virtualization R&D
>>> Red Hat Israel
>>>
>>> phone: +972-9-7692018
>>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>>>
>>
>>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-users] [ANN] oVirt 3.6.5 Final Release is now available

2016-05-09 Thread Nadav Goldin
Hey,
we are working on adding mirror site health check. In the mean while as far
as I can tell the only site
which is out of sync is indeed mirror.nluug.nl, it will be removed from the
mirror file until its synced again(overall there are 9 sites).

Nadav



On Mon, May 9, 2016 at 4:04 PM, Nicolas Ecarnot  wrote:

> Le 26/04/2016 16:22, Gianluca Cecchi a écrit :
>
>> On Tue, Apr 26, 2016 at 11:32 AM, Sandro Bonazzola wrote:
>>
>>
>>
>>
>> please try again, maybe mirrors are out of sync.
>> try commenting mirrorlist and enabling baseurl for the ovirt repos.
>> Thanks,
>>
>>
>> Yes, it worked this way, as previously confirmed in a private message
>> also from Gary Lloyd (thanks!)
>> It was just to notice this misalignment: a new installing user (nearby
>> the broken mirror) will get 3.6.3 and not 3.6.5 and neither 3.6.4, as
>> the reported mirror site missed that too (3.6.4 released on late March)
>> and is not aligned since more than one month now...
>>
>> HIH,
>> Gianluca
>>
>>
>> ___
>> Users mailing list
>> us...@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
> Hello oVirt world,
>
> Please note that the workaround described above is working, and please
> note that without it, no upgrade is returned (staying in 3.6.3).
> Though not knowing who's in charge of managing the repositories mirroring,
> it may be worth having a look.
>
> Have a nice day.
>
> --
> Nicolas ECARNOT
>
> ___
> Users mailing list
> us...@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: restricting check patch/merged parallel jobs

2016-05-11 Thread Nadav Goldin
I enhanced this[1] graph to compare slaves utilization vs build queue, note
that
the slaves utilization is measured in percentages and the number of builds
in the queue
is absolute. basically when the red lines are high(large queue size) and
the green ones(slaves
utilization) are low we could have possibly had more builds running. we can
see in the past few days
we've reached nice utilization of ~ 90% and following that the queue size
decreased pretty
quickly, on the other hand there were times of only 16% utilization and a
large queue ~ 70.
last I checked the least significant problem is the OS as most standard-ci
jobs
are agnostic to EL/FC, usually it was the jobs limit, or sudden peeks in
patches sent
but I didn't get to add 'reason each job is waiting' metric yet, so its
just a feeling.

maybe the Priority Sorter Plugin[2] which comes bundled with Jenkins
could address the problem of jobs waiting 'unfairly' a long time in the
queue,
though it will require to define the priorities in the yamls.



[1]
http://graphite.phx.ovirt.org/dashboard/db/jenkins-monitoring?panelId=16&fullscreen&from=146265480&to=1462966158602&var-average_interval=12h&var-filtered_labels=All&var-filtered_jobs_labels=All
[2] https://wiki.jenkins-ci.org/display/JENKINS/Priority+Sorter+Plugin

On Wed, May 11, 2016 at 1:43 PM, Sandro Bonazzola 
wrote:

>
>
> On Wed, May 11, 2016 at 12:34 PM, Eyal Edri  wrote:
>
>> From what I saw, it was mostly ovirt-engine and vdsm jobs pending on the
>> queue while other slaves are idle.
>> we have over 40 slaves and we're about to add more, so I don't think that
>> will be an issue and IMO 3 per job is not enough, especially if you get
>> idle slaves.
>>
>>
> +1 on raising then.
>
>
>
>> We are thinking on a more dynamic approach of dynamic vm allocation on
>> demand, so in the long run we'll have more control over it,
>> for now i'm monitoring the queue size and slaves on a regular basis [1],
>> so if anything will get blocked too much time we'll act and adjust
>> accordingly.
>>
>>
>> [1] http://graphite.phx.ovirt.org/dashboard/db/jenkins-monitoring
>>
>> On Wed, May 11, 2016 at 1:10 PM, Sandro Bonazzola 
>> wrote:
>>
>>>
>>>
>>> On Tue, May 10, 2016 at 1:01 PM, Eyal Edri  wrote:
>>>
 Shlomi,
 Can you submit a patch to increase the limit to 6 for (i think all jobs
 are using the same yaml template) and we'll continue to monitor to queue
 and see if there is an improvement in the utilization of slaves?

>>>
>>> Issue was that long lasting jobs caused queue to increase too much.
>>> Example: a patch set rebased on master and merged will cause triggering
>>> of check-merged jobs, upgrade jobs, ...; running 6 instance of each of them
>>> will cause all other projects to be queued for a lot of time.
>>>
>>>
>>>

 E.

 On Tue, May 10, 2016 at 1:58 PM, David Caro  wrote:

> On 05/10 13:54, Eyal Edri wrote:
> > Is there any reason we're limiting the amount of check patch & check
> merged
> > jobs to run only 3 in parallel?
> >
>
> We had some mess in the past where enabling parallel runs did not
> really force
> not using the same slave at the same time, I guess we never reenabled
> them.
>
> > Each jobs runs in mock and on its own VM, anything presenting us from
> > removing this limitation so we won't have idle slaves while other
> jobs are
> > in the queue?
> >
> > We can increase it at least to a higher level if we won't one
> specific job
> > to take over all slaves and starve other jobs, but i think
> ovirt-engine
> > jobs are probably the biggest consumer of ci, so the threshold
> should be
> > updated.
>
> +1
>
> >
> > --
> > Eyal Edri
> > Associate Manager
> > RHEV DevOps
> > EMEA ENG Virtualization R&D
> > Red Hat Israel
> >
> > phone: +972-9-7692018
> > irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
>
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>



 --
 Eyal Edri
 Associate Manager
 RHEV DevOps
 EMEA ENG Virtualization R&D
 Red Hat Israel

 phone: +972-9-7692018
 irc: eedri (on #tlv #rhev-dev #rhev-integ)

 ___
 Infra mailing list
 Infra@ovirt.org
 http://lists.ovirt.org/mailman/listinfo/infra


>>>
>>>
>>> --
>>> Sandro Bonazzola
>>> Better technology. Faster innovation. Powered by community collaboration.
>>> See how it works at redhat.com
>>>
>>
>>
>>
>> --
>> Eyal Edri
>> As

ovirt-srv11

2016-05-11 Thread Nadav Goldin
Hi,
ovirt-srv11 host is in an empty cluster called 'Production_CentOS', its
quite a strong machine with 251GB of ram, currently it has no VMs and as
far as I can tell isn't used at all.
I want to move it to the 'Jenkins_CentOS' cluster in order to add more VMs
and later upgrade the older clusters to el7(if we have enough slaves in the
Jenkins_CentOS cluster, we could just take the VMs down in the Jenkins
cluster and upgrade). this is unrelated to the new hosts ovirt-srv17-26.

I'm not sure why it was put there, so posting here if anyone objects or I'm
missing something


Thanks

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: restricting check patch/merged parallel jobs

2016-05-11 Thread Nadav Goldin
On Wed, May 11, 2016 at 4:19 PM, David Caro  wrote:

> You are not having into account yet that we are not using most of the
> hardware
> we have yet, that will allow us to have more than twice the amount of
> slaves we
> have now too (given that we solve any other bottleneck, like nfs storage)
>

I agree. I think best thing to begin with would be to add as many slaves
as we can and remove all restrictions from the jobs and monitor the queue.

> Last time we tried using it, it ended messing up all the executions, mixing
> slaves and creating a lot of failures, if you try it out, be very vigilant

good to know
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: capacity for more slaves

2016-05-12 Thread Nadav Goldin
Hi, I added few more el7 slaves, there is a (relatively) new template
'centos72-jenkins-slave' in the Jenkins_CentOS cluster.
iirc there was a discussion few months ago about our IPs limit, I think
we're approaching that(there are
102 fixed addresses in foreman.phx.ovirt.org dhcpd server, and this is
without the old ovirt-srv* which are not managed there)


On Tue, May 10, 2016 at 7:04 PM, David Caro  wrote:

> On 05/10 18:02, David Caro wrote:
> > On 05/10 18:53, Eyal Edri wrote:
> > > Looking at the load on our hypervisors I'm sure we can add more slaves
> to
> > > jenkins.
> > > Is there a documented procedure on how to add a new slave so anyone in
> the
> > > team can do it?
> > >
> >
> > I remember writing something about the templates and such, will look for
> it.
> >
> > Though it might be a bit old, not sure if we have changed anything in
> the last
> > months, maybe someone has newer info.
>
>
> There's some docs here:
>
>
> http://ovirt-infra-docs.readthedocs.io/en/latest/Phoenix_Lab/oVirt_Hosts.html#ovirt-datacenter-organization
>
> >
> > >
> > >
> > >
> > >
> > > --
> > > Eyal Edri
> > > Associate Manager
> > > RHEV DevOps
> > > EMEA ENG Virtualization R&D
> > > Red Hat Israel
> > >
> > > phone: +972-9-7692018
> > > irc: eedri (on #tlv #rhev-dev #rhev-integ)
> >
> > > ___
> > > Infra mailing list
> > > Infra@ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/infra
> >
> >
> > --
> > David Caro
> >
> > Red Hat S.L.
> > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> >
> > Tel.: +420 532 294 605
> > Email: dc...@redhat.com
> > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > Web: www.redhat.com
> > RHT Global #: 82-62605
>
>
>
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
>
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: capacity for more slaves

2016-05-12 Thread Nadav Goldin
ok I'll add also FC23, though I think the problem is the vdsm jobs, not sure
why they use only fc* instead of the el7 ones

On Thu, May 12, 2016 at 4:33 PM, David Caro  wrote:

> On 05/12 12:46, Nadav Goldin wrote:
> > Hi, I added few more el7 slaves, there is a (relatively) new template
> > 'centos72-jenkins-slave' in the Jenkins_CentOS cluster.
> > iirc there was a discussion few months ago about our IPs limit, I think
> > we're approaching that(there are
> > 102 fixed addresses in foreman.phx.ovirt.org dhcpd server, and this is
> > without the old ovirt-srv* which are not managed there)
>
> Actually, we are scarce on fc23 slaves, not el (right now we have 28 idle
> el7
> slaves, and I've been waiting for more than 45min for a lago check-patch to
> start on a fc23 slave)
>
> >
> >
> > On Tue, May 10, 2016 at 7:04 PM, David Caro  wrote:
> >
> > > On 05/10 18:02, David Caro wrote:
> > > > On 05/10 18:53, Eyal Edri wrote:
> > > > > Looking at the load on our hypervisors I'm sure we can add more
> slaves
> > > to
> > > > > jenkins.
> > > > > Is there a documented procedure on how to add a new slave so
> anyone in
> > > the
> > > > > team can do it?
> > > > >
> > > >
> > > > I remember writing something about the templates and such, will look
> for
> > > it.
> > > >
> > > > Though it might be a bit old, not sure if we have changed anything in
> > > the last
> > > > months, maybe someone has newer info.
> > >
> > >
> > > There's some docs here:
> > >
> > >
> > >
> http://ovirt-infra-docs.readthedocs.io/en/latest/Phoenix_Lab/oVirt_Hosts.html#ovirt-datacenter-organization
> > >
> > > >
> > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > Eyal Edri
> > > > > Associate Manager
> > > > > RHEV DevOps
> > > > > EMEA ENG Virtualization R&D
> > > > > Red Hat Israel
> > > > >
> > > > > phone: +972-9-7692018
> > > > > irc: eedri (on #tlv #rhev-dev #rhev-integ)
> > > >
> > > > > ___
> > > > > Infra mailing list
> > > > > Infra@ovirt.org
> > > > > http://lists.ovirt.org/mailman/listinfo/infra
> > > >
> > > >
> > > > --
> > > > David Caro
> > > >
> > > > Red Hat S.L.
> > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > > >
> > > > Tel.: +420 532 294 605
> > > > Email: dc...@redhat.com
> > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > > Web: www.redhat.com
> > > > RHT Global #: 82-62605
> > >
> > >
> > >
> > > > ___
> > > > Infra mailing list
> > > > Infra@ovirt.org
> > > > http://lists.ovirt.org/mailman/listinfo/infra
> > >
> > >
> > > --
> > > David Caro
> > >
> > > Red Hat S.L.
> > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > >
> > > Tel.: +420 532 294 605
> > > Email: dc...@redhat.com
> > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > Web: www.redhat.com
> > > RHT Global #: 82-62605
> > >
> > > ___
> > > Infra mailing list
> > > Infra@ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/infra
> > >
> > >
>
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
>
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: capacity for more slaves

2016-05-12 Thread Nadav Goldin
sure, added 4 more FC23 slaves, and on the way hit the IPs limit,
had to remove some old unused testing VMs to release IPs.
we have some useless leases there such as an IP for each template,
though I'm not 100% sure if its safe to simply delete them from foreman
(they are saved as templates in the engine, so theoretically shouldn't
be a problem)




On Thu, May 12, 2016 at 4:53 PM, David Caro  wrote:

> On 05/12 16:50, Nadav Goldin wrote:
> > ok I'll add also FC23, though I think the problem is the vdsm jobs, not
> sure
> > why they use only fc* instead of the el7 ones
>
> Thanks man, that's because (at least the check-merged) use libvirt to
> start up
> vms, and trying to use libvirt from a fc23 chroot on an el7 host ends up in
> error, so they have to run on fc23/21 slaves
>
> >
> > On Thu, May 12, 2016 at 4:33 PM, David Caro  wrote:
> >
> > > On 05/12 12:46, Nadav Goldin wrote:
> > > > Hi, I added few more el7 slaves, there is a (relatively) new template
> > > > 'centos72-jenkins-slave' in the Jenkins_CentOS cluster.
> > > > iirc there was a discussion few months ago about our IPs limit, I
> think
> > > > we're approaching that(there are
> > > > 102 fixed addresses in foreman.phx.ovirt.org dhcpd server, and this
> is
> > > > without the old ovirt-srv* which are not managed there)
> > >
> > > Actually, we are scarce on fc23 slaves, not el (right now we have 28
> idle
> > > el7
> > > slaves, and I've been waiting for more than 45min for a lago
> check-patch to
> > > start on a fc23 slave)
> > >
> > > >
> > > >
> > > > On Tue, May 10, 2016 at 7:04 PM, David Caro 
> wrote:
> > > >
> > > > > On 05/10 18:02, David Caro wrote:
> > > > > > On 05/10 18:53, Eyal Edri wrote:
> > > > > > > Looking at the load on our hypervisors I'm sure we can add more
> > > slaves
> > > > > to
> > > > > > > jenkins.
> > > > > > > Is there a documented procedure on how to add a new slave so
> > > anyone in
> > > > > the
> > > > > > > team can do it?
> > > > > > >
> > > > > >
> > > > > > I remember writing something about the templates and such, will
> look
> > > for
> > > > > it.
> > > > > >
> > > > > > Though it might be a bit old, not sure if we have changed
> anything in
> > > > > the last
> > > > > > months, maybe someone has newer info.
> > > > >
> > > > >
> > > > > There's some docs here:
> > > > >
> > > > >
> > > > >
> > >
> http://ovirt-infra-docs.readthedocs.io/en/latest/Phoenix_Lab/oVirt_Hosts.html#ovirt-datacenter-organization
> > > > >
> > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > --
> > > > > > > Eyal Edri
> > > > > > > Associate Manager
> > > > > > > RHEV DevOps
> > > > > > > EMEA ENG Virtualization R&D
> > > > > > > Red Hat Israel
> > > > > > >
> > > > > > > phone: +972-9-7692018
> > > > > > > irc: eedri (on #tlv #rhev-dev #rhev-integ)
> > > > > >
> > > > > > > ___
> > > > > > > Infra mailing list
> > > > > > > Infra@ovirt.org
> > > > > > > http://lists.ovirt.org/mailman/listinfo/infra
> > > > > >
> > > > > >
> > > > > > --
> > > > > > David Caro
> > > > > >
> > > > > > Red Hat S.L.
> > > > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > > > > >
> > > > > > Tel.: +420 532 294 605
> > > > > > Email: dc...@redhat.com
> > > > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > > > > Web: www.redhat.com
> > > > > > RHT Global #: 82-62605
> > > > >
> > > > >
> > > > >
> > > > > > ___
> > > > > > Infra mailing list
> > > > > > Infra@ovirt.org
> > > > > > http://lists.ovirt.org/mailman/listinfo/infra
> > > > >
> > > > >
> > > > > --
> > > > > David Caro
> > > > >
> > > > > Red Hat S.L.
> > > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > > > >
> > > > > Tel.: +420 532 294 605
> > > > > Email: dc...@redhat.com
> > > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > > > Web: www.redhat.com
> > > > > RHT Global #: 82-62605
> > > > >
> > > > > ___
> > > > > Infra mailing list
> > > > > Infra@ovirt.org
> > > > > http://lists.ovirt.org/mailman/listinfo/infra
> > > > >
> > > > >
> > >
> > > > ___
> > > > Infra mailing list
> > > > Infra@ovirt.org
> > > > http://lists.ovirt.org/mailman/listinfo/infra
> > >
> > >
> > > --
> > > David Caro
> > >
> > > Red Hat S.L.
> > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > >
> > > Tel.: +420 532 294 605
> > > Email: dc...@redhat.com
> > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > Web: www.redhat.com
> > > RHT Global #: 82-62605
> > >
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ovirt-srv11

2016-05-12 Thread Nadav Goldin
>
> If we need that extra server in Production DC (for hosted engine
> redundancy and to allow maintenance) then lets take the lower end new
> servers from 17-26 and replace it with the strong one.
> We need to utilize our servers, I don't think we're at 50% utilization
> even, looking at the memory consumption last time i checked when all slaves
> were working
>
the problem is that we won't have live migration in the Production cluster
if we add
the new hosts there(because of the multi node NUMA mismatch), we could put
2 new
hosts in the Production cluster but then we would have to schedule a
downtime
in order to move them.

I think that coupling the engine upgrade with the new slaves is not
necessary.
we can start by installing the hook on ovirt-srv11 and spinning up new
slaves,
that way we can also test the hook in production. because there is no NFS
involved,
and live migration isn't working with the hook, the host is pretty much
self-contained
and this can't harm anything.


On Thu, May 12, 2016 at 4:59 PM, Anton Marchukov 
wrote:

> I was talking about using the hooks before installing the ssds, but if
>> that can
>> be done reliably also before the upgrade it's also a solution that will
>> help
>> scratch our current itches sooner.
>>
>
> That's about it. The hook in 3.6 contains the patch I submitted. It does
> not work without it at all.
> Although you can use rpm from 3.5 in 3.6.
>
>
>> ^ well, the machines being slaves I think there's no problem having to
>> stop
>> them to migrate them, we have already experience with both fabric and
>> jenkins
>> so I guess it should not be hard to automate with those tools ;)
>>
>
> Yes there is no problem with that. But I am not aware about "stop all
> slaves on the host" feature in
> oVirt so that would be either manually clicking or we need to fabricate
> it. Not a big deal too.
>
> --
> Anton Marchukov
> Senior Software Engineer - RHEV CI - Red Hat
>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Maintenance and reboot of ovirt-srv17

2016-05-16 Thread Nadav Goldin
Hi David,
I also had similar problem last week that some el7 vms didn't have nested
enabled
although all configurations were OK. Restarting helped to some and for
others I had
to manually reload the module which solved it without rebooting.


On Tue, May 17, 2016 at 12:47 AM, David Caro  wrote:

> On 05/16 23:39, David Caro wrote:
> > On 05/16 23:36, David Caro wrote:
> > > On 05/16 23:07, David Caro wrote:
> > > >
> > > > Hey, just everyone is on the same page, I'm putting into maintenance
> and will
> > > > reboot ovirt-srv17 to enable nested vms on it (the config is there,
> but the
> > > > host was not restarted and thus the params were not applied to the
> module)
> > >
> > > Finished with ovirt-srv17, now with ovirt-srv01
> >
> >
> > Sorry, srv01 has nested enabled... debugging
>
> It was not 01, but ovirt-srv14 the host that had issues, and the issue was
> not
> that nested was not enabled, but for some reason, the vm did not pick it up
> (maybe it was migrated from a non-nested host) and had to restart the vm to
> refresh the hardware capabilities.
>
> > >
> > > >
> > > >
> > > > --
> > > > David Caro
> > > >
> > > > Red Hat S.L.
> > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > > >
> > > > Tel.: +420 532 294 605
> > > > Email: dc...@redhat.com
> > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > > Web: www.redhat.com
> > > > RHT Global #: 82-62605
> > >
> > >
> > >
> > > --
> > > David Caro
> > >
> > > Red Hat S.L.
> > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> > >
> > > Tel.: +420 532 294 605
> > > Email: dc...@redhat.com
> > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > > Web: www.redhat.com
> > > RHT Global #: 82-62605
> >
> >
> >
> > > ___
> > > Infra mailing list
> > > Infra@ovirt.org
> > > http://lists.ovirt.org/mailman/listinfo/infra
> >
> >
> > --
> > David Caro
> >
> > Red Hat S.L.
> > Continuous Integration Engineer - EMEA ENG Virtualization R&D
> >
> > Tel.: +420 532 294 605
> > Email: dc...@redhat.com
> > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> > Web: www.redhat.com
> > RHT Global #: 82-62605
>
>
>
> > ___
> > Infra mailing list
> > Infra@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/infra
>
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


ovirt-optimizer dependencies on Jenkins slaves

2016-05-22 Thread Nadav Goldin
Hi Martin,
I want to confirm with you that ovirt-optimizer was moved to
standard-ci, as we still have some puppet code that installs its
dependencies on all slaves(the only package is 'symlinks')
is it ok to remove it?(should not affect anything that is running under mock)




Thanks,
Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [oVirt Jenkins] ovirt-engine_master_upgrade-from-master_el7_merged - Build # 431 - Failure!

2016-05-29 Thread Nadav Goldin
this seems to be caused by:

[INFO] Compiling module org.ovirt.engine.ui.webadmin.WebAdmin
[INFO]Validating units:
[INFO]   [ERROR] Errors in
'jar:file:/home/jenkins/workspace/ovirt-engine_master_upgrade-from-master_el7_merged/tmp_repo/ovirt-engine-4.1.0/backend/manager/modules/common/target/common-4.1.0-SNAPSHOT-sources.jar!/org/ovirt/engine/core/common/action/VdsOperationActionParameters.java'
[INFO]  [ERROR] Line 23: No source code is available for type
org.ovirt.engine.core.common.businessentities.HostedEngineDeployConfiguration;
did you forget to inherit a required module?
[INFO]   [WARN] Warnings in
'jar:file:/home/jenkins/workspace/ovirt-engine_master_upgrade-from-master_el7_merged/tmp_repo/ovirt-engine-4.1.0/frontend/webadmin/modules/gwt-common/target/gwt-common-4.1.0-SNAPSHOT.jar!/org/ovirt/engine/ui/common/uicommon/SpicePluginImpl.java'

(thought it was caused by the recent puppet upgrade so searched the logs)


On Sun, May 29, 2016 at 10:44 PM,   wrote:
> Project: 
> http://jenkins.ovirt.org/job/ovirt-engine_master_upgrade-from-master_el7_merged/
> Build: 
> http://jenkins.ovirt.org/job/ovirt-engine_master_upgrade-from-master_el7_merged/431/
> Build Number: 431
> Build Status:  Failure
> Triggered By: Triggered by Gerrit: https://gerrit.ovirt.org/43562
>
> -
> Changes Since Last Success:
> -
> Changes for Build #431
> [Eyal Edri] fixing 4.0 publisher to push to 4.0 and not master
>
> [Roy Golan] host-deploy: Support deploying Hosted Engine components
>
>
>
>
> -
> Failed Tests:
> -
> No tests ran.
>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


fc23 jenkins slaves

2016-05-29 Thread Nadav Goldin
Hi all,
I merged [1](and a 'hotfix' [2]) which basically allows to run our
puppet code on fc23 and cleans up some packages we had. This caused a
restart of postgresql on all slaves, so far I did not catch any jobs
failing due to that.

the second implication is that we have zero slaves in 'error
state'[3]**, so we have a 'fresh ground' and can monitor from now on
if new puppet errors pop up.

Thanks,

Nadav.


[1] https://gerrit.ovirt.org/#/c/57847/
[2] https://gerrit.ovirt.org/#/c/58229/
[3] https://foreman.ovirt.org/dashboard
** to the time of writing :)
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: fc23 jenkins slaves

2016-05-30 Thread Nadav Goldin
Thanks :)
I'll open another ticket on that, I'll re-enable it only for 'puppet'
errors and not for unsynchronized hosts, as this includes also VMs
that were shutdown(and was causing most of the alerts flood last time)


On Mon, May 30, 2016 at 10:41 AM, Eyal Edri  wrote:
> Kudos for all the work Nadav, seems quite a rework and improvement long
> overdue :)
> Did you re-enable the alerts on errors in VMs in icigna?
>
> /e
>
> On Mon, May 30, 2016 at 9:23 AM, Barak Korren  wrote:
>>
>> >
>> > the second implication is that we have zero slaves in 'error
>> > state'[3]**, so we have a 'fresh ground' and can monitor from now on
>> > if new puppet errors pop up.
>> >
>> Great work!
>>
>>
>> --
>> Barak Korren
>> bkor...@redhat.com
>> RHEV-CI Team
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>
>
>
> --
> Eyal Edri
> Associate Manager
> RHEV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Maintenance and reboot of ovirt-srv17

2016-05-31 Thread Nadav Goldin
bumping this old thread:
after [1] was merged we now get alerts for VMs that don't have nested
enabled through puppet:
/Stage[main]/Ovirt_jenkins_slave::Base/Kmod::Load[kvm_intel]/Exec[modprobe
kvm_intel]/returns change from notrun to 0 failed: modprobe kvm_intel
returned 1 instead of one of [0]

assuming the hypervisor is configured properly(no puppet there yet) a
shutdown/run of the VM from the engine resolves the issue(got few of
those today)

[1] https://gerrit.ovirt.org/#/c/58338/

On Tue, May 17, 2016 at 12:23 PM, David Caro  wrote:
> On 05/17 09:59, Nadav Goldin wrote:
>> Hi David,
>> I also had similar problem last week that some el7 vms didn't have nested
>> enabled
>> although all configurations were OK. Restarting helped to some and for
>> others I had
>> to manually reload the module which solved it without rebooting.
>
> I prefer rebooting to make sure that it will not go away when we reboot
> (manually reloading the module will not persist on reboot).
> If it does not work by itself after reboot, is an issue and should be fixed 
> too
> (probably config related).
>
>>
>>
>> On Tue, May 17, 2016 at 12:47 AM, David Caro  wrote:
>>
>> > On 05/16 23:39, David Caro wrote:
>> > > On 05/16 23:36, David Caro wrote:
>> > > > On 05/16 23:07, David Caro wrote:
>> > > > >
>> > > > > Hey, just everyone is on the same page, I'm putting into maintenance
>> > and will
>> > > > > reboot ovirt-srv17 to enable nested vms on it (the config is there,
>> > but the
>> > > > > host was not restarted and thus the params were not applied to the
>> > module)
>> > > >
>> > > > Finished with ovirt-srv17, now with ovirt-srv01
>> > >
>> > >
>> > > Sorry, srv01 has nested enabled... debugging
>> >
>> > It was not 01, but ovirt-srv14 the host that had issues, and the issue was
>> > not
>> > that nested was not enabled, but for some reason, the vm did not pick it up
>> > (maybe it was migrated from a non-nested host) and had to restart the vm to
>> > refresh the hardware capabilities.
>> >
>> > > >
>> > > > >
>> > > > >
>> > > > > --
>> > > > > David Caro
>> > > > >
>> > > > > Red Hat S.L.
>> > > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
>> > > > >
>> > > > > Tel.: +420 532 294 605
>> > > > > Email: dc...@redhat.com
>> > > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
>> > > > > Web: www.redhat.com
>> > > > > RHT Global #: 82-62605
>> > > >
>> > > >
>> > > >
>> > > > --
>> > > > David Caro
>> > > >
>> > > > Red Hat S.L.
>> > > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
>> > > >
>> > > > Tel.: +420 532 294 605
>> > > > Email: dc...@redhat.com
>> > > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
>> > > > Web: www.redhat.com
>> > > > RHT Global #: 82-62605
>> > >
>> > >
>> > >
>> > > > ___
>> > > > Infra mailing list
>> > > > Infra@ovirt.org
>> > > > http://lists.ovirt.org/mailman/listinfo/infra
>> > >
>> > >
>> > > --
>> > > David Caro
>> > >
>> > > Red Hat S.L.
>> > > Continuous Integration Engineer - EMEA ENG Virtualization R&D
>> > >
>> > > Tel.: +420 532 294 605
>> > > Email: dc...@redhat.com
>> > > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
>> > > Web: www.redhat.com
>> > > RHT Global #: 82-62605
>> >
>> >
>> >
>> > > ___
>> > > Infra mailing list
>> > > Infra@ovirt.org
>> > > http://lists.ovirt.org/mailman/listinfo/infra
>> >
>> >
>> > --
>> > David Caro
>> >
>> > Red Hat S.L.
>> > Continuous Integration Engineer - EMEA ENG Virtualization R&D
>> >
>> > Tel.: +420 532 294 605
>> > Email: dc...@redhat.com
>> > IRC: dcaro|dcaroest@{freenode|oftc|redhat}
>> > Web: www.redhat.com
>> > RHT Global #: 82-62605
>> >
>> > ___
>> > Infra mailing list
>> > Infra@ovirt.org
>> > http://lists.ovirt.org/mailman/listinfo/infra
>> >
>> >
>
> --
> David Caro
>
> Red Hat S.L.
> Continuous Integration Engineer - EMEA ENG Virtualization R&D
>
> Tel.: +420 532 294 605
> Email: dc...@redhat.com
> IRC: dcaro|dcaroest@{freenode|oftc|redhat}
> Web: www.redhat.com
> RHT Global #: 82-62605
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: PHX updates - hypervisors in Jenkins datacenter

2016-06-02 Thread Nadav Goldin
On Thu, Jun 2, 2016 at 12:40 PM, Evgheni Dereveanchin
 wrote:
> fc21-vm01.phx.ovirt.org
> el6-vm06.phx.ovirt.org


those are on jenkins-old.ovirt.org
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ovirt-engine_master_upgrade-from-4.0_el7_merged

2016-06-19 Thread Nadav Goldin
I'm working on it(currently there is some problem with r10k, trying to fix)


On Sun, Jun 19, 2016 at 4:42 PM, Shlomo Ben David  wrote:
> This patch updates the infra-puppet project but i couldn't find any git use
> of this project in the specified job, can you guide me please where to
> search?
>
> Best Regards,
>
> Shlomi Ben-David | Software Engineer | Red Hat ISRAEL
> Phone: +972-54-8008858
> IRC: shlomibendavid (on #rhev-integ, #rhev-dev, #rhev-ci)
>
> OPEN SOURCE - 1 4 011 && 011 4 1
>
> On Sun, Jun 19, 2016 at 3:05 PM, Eyal Edri  wrote:
>>
>> These repos are obsolete and should be replaced, Sandro already sent a
>> patch [1].
>> If you can verify it, I'll merge it.
>>
>>
>> [1] https://gerrit.ovirt.org/#/c/56821/
>>
>> On Sun, Jun 19, 2016 at 2:12 PM, Shlomo Ben David 
>> wrote:
>>>
>>> Hi,
>>>
>>> JOB: ovirt-engine_master_upgrade-from-4.0_el7_merged
>>> STATUS: Fail
>>> ERROR:
>>> http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-7/x86_64/repodata/repomd.xml:
>>> [Errno 14] HTTP Error 404 - Not Found
>>> ...
>>> ...
>>> 
>>>
>>> + [[ UPGRADE::WAIT_FOR_ENGINE == \F\I\N\I\S\H\E\D ]]
>>> + echo 'FAILED::UPGRADE::WAIT_FOR_ENGINE:: Unrecoverable failure,
>>> exitting'
>>>
>>>
>>> REASON: Trying to get repo info from non exist location:
>>> http://download.gluster.org/pub/gluster/glusterfs/LATEST/EPEL.repo/epel-7/x86_64/repodata/repomd.xml
>>>
>>> Best Regards,
>>>
>>> Shlomi Ben-David | Software Engineer | Red Hat ISRAEL
>>> Phone: +972-54-8008858
>>> IRC: shlomibendavid (on #rhev-integ, #rhev-dev, #rhev-ci)
>>>
>>> OPEN SOURCE - 1 4 011 && 011 4 1
>>>
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
>>
>>
>>
>> --
>> Eyal Edri
>> Associate Manager
>> RHEV DevOps
>> EMEA ENG Virtualization R&D
>> Red Hat Israel
>>
>> phone: +972-9-7692018
>> irc: eedri (on #tlv #rhev-dev #rhev-integ)
>
>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


[Attention] Jenkins maintenance today(24/06/2016 01:00 AM TLV)

2016-06-23 Thread Nadav Goldin
Hi,
As part of an infrastructure upgrade, in approximately one hour at
01:00 AM TLV, http://jenkins.ovirt.org will be shut down for
maintenance, expected downtime is 15 minutes.
Patches sent during the downtime will be checked afterwards, patches
sent around 40 minutes prior to the downtime might not get checked.

If patches you sent did not trigger CI, you can login after the
downtime and re-trigger them manually.

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [Attention] Jenkins maintenance today(24/06/2016 01:00 AM TLV)

2016-06-23 Thread Nadav Goldin
Jenkins is back up to normal.



On Fri, Jun 24, 2016 at 12:07 AM, Nadav Goldin  wrote:
> Hi,
> As part of an infrastructure upgrade, in approximately one hour at
> 01:00 AM TLV, http://jenkins.ovirt.org will be shut down for
> maintenance, expected downtime is 15 minutes.
> Patches sent during the downtime will be checked afterwards, patches
> sent around 40 minutes prior to the downtime might not get checked.
>
> If patches you sent did not trigger CI, you can login after the
> downtime and re-trigger them manually.
>
> Thanks,
>
> Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Migration of Jenkins VM to new Cluster failed

2016-06-23 Thread Nadav Goldin
Hi Evgheni,
Unfortunately migrating the Jenkins VM failed, to (my) luck its back
running in the old Production cluster. so we could track this I am
listing again the steps taken today:

1. around 18:00 TLV time, I triggered a snapshot of the VM. This not
only failed but caused the Jenkins VM to be none-responsive for a few
minutes. More distributing is that although in the 'event's in the
engine it announced a failure, under 'snapshots' the new snapshot was
listed under status 'ok'. this also caused few CI failures(which were
re-triggered).

2. As snapshot seems like a no-option, I created a new VM in the
production cluster jenkins-2.phx.ovirt.org, and downloaded the latest
backup from backup.phx.ovirt.org, so in case of a failure we could
change the DNS and use it(keep in mind this backup does not have any
builds, only logs/configs)

3. I shut down the VM from the engine - it was hanging for a few
minutes in 'shutting down' and then announced 'shutdown failed', which
caused it to appear again in 'up' state but it was non responsive.
virsh -r --list also stated it was up.

4. I triggered another shutdown, which succeeded. As I didn't want to
risk it any more I let it boot in the same cluster, which was also
successful.

I've attached some parts of engine.log, from a quick look on vdsm.log
I didn't see anything but could help if someone else have a look(this
is ovirt-srv02). the relevant log times for the shut down failure are
from '2016-06-23 16:15'.

Either way until we find the problem, I'm not sure we should risk it
before we have a proper recovery plan. One brute-force option is using
rsync from jenkins.phx.ovirt.org:/var/lib/data/jenkins to jenkins-2,
with jenkins daemon itself shut down on 'jenkins-2', then we could
schedule a downtime on jenkins.phx.ovirt.org, wait that everything is
synced, and stop jenkins(and puppet), then start jenkins daemon on
jenkins-2 and change the DNS cname of jenkins.ovirt.org to point to
it. if everything goes smooth it should run fine, and if not, we still
have jenkins.phx.ovirt.org running.

another option is to unmount /var/lib/data/  and mount it back to
jenkins-2, though then we might be in trouble if something goes wrong
on the way.


Nadav.
engine.log
snapshot event
2016-06-23 09:06:49,592 INFO  
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-44) VM jenkins-phx-ovirt-org 
e7a7b735-0310-4f88-9ed9-4fed85835a01 moved from Up --> Paused
, Custom Event ID: -1, Message: Failed to create live snapshot 
'ngoldin_before_cluster_move' for VM 'jenkins-phx-ovirt-org'. VM restart is 
recommended. Note that using the created snapshot might cause data 
inconsistency.
, Custom Event ID: -1, Message: Failed to complete snapshot 
'ngoldin_before_cluster_move' creation for VM 'jenkins-phx-ovirt-org'.
2016-06-23 09:17:29,020 INFO  
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-69) VM jenkins-phx-ovirt-org 
e7a7b735-0310-4f88-9ed9-4fed85835a01 moved from Paused --> Up

failed shutdown
2016-06-23 15:59:20,348 INFO  [org.ovirt.engine.core.bll.ShutdownVmCommand] 
(org.ovirt.thread.pool-8-thread-25) [52b9dd27] Entered (VM 
jenkins-phx-ovirt-org).
2016-06-23 15:59:20,349 INFO  [org.ovirt.engine.core.bll.ShutdownVmCommand] 
(org.ovirt.thread.pool-8-thread-25) [52b9dd27] Sending shutdown command for VM 
jenkins-phx-ovirt-org.
2016-06-23 15:59:20,446 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(org.ovirt.thread.pool-8-thread-25) [52b9dd27] Correlation ID: 52b9dd27, Job 
ID: f1f0d78e-ae68-465e-a3c1-e46d146fc2e7, Call Stack: null, Custom Event ID: 
-1, Message: VM shutdown initiated by admin on VM jenkins-phx-ovirt-org (Host: 
ovirt-srv02) (Reason: Not Specified).
2016-06-23 16:04:20,556 INFO  
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] 
(DefaultQuartzScheduler_Worker-20) [2d2d1b3a] VM jenkins-phx-ovirt-org 
e7a7b735-0310-4f88-9ed9-4fed85835a01 moved from PoweringDown --> Up
2016-06-23 16:04:20,628 WARN  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] 
(DefaultQuartzScheduler_Worker-20) [2d2d1b3a] Correlation ID: null, Call Stack: 
null, Custom Event ID: -1, Message: Shutdown of VM jenkins-phx-ovirt-org failed.


___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


[FYI] aritfactory.ovirt.org upgrade

2016-06-26 Thread Nadav Goldin
Hi,
Yesterday the artifactory VM was upgraded to a new version, which is also
puppet-managed. The new server is at artifactory-2.phx.ovirt.org and
artifactory.ovirt.org resolves to it. All users were migrated. So far I
have seen no failures in CI, but if you see anything related, please ping
me.

For recovery purposes, the old VM is still available at
artifactory.phx.ovirt.org.

Thanks,

Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Change in ovirt-engine[ovirt-engine-4.0]: aaa: Empty keystore passwd should be passed as null to KeySt...

2016-06-27 Thread Nadav Goldin
should be ok now -
re-triggered
http://jenkins.ovirt.org/job/ovirt-engine_4.0_find-bugs_created_test/37


On Mon, Jun 27, 2016 at 10:12 PM, Martin Perina  wrote:

> Hi,
>
> could you please take a look?
>
> Thanks
>
> Martin
>
>
> On Mon, Jun 27, 2016 at 9:08 PM, Jenkins CI 
> wrote:
>
>> Jenkins CI has posted comments on this change.
>>
>> Change subject: aaa: Empty keystore passwd should be passed as null to
>> KeyStore.load
>> ..
>>
>>
>> Patch Set 1:
>>
>> Build Failed
>>
>> http://jenkins.ovirt.org/job/ovirt-engine_4.0_find-bugs_created_test/35/
>> : There was an infra issue, please contact infra@ovirt.org
>>
>> http://jenkins.ovirt.org/job/ovirt-engine_4.0_find-bugs_gerrit/177/ :
>> SUCCESS
>>
>> http://jenkins.ovirt.org/job/ovirt-engine_4.0_check-patch-el7-x86_64/316/
>> : SUCCESS
>>
>> http://jenkins.ovirt.org/job/ovirt-engine_4.0_check-patch-fc23-x86_64/315/
>> : SUCCESS
>>
>> --
>> To view, visit https://gerrit.ovirt.org/59857
>> To unsubscribe, visit https://gerrit.ovirt.org/settings
>>
>> Gerrit-MessageType: comment
>> Gerrit-Change-Id: I0f9c3706ffe150f848f231ace3ec5995d3ee2d85
>> Gerrit-PatchSet: 1
>> Gerrit-Project: ovirt-engine
>> Gerrit-Branch: ovirt-engine-4.0
>> Gerrit-Owner: Martin Peřina 
>> Gerrit-Reviewer: Jenkins CI
>> Gerrit-Reviewer: Martin Peřina 
>> Gerrit-Reviewer: Ravi Nori 
>> Gerrit-Reviewer: gerrit-hooks 
>> Gerrit-HasComments: No
>>
>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-infra] jenkins is misbehaving

2016-06-29 Thread Nadav Goldin
1. Its the second time it happens this week[1]
2. Around a month a go, I did a log analyse of how often this happens,
and it was more than 10 times a week.
3. After Shlomi resolved few issues on Gerrit, it seem to have gone away.
My guess is that this is network related or overload on Gerrit - it
either fails when trying to connect to Gerrit, or while cloning(like
in [1]). I didn't find any consistency in the error, which makes it
hard to reproduce. The current re-trigger Anton did was on a BM metal
slave, so I doubt its related to overload on the Jenkins slave itself.

[1] 
http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3281/console

On Wed, Jun 29, 2016 at 7:47 PM, Eyal Edri  wrote:
> Shlomi arw you running anything on gerrit now?
> If you're copying the content please stop as it might affect gerrit
> performance.
>
> On Jun 29, 2016 7:37 PM, "Anton Marchukov"  wrote:
>>
>> Hello All.
>>
>> I tried to clone manually and this works:
>>
>> [amarchuk@ovirt-srv22 ~]$ git clone
>> git://gerrit.ovirt.org/ovirt-engine.git
>> Cloning into 'ovirt-engine'...
>> remote: Counting objects: 784726, done.
>> remote: Compressing objects: 100% (204209/204209), done.
>> remote: Total 784726 (delta 360293), reused 777805 (delta 358840)
>> Receiving objects: 100% (784726/784726), 136.26 MiB | 28.66 MiB/s, done.
>> Resolving deltas: 100% (360293/360293), done.
>>
>>
>> But failed in the job
>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3379/console
>>
>> So it is either not 100% reproducible or some Jenkins issue.
>>
>> Anybody did anything on Jenkins recently that can be correlated with this?
>>
>> Also gerrit plugin started to loose events...
>>
>> Anton.
>>
>> On Wed, Jun 29, 2016 at 6:14 PM, Piotr Kliczewski
>>  wrote:
>>>
>>> Some time ago jenkins did not update the patches with the score. Now I
>>> see that builds are not triggered. One of the builds that I triggered
>>> manually [1] failed with:
>>>
>>> 16:04:47 ERROR: Timeout after 10 minutes
>>> 16:04:47 ERROR: Error cloning remote repo 'origin'
>>> 16:04:47 hudson.plugins.git.GitException: Command "git fetch --tags
>>> --progress git://gerrit.ovirt.org/ovirt-engine.git
>>> +refs/heads/*:refs/remotes/origin/*" returned status code 143:
>>> 16:04:47 stdout:
>>> 16:04:47 stderr:
>>> 16:04:47 at
>>> org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1640)
>>>
>>> Thanks,
>>> Piotr
>>>
>>> [1]
>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-el7-x86_64/3381/console
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>>
>>
>> --
>> Anton Marchukov
>> Senior Software Engineer - RHEV CI - Red Hat
>>
>>
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-infra] jenkins is misbehaving

2016-06-29 Thread Nadav Goldin
I was more or less able to reproduce the problem. I ran git clone on
ovirt-engine.git from one of the VMs in the Jenkins_CentOS cluster for
200 times, with time out set to 90 seconds, and 15 seconds between
each clone. It had 13/200 failures, which is exactly 6.5%. This
explains why we don't see it often, it might be more severe as this
testing was done during the night when Jenkins/Gerrit aren't busy.
During that time there were few, but not 13, exceptions in gerrit's
error_log:
[2016-06-29 19:18:26,933] [NioProcessor-1] WARN
com.google.gerrit.sshd.GerritServerSession : Exception caught
org.apache.sshd.common.SshException: Received 97 on unknown channel 0
at 
org.apache.sshd.common.session.AbstractConnectionService.getChannel(AbstractConnectionService.java:301)
.
Since it doesn't have the client IP log, its hard to tell if it is
correlated, even if it is, not all attempts reached to the exception
log. So there is a problem, independent of Jenkins itself. Will need
to dig deeper to find out what is causing it..




On Wed, Jun 29, 2016 at 8:22 PM, Nadav Goldin  wrote:
> 1. Its the second time it happens this week[1]
> 2. Around a month a go, I did a log analyse of how often this happens,
> and it was more than 10 times a week.
> 3. After Shlomi resolved few issues on Gerrit, it seem to have gone away.
> My guess is that this is network related or overload on Gerrit - it
> either fails when trying to connect to Gerrit, or while cloning(like
> in [1]). I didn't find any consistency in the error, which makes it
> hard to reproduce. The current re-trigger Anton did was on a BM metal
> slave, so I doubt its related to overload on the Jenkins slave itself.
>
> [1] 
> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3281/console
>
> On Wed, Jun 29, 2016 at 7:47 PM, Eyal Edri  wrote:
>> Shlomi arw you running anything on gerrit now?
>> If you're copying the content please stop as it might affect gerrit
>> performance.
>>
>> On Jun 29, 2016 7:37 PM, "Anton Marchukov"  wrote:
>>>
>>> Hello All.
>>>
>>> I tried to clone manually and this works:
>>>
>>> [amarchuk@ovirt-srv22 ~]$ git clone
>>> git://gerrit.ovirt.org/ovirt-engine.git
>>> Cloning into 'ovirt-engine'...
>>> remote: Counting objects: 784726, done.
>>> remote: Compressing objects: 100% (204209/204209), done.
>>> remote: Total 784726 (delta 360293), reused 777805 (delta 358840)
>>> Receiving objects: 100% (784726/784726), 136.26 MiB | 28.66 MiB/s, done.
>>> Resolving deltas: 100% (360293/360293), done.
>>>
>>>
>>> But failed in the job
>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3379/console
>>>
>>> So it is either not 100% reproducible or some Jenkins issue.
>>>
>>> Anybody did anything on Jenkins recently that can be correlated with this?
>>>
>>> Also gerrit plugin started to loose events...
>>>
>>> Anton.
>>>
>>> On Wed, Jun 29, 2016 at 6:14 PM, Piotr Kliczewski
>>>  wrote:
>>>>
>>>> Some time ago jenkins did not update the patches with the score. Now I
>>>> see that builds are not triggered. One of the builds that I triggered
>>>> manually [1] failed with:
>>>>
>>>> 16:04:47 ERROR: Timeout after 10 minutes
>>>> 16:04:47 ERROR: Error cloning remote repo 'origin'
>>>> 16:04:47 hudson.plugins.git.GitException: Command "git fetch --tags
>>>> --progress git://gerrit.ovirt.org/ovirt-engine.git
>>>> +refs/heads/*:refs/remotes/origin/*" returned status code 143:
>>>> 16:04:47 stdout:
>>>> 16:04:47 stderr:
>>>> 16:04:47 at
>>>> org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1640)
>>>>
>>>> Thanks,
>>>> Piotr
>>>>
>>>> [1]
>>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-el7-x86_64/3381/console
>>>> ___
>>>> Infra mailing list
>>>> Infra@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
>>>
>>>
>>>
>>> --
>>> Anton Marchukov
>>> Senior Software Engineer - RHEV CI - Red Hat
>>>
>>>
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
>>
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-infra] jenkins is misbehaving

2016-07-05 Thread Nadav Goldin
Opened a ticket for tracking[1], still didn't get to do more tests.
Also, this never happens in VDSM, probably because ovirt-engine.git is
244MB while vdsm.git is only 52MB.

another failure:
http://jenkins.ovirt.org/job/ovirt-engine_3.6_upgrade-db-from-3.6_el6_created/44/





[1] https://ovirt-jira.atlassian.net/browse/OVIRT-403

On Thu, Jun 30, 2016 at 4:11 AM, Nadav Goldin  wrote:
> I was more or less able to reproduce the problem. I ran git clone on
> ovirt-engine.git from one of the VMs in the Jenkins_CentOS cluster for
> 200 times, with time out set to 90 seconds, and 15 seconds between
> each clone. It had 13/200 failures, which is exactly 6.5%. This
> explains why we don't see it often, it might be more severe as this
> testing was done during the night when Jenkins/Gerrit aren't busy.
> During that time there were few, but not 13, exceptions in gerrit's
> error_log:
> [2016-06-29 19:18:26,933] [NioProcessor-1] WARN
> com.google.gerrit.sshd.GerritServerSession : Exception caught
> org.apache.sshd.common.SshException: Received 97 on unknown channel 0
> at 
> org.apache.sshd.common.session.AbstractConnectionService.getChannel(AbstractConnectionService.java:301)
> .
> Since it doesn't have the client IP log, its hard to tell if it is
> correlated, even if it is, not all attempts reached to the exception
> log. So there is a problem, independent of Jenkins itself. Will need
> to dig deeper to find out what is causing it..
>
>
>
>
> On Wed, Jun 29, 2016 at 8:22 PM, Nadav Goldin  wrote:
>> 1. Its the second time it happens this week[1]
>> 2. Around a month a go, I did a log analyse of how often this happens,
>> and it was more than 10 times a week.
>> 3. After Shlomi resolved few issues on Gerrit, it seem to have gone away.
>> My guess is that this is network related or overload on Gerrit - it
>> either fails when trying to connect to Gerrit, or while cloning(like
>> in [1]). I didn't find any consistency in the error, which makes it
>> hard to reproduce. The current re-trigger Anton did was on a BM metal
>> slave, so I doubt its related to overload on the Jenkins slave itself.
>>
>> [1] 
>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3281/console
>>
>> On Wed, Jun 29, 2016 at 7:47 PM, Eyal Edri  wrote:
>>> Shlomi arw you running anything on gerrit now?
>>> If you're copying the content please stop as it might affect gerrit
>>> performance.
>>>
>>> On Jun 29, 2016 7:37 PM, "Anton Marchukov"  wrote:
>>>>
>>>> Hello All.
>>>>
>>>> I tried to clone manually and this works:
>>>>
>>>> [amarchuk@ovirt-srv22 ~]$ git clone
>>>> git://gerrit.ovirt.org/ovirt-engine.git
>>>> Cloning into 'ovirt-engine'...
>>>> remote: Counting objects: 784726, done.
>>>> remote: Compressing objects: 100% (204209/204209), done.
>>>> remote: Total 784726 (delta 360293), reused 777805 (delta 358840)
>>>> Receiving objects: 100% (784726/784726), 136.26 MiB | 28.66 MiB/s, done.
>>>> Resolving deltas: 100% (360293/360293), done.
>>>>
>>>>
>>>> But failed in the job
>>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3379/console
>>>>
>>>> So it is either not 100% reproducible or some Jenkins issue.
>>>>
>>>> Anybody did anything on Jenkins recently that can be correlated with this?
>>>>
>>>> Also gerrit plugin started to loose events...
>>>>
>>>> Anton.
>>>>
>>>> On Wed, Jun 29, 2016 at 6:14 PM, Piotr Kliczewski
>>>>  wrote:
>>>>>
>>>>> Some time ago jenkins did not update the patches with the score. Now I
>>>>> see that builds are not triggered. One of the builds that I triggered
>>>>> manually [1] failed with:
>>>>>
>>>>> 16:04:47 ERROR: Timeout after 10 minutes
>>>>> 16:04:47 ERROR: Error cloning remote repo 'origin'
>>>>> 16:04:47 hudson.plugins.git.GitException: Command "git fetch --tags
>>>>> --progress git://gerrit.ovirt.org/ovirt-engine.git
>>>>> +refs/heads/*:refs/remotes/origin/*" returned status code 143:
>>>>> 16:04:47 stdout:
>>>>> 16:04:47 stderr:
>>>>> 16:04:47 at
>>>>> org.jenkinsci.plugins.gitclient.CliGitAPIImpl.launchCommandIn(CliGitAPIImpl.java:1640)
>>>>>
>>>>> Thanks,
>>>>> Piotr
>>>>>
>>>>> [1]
>>>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-el7-x86_64/3381/console
>>>>> ___
>>>>> Infra mailing list
>>>>> Infra@ovirt.org
>>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Anton Marchukov
>>>> Senior Software Engineer - RHEV CI - Red Hat
>>>>
>>>>
>>>> ___
>>>> Infra mailing list
>>>> Infra@ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>>
>>>
>>> ___
>>> Infra mailing list
>>> Infra@ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/infra
>>>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-infra] jenkins is misbehaving

2016-07-05 Thread Nadav Goldin
fixing wrong ticket link:
[1] https://ovirt-jira.atlassian.net/browse/OVIRT-619


On Tue, Jul 5, 2016 at 11:45 AM, Nadav Goldin  wrote:
> Opened a ticket for tracking[1], still didn't get to do more tests.
> Also, this never happens in VDSM, probably because ovirt-engine.git is
> 244MB while vdsm.git is only 52MB.
>
> another failure:
> http://jenkins.ovirt.org/job/ovirt-engine_3.6_upgrade-db-from-3.6_el6_created/44/
>
>
>
>
>
> [1] https://ovirt-jira.atlassian.net/browse/OVIRT-403
>
> On Thu, Jun 30, 2016 at 4:11 AM, Nadav Goldin  wrote:
>> I was more or less able to reproduce the problem. I ran git clone on
>> ovirt-engine.git from one of the VMs in the Jenkins_CentOS cluster for
>> 200 times, with time out set to 90 seconds, and 15 seconds between
>> each clone. It had 13/200 failures, which is exactly 6.5%. This
>> explains why we don't see it often, it might be more severe as this
>> testing was done during the night when Jenkins/Gerrit aren't busy.
>> During that time there were few, but not 13, exceptions in gerrit's
>> error_log:
>> [2016-06-29 19:18:26,933] [NioProcessor-1] WARN
>> com.google.gerrit.sshd.GerritServerSession : Exception caught
>> org.apache.sshd.common.SshException: Received 97 on unknown channel 0
>> at 
>> org.apache.sshd.common.session.AbstractConnectionService.getChannel(AbstractConnectionService.java:301)
>> .
>> Since it doesn't have the client IP log, its hard to tell if it is
>> correlated, even if it is, not all attempts reached to the exception
>> log. So there is a problem, independent of Jenkins itself. Will need
>> to dig deeper to find out what is causing it..
>>
>>
>>
>>
>> On Wed, Jun 29, 2016 at 8:22 PM, Nadav Goldin  wrote:
>>> 1. Its the second time it happens this week[1]
>>> 2. Around a month a go, I did a log analyse of how often this happens,
>>> and it was more than 10 times a week.
>>> 3. After Shlomi resolved few issues on Gerrit, it seem to have gone away.
>>> My guess is that this is network related or overload on Gerrit - it
>>> either fails when trying to connect to Gerrit, or while cloning(like
>>> in [1]). I didn't find any consistency in the error, which makes it
>>> hard to reproduce. The current re-trigger Anton did was on a BM metal
>>> slave, so I doubt its related to overload on the Jenkins slave itself.
>>>
>>> [1] 
>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3281/console
>>>
>>> On Wed, Jun 29, 2016 at 7:47 PM, Eyal Edri  wrote:
>>>> Shlomi arw you running anything on gerrit now?
>>>> If you're copying the content please stop as it might affect gerrit
>>>> performance.
>>>>
>>>> On Jun 29, 2016 7:37 PM, "Anton Marchukov"  wrote:
>>>>>
>>>>> Hello All.
>>>>>
>>>>> I tried to clone manually and this works:
>>>>>
>>>>> [amarchuk@ovirt-srv22 ~]$ git clone
>>>>> git://gerrit.ovirt.org/ovirt-engine.git
>>>>> Cloning into 'ovirt-engine'...
>>>>> remote: Counting objects: 784726, done.
>>>>> remote: Compressing objects: 100% (204209/204209), done.
>>>>> remote: Total 784726 (delta 360293), reused 777805 (delta 358840)
>>>>> Receiving objects: 100% (784726/784726), 136.26 MiB | 28.66 MiB/s, done.
>>>>> Resolving deltas: 100% (360293/360293), done.
>>>>>
>>>>>
>>>>> But failed in the job
>>>>> http://jenkins.ovirt.org/job/ovirt-engine_master_check-patch-fc23-x86_64/3379/console
>>>>>
>>>>> So it is either not 100% reproducible or some Jenkins issue.
>>>>>
>>>>> Anybody did anything on Jenkins recently that can be correlated with this?
>>>>>
>>>>> Also gerrit plugin started to loose events...
>>>>>
>>>>> Anton.
>>>>>
>>>>> On Wed, Jun 29, 2016 at 6:14 PM, Piotr Kliczewski
>>>>>  wrote:
>>>>>>
>>>>>> Some time ago jenkins did not update the patches with the score. Now I
>>>>>> see that builds are not triggered. One of the builds that I triggered
>>>>>> manually [1] failed with:
>>>>>>
>>>>>> 16:04:47 ERROR: Timeout after 10 minutes
>>>>>> 16:04:47 ERROR: Error cloning remote repo 'origin'
>>>>>> 16:04:47 hudson.plugins.git.GitEx

[FYI] Jenkins maintenance today 18:30 TLV

2016-07-06 Thread Nadav Goldin
Hi,
http://jenkins.ovirt.org will be restarted for plugin updates today at
18:30 TLV time, expected downtime is 30 minutes. Patches sent 30
minutes before might not get checked, patches sent during the downtime
will get checked when Jenkins is back.

If patches you sent did not trigger CI, you can login after the
downtime and re-trigger them manually.

Thanks,
Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [FYI] Jenkins maintenance today 18:30 TLV

2016-07-06 Thread Nadav Goldin
Jenkins is back up. If you see any issues please email infra-supp...@ovirt.org

Thanks,
Nadav.


On Wed, Jul 6, 2016 at 5:48 PM, Nadav Goldin  wrote:
> Hi,
> http://jenkins.ovirt.org will be restarted for plugin updates today at
> 18:30 TLV time, expected downtime is 30 minutes. Patches sent 30
> minutes before might not get checked, patches sent during the downtime
> will get checked when Jenkins is back.
>
> If patches you sent did not trigger CI, you can login after the
> downtime and re-trigger them manually.
>
> Thanks,
> Nadav.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: Build failed in Jenkins: ovirt_master_system-tests #243

2016-07-07 Thread Nadav Goldin
Seems like [1], as ovirt-srv19  has fresh new FC24 installation,
virtlogd is not enabled by default:
● virtlogd.service - Virtual machine log manager
   Loaded: loaded (/usr/lib/systemd/system/virtlogd.service; indirect;
vendor preset: disabled)
   Active: inactive (dead)
 Docs: man:virtlogd(8)
   http://libvirt.org
we can add it to puppet for now.

[1] https://bugzilla.redhat.com/show_bug.cgi?id=1290357


On Thu, Jul 7, 2016 at 6:49 PM, Eyal Edri  wrote:
> This looks like a bug in libvirt?
> Tolik mentioned something in a socket name which is too long, anyone seen it
> before?
>
> 15:37:11 libvirt: XML-RPC error : Failed to connect socket to
> '/var/run/libvirt/virtlogd-sock': No such file or directory
> 15:37:11 * Starting VM lago_basic_suite_master_storage: ERROR (in
> 0:00:00)
> 15:37:11   # Start vms: ERROR (in 0:00:00)
> 15:37:11   # Destroy network lago_basic_suite_master_lago:
> 15:37:11   # Destroy network lago_basic_suite_master_lago: ERROR (in
> 0:00:00)
> 15:37:11 @ Start Prefix: ERROR (in 0:00:00)
> 15:37:11 Error occured, aborting
> 15:37:11 Traceback (most recent call last):
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/cmd.py", line 691, in
> main
> 15:37:11 cli_plugins[args.verb].do_run(args)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/plugins/cli.py", line
> 180, in do_run
> 15:37:11 self._do_run(**vars(args))
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/utils.py", line 488,
> in wrapper
> 15:37:11 return func(*args, **kwargs)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/utils.py", line 499,
> in wrapper
> 15:37:11 return func(*args, prefix=prefix, **kwargs)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/cmd.py", line 255, in
> do_start
> 15:37:11 prefix.start(vm_names=vm_names)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/prefix.py", line 958,
> in start
> 15:37:11 self.virt_env.start(vm_names=vm_names)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/virt.py", line 182,
> in start
> 15:37:11 vm.start()
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/plugins/vm.py", line
> 247, in start
> 15:37:11 return self.provider.start(*args, **kwargs)
> 15:37:11   File "/usr/lib/python2.7/site-packages/lago/vm.py", line 93, in
> start
> 15:37:11 self.libvirt_con.createXML(self._libvirt_xml())
> 15:37:11   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 3611,
> in createXML
> 15:37:11 if ret is None:raise libvirtError('virDomainCreateXML()
> failed', conn=self)
> 15:37:11 libvirtError: Failed to connect socket to
> '/var/run/libvirt/virtlogd-sock': No such file or directory
> 15:37:11 #
>
>
> On Thu, Jul 7, 2016 at 6:37 PM,  wrote:
>>
>> See 
>>
>> Changes:
>>
>> [Eyal Edri] add hystrix deps to yum repos include list
>>
>> [Eyal Edri] refresh fedora versions and release versions for ovirt-engine
>>
>> [Sandro Bonazzola] ovirt-engine_upgrade-db: drop 3.6.7 jobs
>>
>> [Shirly Radco] Replacing jpackage repo for 3.6 dwh
>>
>> --
>> [...truncated 485 lines...]
>> ##  rc = 1
>> ##
>> ##! ERROR v
>> ##! Last 20 log enties:
>> logs/mocker-fedora-23-x86_64.fc23.basic_suite_master.sh/basic_suite_master.sh.log
>> ##!
>>   File "/usr/lib/python2.7/site-packages/lago/cmd.py", line 255, in
>> do_start
>> prefix.start(vm_names=vm_names)
>>   File "/usr/lib/python2.7/site-packages/lago/prefix.py", line 958, in
>> start
>> self.virt_env.start(vm_names=vm_names)
>>   File "/usr/lib/python2.7/site-packages/lago/virt.py", line 182, in start
>> vm.start()
>>   File "/usr/lib/python2.7/site-packages/lago/plugins/vm.py", line 247, in
>> start
>> return self.provider.start(*args, **kwargs)
>>   File "/usr/lib/python2.7/site-packages/lago/vm.py", line 93, in start
>> self.libvirt_con.createXML(self._libvirt_xml())
>>   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 3611, in
>> createXML
>> if ret is None:raise libvirtError('virDomainCreateXML() failed',
>> conn=self)
>> libvirtError: Failed to connect socket to
>> '/var/run/libvirt/virtlogd-sock': No such file or directory
>> #
>>  Cleaning up
>> --- Cleaning with lago
>> --- Cleaning with lago done
>>  Cleanup done
>> Took 197 seconds
>> ===
>> ##!
>> ##! ERROR ^^
>> ##!
>> ##
>> Build step 'Execute shell' marked build as failure
>> Performing Post build task...
>> Match found for :.* : True
>> Logical operation result is TRUE
>> Running script  : #!/bin/bash -xe
>> echo 'shell_scripts/system_tests.collect_logs.sh'
>>
>> #

Re: [lago-devel] lago_master_check-patch-fc23-x86_64 #224 Console [Jenkins]

2016-09-06 Thread Nadav Goldin
Looks like [1] indeed. I ran my failed build with jenkins repo prior
to the patch and it works.

Didi, any ideas?


Thanks,

Nadav.


http://jenkins.ovirt.org/job/lago_master_check-merged-fc23-x86_64/200/
(reverted - works)
http://jenkins.ovirt.org/job/lago_master_check-merged-fc23-x86_64/196/
(with [1] fails)

[1] https://gerrit.ovirt.org/#/c/63249/

On Tue, Sep 6, 2016 at 10:40 AM, Nadav Goldin  wrote:
> Ran into the same problem yesterday, I suspect this patch [1] to
> mock_runner.sh. The patch might be legit but exposed some other
> caching issues.
> Investigating..
>
>
> [1] https://gerrit.ovirt.org/#/c/63249/
>
> On Tue, Sep 6, 2016 at 10:09 AM, Yaniv Kaul  wrote:
>> New breakage (that I hope) has nothing to do with my patch:
>>
>> Error: nothing provides repoman >= 2.0.12 needed by
>> python-lago-ovirt-0.24.1-1.fc23.noarch
>>
>> Anyone?
>>
>> http://jenkins.ovirt.org/job/lago_master_check-patch-fc23-x86_64/224/console
>>
>> ___
>> lago-devel mailing list
>> lago-de...@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/lago-devel
>>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is WARNING **

2016-09-11 Thread Nadav Goldin
Hi,
looks like http://www.gtlib.gatech.edu/pub/oVirt/pub is unreachable in
the past 24 hours or so, can you check and update us?

Thanks,

Nadav



-- Forwarded message --
From: icinga 
Date: Sun, Sep 11, 2016 at 12:33 PM
Subject: ** PROBLEM Service Alert:
ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site
last sync is WARNING **
To: ngol...@redhat.com


* Icinga *

Notification Type: PROBLEM

Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
Host: ovirt-mirrorchecker
Address: 66.187.230.105
State: WARNING

Date/Time: Sun Sept 11 09:33:58 UTC 2016

Additional Info:

WARNING - 111890 seconds since last sync, which are 31.0806 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is WARNING **

2016-09-25 Thread Nadav Goldin
Hi,
Ping again :) looks like its been down during most of the weekend.

As a side note - you mind if we add 'gt...@gtlib.gatech.edu' email
address to this specific Icinga alert?


Thanks!

Nadav.



-- Forwarded message --
From: icinga 
Date: Sun, Sep 25, 2016 at 11:49 PM
Subject: ** PROBLEM Service Alert:
ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site
last sync is WARNING **
To: ngol...@redhat.com


* Icinga *

Notification Type: PROBLEM

Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
Host: ovirt-mirrorchecker
Address: 66.187.230.105
State: WARNING

Date/Time: Sun Sept 25 20:49:57 UTC 2016

Additional Info:

WARNING - 237933 seconds since last sync, which are 66.0925 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is WARNING **

2016-09-25 Thread Nadav Goldin
Hi,
Not sure if this is the right address - sent this to
'gt...@gtlib.gatech.edu' but got an out of office reply from Neil
Bright.
The oVirt mirror site hosted at www.gtlib.gatech.edu/pub/oVirt/pub
looks down for more than 3 days(actually looks like the entire
gtlib.gatech.edu domain is down). Can you check?

Thanks,

Nadav.



-- Forwarded message ------
From: Nadav Goldin 
Date: Mon, Sep 26, 2016 at 12:09 AM
Subject: Fwd: ** PROBLEM Service Alert:
ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site
last sync is WARNING **
To: gt...@gtlib.gatech.edu
Cc: infra 


Hi,
Ping again :) looks like its been down during most of the weekend.

As a side note - you mind if we add 'gt...@gtlib.gatech.edu' email
address to this specific Icinga alert?


Thanks!

Nadav.



-- Forwarded message --
From: icinga 
Date: Sun, Sep 25, 2016 at 11:49 PM
Subject: ** PROBLEM Service Alert:
ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site
last sync is WARNING **
To: ngol...@redhat.com


* Icinga *

Notification Type: PROBLEM

Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
Host: ovirt-mirrorchecker
Address: 66.187.230.105
State: WARNING

Date/Time: Sun Sept 25 20:49:57 UTC 2016

Additional Info:

WARNING - 237933 seconds since last sync, which are 66.0925 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/ftp.snt.utwente.nl/pub/software/ovirt mirror site last sync is WARNING **

2016-10-02 Thread Nadav Goldin
Hi,
It looks like ovirt mirror site is not syncing for the past 48 hours
or so. From the logs I see you are connecting and running rsync, but
the files aren't synchronized.
For example:
http://ftp.snt.utwente.nl/pub/software/ovirt/ovirt-master-snapshot/src/ovirt-engine-sdk/ovirt-engine-sdk-4.1.0.alpha0.20160929gitfd328cd.tar.gz(30/09/2016)
http://resources.ovirt.org/pub/ovirt-master-snapshot/src/ovirt-engine-sdk/ovirt-engine-sdk-4.1.0.alpha0.20160929gitfd328cd.tar.gz
(updated 2/10/2016)

http://resources.ovirt.org/pub/ovirt-master-snapshot/rpm/fc24/x86_64/vdsm-hook-extra-ipv4-addrs-4.18.999-661.git3267d12.fc24.x86_64.rpm
(doesn't exist @ utwente.nl)
and the  '.timestamp' files on which this alert is triggered.

Also, I would suggest adding '--delete'(if you don't have it yet) to
the rsync command to save some space and to avoid storing unneeded
artifacts.

Can you please have a look?

Thanks!

Nadav.


-- Forwarded message --
From: icinga 
Date: Sun, Oct 2, 2016 at 11:59 AM
Subject: ** PROBLEM Service Alert:
ovirt-mirrorchecker/ftp.snt.utwente.nl/pub/software/ovirt mirror site
last sync is WARNING **
To: ngol...@redhat.com


* Icinga *

Notification Type: PROBLEM

Service: ftp.snt.utwente.nl/pub/software/ovirt mirror site last sync
Host: ovirt-mirrorchecker
Address: 66.187.230.105
State: WARNING

Date/Time: Sun Oct 2 08:59:57 UTC 2016

Additional Info:

WARNING - 194861 seconds since last sync, which are 54.1281 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [Ftpcom] Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/ftp.snt.utwente.nl/pub/software/ovirt mirror site last sync is WARNING **

2016-10-04 Thread Nadav Goldin
Hi Maarten,
Thanks for the reply. It seems the synchronization is still broken,
from the logs I see an attempted rsync connection at:
> START:Tue Oct  4 02:00:24 UTC 2016
>END:Tue Oct  4 02:23:55 UTC 2016
But looking at the timestamps, all files I sampled still date back to
29/09/2016.


> In other news; we have replacement hardware underway for the mirror, so
> I expect a maintenance window before the end of the year. If it'll
> impact the ovirt mirror in a meaningful way, we'll reach out.
Great, we'll appreciate that.


Nadav.




On Sun, Oct 2, 2016 at 9:34 PM, Maarten Aertsen  wrote:
> Hi Nadav,
>
> On 2016-10-02 11:22, Nadav Goldin wrote:
>> It looks like ovirt mirror site is not syncing for the past 48 hours
>> or so. From the logs I see you are connecting and running rsync, but
>> the files aren't synchronized.
>
> We're currently experiencing a pretty big spike for one of the other
> projects we're mirroring, so our own monitoring shows similar warnings.
> The rsync is currently running but proceeding very slowly. We expect
> traffic to slow down shortly; please let us know if things do not return
> to normal within one or two days.
>
> In other news; we have replacement hardware underway for the mirror, so
> I expect a maintenance window before the end of the year. If it'll
> impact the ovirt mirror in a meaningful way, we'll reach out.
>
> best regards, Maarten
>
> --
> Maarten Aertsen
> SNT FTPcom
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [Ftpcom] Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/ftp.snt.utwente.nl/pub/software/ovirt mirror site last sync is WARNING **

2016-10-05 Thread Nadav Goldin
Hi again Maarten,
any updates on this?

Thanks,
Nadav.


On Tue, Oct 4, 2016 at 12:04 PM, Nadav Goldin  wrote:
> Hi Maarten,
> Thanks for the reply. It seems the synchronization is still broken,
> from the logs I see an attempted rsync connection at:
>> START:Tue Oct  4 02:00:24 UTC 2016
>>END:Tue Oct  4 02:23:55 UTC 2016
> But looking at the timestamps, all files I sampled still date back to
> 29/09/2016.
>
>
>> In other news; we have replacement hardware underway for the mirror, so
>> I expect a maintenance window before the end of the year. If it'll
>> impact the ovirt mirror in a meaningful way, we'll reach out.
> Great, we'll appreciate that.
>
>
> Nadav.
>
>
>
>
> On Sun, Oct 2, 2016 at 9:34 PM, Maarten Aertsen  
> wrote:
>> Hi Nadav,
>>
>> On 2016-10-02 11:22, Nadav Goldin wrote:
>>> It looks like ovirt mirror site is not syncing for the past 48 hours
>>> or so. From the logs I see you are connecting and running rsync, but
>>> the files aren't synchronized.
>>
>> We're currently experiencing a pretty big spike for one of the other
>> projects we're mirroring, so our own monitoring shows similar warnings.
>> The rsync is currently running but proceeding very slowly. We expect
>> traffic to slow down shortly; please let us know if things do not return
>> to normal within one or two days.
>>
>> In other news; we have replacement hardware underway for the mirror, so
>> I expect a maintenance window before the end of the year. If it'll
>> impact the ovirt mirror in a meaningful way, we'll reach out.
>>
>> best regards, Maarten
>>
>> --
>> Maarten Aertsen
>> SNT FTPcom
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [Ftpcom] Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/ftp.snt.utwente.nl/pub/software/ovirt mirror site last sync is WARNING **

2016-10-05 Thread Nadav Goldin
My bad, I see it was fixed during the night,

Thanks!

Nadav.

On Thu, Oct 6, 2016 at 9:33 AM, Nadav Goldin  wrote:
> Hi again Maarten,
> any updates on this?
>
> Thanks,
> Nadav.
>
>
> On Tue, Oct 4, 2016 at 12:04 PM, Nadav Goldin  wrote:
>> Hi Maarten,
>> Thanks for the reply. It seems the synchronization is still broken,
>> from the logs I see an attempted rsync connection at:
>>> START:Tue Oct  4 02:00:24 UTC 2016
>>>END:Tue Oct  4 02:23:55 UTC 2016
>> But looking at the timestamps, all files I sampled still date back to
>> 29/09/2016.
>>
>>
>>> In other news; we have replacement hardware underway for the mirror, so
>>> I expect a maintenance window before the end of the year. If it'll
>>> impact the ovirt mirror in a meaningful way, we'll reach out.
>> Great, we'll appreciate that.
>>
>>
>> Nadav.
>>
>>
>>
>>
>> On Sun, Oct 2, 2016 at 9:34 PM, Maarten Aertsen  
>> wrote:
>>> Hi Nadav,
>>>
>>> On 2016-10-02 11:22, Nadav Goldin wrote:
>>>> It looks like ovirt mirror site is not syncing for the past 48 hours
>>>> or so. From the logs I see you are connecting and running rsync, but
>>>> the files aren't synchronized.
>>>
>>> We're currently experiencing a pretty big spike for one of the other
>>> projects we're mirroring, so our own monitoring shows similar warnings.
>>> The rsync is currently running but proceeding very slowly. We expect
>>> traffic to slow down shortly; please let us know if things do not return
>>> to normal within one or two days.
>>>
>>> In other news; we have replacement hardware underway for the mirror, so
>>> I expect a maintenance window before the end of the year. If it'll
>>> impact the ovirt mirror in a meaningful way, we'll reach out.
>>>
>>> best regards, Maarten
>>>
>>> --
>>> Maarten Aertsen
>>> SNT FTPcom
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: new jenkins slaves added

2016-10-13 Thread Nadav Goldin
Amazing, great work!


On Tue, Oct 11, 2016 at 4:34 PM, Evgheni Dereveanchin
 wrote:
> Hi everyone,
>
> As part of the process of network reorganization and
> resource usage optimization, I've spawned 34 new VMs
> to be used as Jenkins slaves. They are added to Jenkins
> and their hostnames are in workers-phx.ovirt.org zone.
>
> These reside in an isolated subnet inside the PHX
> datacenter and access most resources through NAT.
>
> If you need to log in to these VMs please use
> the PHX Foreman proxy as a jump host.
>
> There are both CentOS7 and Fedora 24 systems in this
> batch which should work just as existing ones, but
> if you see any issues - please let me know. The plan
> is to add more VMs so that we can do more consecutive
> builds.
>
> We currently have 106 slaves of all kinds attached
> to Jenkins and with current resources it should be
> possible to double this number within the
> following weeks.
>
> Regards,
> Evgheni Dereveanchin
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Fwd: ** PROBLEM Service Alert: ovirt-mirrorchecker/mirrors.ibiblio.org/ovirt/pub mirror site last sync is CRITICAL **

2016-10-24 Thread Nadav Goldin
Hi,
It seems like the oVirt mirror at ibiblio.org has not been
synchronized in the past 2 days, from the logs I see there was an
attempted rsync connection but it took way longer than expected(10
hours), so I assume it hanged and eventually failed.

Can you please have a look?

Thanks,

Nadav.


-- Forwarded message --
From: icinga 
Date: Mon, Oct 24, 2016 at 8:14 PM
Subject: ** PROBLEM Service Alert:
ovirt-mirrorchecker/mirrors.ibiblio.org/ovirt/pub mirror site last
sync is CRITICAL **
To: ngol...@redhat.com


* Icinga *

Notification Type: PROBLEM

Service: mirrors.ibiblio.org/ovirt/pub mirror site last sync
Host: ovirt-mirrorchecker
Address: 66.187.230.105
State: CRITICAL

Date/Time: Mon Oct 24 17:14:07 UTC 2016

Additional Info:

CRITICAL - 432040 seconds since last sync, which are 120.0111 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Failure on start VM in ovirt-system-tests from patches merged to master on the 25/10/2016

2016-10-26 Thread Nadav Goldin
Hi,
We have new failure on OST from patches merged to master yesterday,
the failure started after the merge of [1], but as there were quite a
few patches merged quickly I can't make sure it is the one causing
it(OST aren't ran per-patch).
The test that fails is [2] when attempting to start the VM.

The error from the API side:

RequestError:
status: 500
reason: Internal Server Error
detail: javax.ejb.EJBException: java.lang.NullPointerException
at 
org.jboss.as.ejb3.tx.CMTTxInterceptor.handleExceptionInNoTx(CMTTxInterceptor.java:213)
at 
org.jboss.as.ejb3.tx.CMTTxInterceptor.invokeInNoTx(CMTTxInterceptor.java:265)
at org.jboss.as.ejb3.tx.CMTTxInterceptor.supports(CMTTxInterceptor.java:374)
at 
org.jboss.as.ejb3.tx.CMTTxInterceptor.processInvocation(CMTTxInterceptor.java:243)
at 
org.jboss.invocation.InterceptorContext.proceed(InterceptorContext.java:340)


In the engine logs there are a few 'java.lang.NullPointerException' errors:

2016-10-25 11:53:52,845 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
(org.ovirt.thread.pool-6-thread-2) [5e6a88be] Failed to get vds
'd60db21f-95f0-487b-9f17-44861e2610a7', error: null
2016-10-25 11:53:52,864 DEBUG
[org.ovirt.engine.core.utils.timer.FixedDelayJobListener]
(DefaultQuartzScheduler5) [] Rescheduling
DEFAULT.org.ovirt.engine.core.bll.tasks.AsyncTaskManager.timerElapsed#-9223372036854775787
as there is no unfired trigger.
...
2016-10-25 11:53:52,845 DEBUG
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
(org.ovirt.thread.pool-6-thread-2) [5e6a88be] Exception:
java.lang.NullPointerException
at 
org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsStatic(AuditLogableBase.java:633)
[dal.jar:]
at 
org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsName(AuditLogableBase.java:504)
[dal.jar:]
...
2016-10-25 11:53:52,837 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
(org.ovirt.thread.pool-6-thread-2) [5e6a88be] Failed to get vds
'd60db21f-95f0-487b-9f17-44861e2610a7', error: null
2016-10-25 11:53:52,837 DEBUG
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
(org.ovirt.thread.pool-6-thread-2) [5e6a88be] Exception:
java.lang.NullPointerException
at 
org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsStatic(AuditLogableBase.java:633)
[dal.jar:]
at 
org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsName(AuditLogableBase.java:504)
[dal.jar:]
...

The full engine logs can be found here[3] and the entire test suite
logs here[4].

Can anyone have a look?

Thanks,
Nadav.


[1] https://gerrit.ovirt.org/#/c/65198/
[2] 
https://github.com/oVirt/ovirt-system-tests/blob/master/basic_suite_master/test-scenarios/004_basic_sanity.py#L322
[3] 
http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2759/artifact/exported-artifacts/basic_suite_master.sh-fc24/exported-artifacts/test_logs/basic_suite_master/post-004_basic_sanity.py/*zip*/post-004_basic_sanity.py.zip
[4] 
http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2759/artifact/exported-artifacts/basic_suite_master.sh-fc24/exported-artifacts/
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ** PROBLEM Service Alert: ovirt-mirrorchecker/mirrors.ibiblio.org/ovirt/pub mirror site last sync is CRITICAL **

2016-10-26 Thread Nadav Goldin
Working again since yesterday,

Thanks!

Nadav.


On Mon, Oct 24, 2016 at 10:46 PM, Cristóbal Palmer  wrote:
> We have an NFS issue that I'm trying to debug with our storage guy. It
> seems to have started on Friday morning and is impacting both our
> login host and the host that does rsync jobs for mirrors. Please check
> back with us if this is not resolved by 9am Wednesday.
>
> Thanks!
> CMP
>
> On Mon, Oct 24, 2016 at 2:20 PM, Nadav Goldin  wrote:
>> Hi,
>> It seems like the oVirt mirror at ibiblio.org has not been
>> synchronized in the past 2 days, from the logs I see there was an
>> attempted rsync connection but it took way longer than expected(10
>> hours), so I assume it hanged and eventually failed.
>>
>> Can you please have a look?
>>
>> Thanks,
>>
>> Nadav.
>>
>>
>> -- Forwarded message --
>> From: icinga 
>> Date: Mon, Oct 24, 2016 at 8:14 PM
>> Subject: ** PROBLEM Service Alert:
>> ovirt-mirrorchecker/mirrors.ibiblio.org/ovirt/pub mirror site last
>> sync is CRITICAL **
>> To: ngol...@redhat.com
>>
>>
>> * Icinga *
>>
>> Notification Type: PROBLEM
>>
>> Service: mirrors.ibiblio.org/ovirt/pub mirror site last sync
>> Host: ovirt-mirrorchecker
>> Address: 66.187.230.105
>> State: CRITICAL
>>
>> Date/Time: Mon Oct 24 17:14:07 UTC 2016
>>
>> Additional Info:
>>
>> CRITICAL - 432040 seconds since last sync, which are 120.0111 hours.
>
>
>
> --
> Cristóbal Palmer
> ibiblio.org
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-devel] Failure on start VM in ovirt-system-tests from patches merged to master on the 25/10/2016

2016-10-26 Thread Nadav Goldin
Unfortunately it is still failing, see[1], the repository used was
built from commit [2]. If you want to check the logs same links
apply(just replace build number 2759->2782)



[1] http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2782
[2] https://gerrit.ovirt.org/#/c/65740/

On Wed, Oct 26, 2016 at 11:33 AM, Tal Nisan  wrote:
> Should be fixed now on master as those patches were just merged:
>
> https://gerrit.ovirt.org/#/c/65738/ - "Move InjectorRule to VdsBroker":
> aligns the InjectorRule's package with the Injector so it can essentially be
> used wherever the Injector is used [+12, -4]
> https://gerrit.ovirt.org/#/c/65739 - "InjectorRule: override injector
> anyway": Fixes a bug in InjectorRule where the Injector is only mocked if
> you call InjectorRule.bind [+6, -1]
> https://gerrit.ovirt.org/#/c/65740 - "core: InjectorRule for injecting
> members": Fixes a bug in InjectorRule to allow using Injector.injectMembers
> when using it [+18, -5]
> https://gerrit.ovirt.org/#/c/65725 - "core: Fix AuditLogging": The actual
> fix. Basically, goes over all the places that create an AuditLoggableBase
> that needs injecting and take care of it [+155, -160]
>
>
> On Wed, Oct 26, 2016 at 10:04 AM, Nadav Goldin  wrote:
>>
>> Hi,
>> We have new failure on OST from patches merged to master yesterday,
>> the failure started after the merge of [1], but as there were quite a
>> few patches merged quickly I can't make sure it is the one causing
>> it(OST aren't ran per-patch).
>> The test that fails is [2] when attempting to start the VM.
>>
>> The error from the API side:
>>
>> RequestError:
>> status: 500
>> reason: Internal Server Error
>> detail: javax.ejb.EJBException: java.lang.NullPointerException
>> at
>> org.jboss.as.ejb3.tx.CMTTxInterceptor.handleExceptionInNoTx(CMTTxInterceptor.java:213)
>> at
>> org.jboss.as.ejb3.tx.CMTTxInterceptor.invokeInNoTx(CMTTxInterceptor.java:265)
>> at
>> org.jboss.as.ejb3.tx.CMTTxInterceptor.supports(CMTTxInterceptor.java:374)
>> at
>> org.jboss.as.ejb3.tx.CMTTxInterceptor.processInvocation(CMTTxInterceptor.java:243)
>> at
>> org.jboss.invocation.InterceptorContext.proceed(InterceptorContext.java:340)
>> 
>>
>> In the engine logs there are a few 'java.lang.NullPointerException'
>> errors:
>>
>> 2016-10-25 11:53:52,845 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
>> (org.ovirt.thread.pool-6-thread-2) [5e6a88be] Failed to get vds
>> 'd60db21f-95f0-487b-9f17-44861e2610a7', error: null
>> 2016-10-25 11:53:52,864 DEBUG
>> [org.ovirt.engine.core.utils.timer.FixedDelayJobListener]
>> (DefaultQuartzScheduler5) [] Rescheduling
>>
>> DEFAULT.org.ovirt.engine.core.bll.tasks.AsyncTaskManager.timerElapsed#-9223372036854775787
>> as there is no unfired trigger.
>> ...
>> 2016-10-25 11:53:52,845 DEBUG
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
>> (org.ovirt.thread.pool-6-thread-2) [5e6a88be] Exception:
>> java.lang.NullPointerException
>> at
>> org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsStatic(AuditLogableBase.java:633)
>> [dal.jar:]
>> at
>> org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsName(AuditLogableBase.java:504)
>> [dal.jar:]
>> ...
>> 2016-10-25 11:53:52,837 INFO
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
>> (org.ovirt.thread.pool-6-thread-2) [5e6a88be] Failed to get vds
>> 'd60db21f-95f0-487b-9f17-44861e2610a7', error: null
>> 2016-10-25 11:53:52,837 DEBUG
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
>> (org.ovirt.thread.pool-6-thread-2) [5e6a88be] Exception:
>> java.lang.NullPointerException
>> at
>> org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsStatic(AuditLogableBase.java:633)
>> [dal.jar:]
>> at
>> org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase.getVdsName(AuditLogableBase.java:504)
>> [dal.jar:]
>> ...
>>
>> The full engine logs can be found here[3] and the entire test suite
>> logs here[4].
>>
>> Can anyone have a look?
>>
>> Thanks,
>> Nadav.
>>
>>
>> [1] https://gerrit.ovirt.org/#/c/65198/
>> [2]
>> https://github.com/oVirt/ovirt-system-tests/blob/master/basic_suite_master/test-scenarios/004_basic_sanity.py#L322
>> [3]
>> http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2759/artifact/exported-artifacts/basic_suite_master.sh-fc24/exported-artifacts/test_logs/basic_suite_master/post-004_basic_sanity.py/*zip*/post-004_basic_sanity.py.zip
>> [4]
>> http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2759/artifact/exported-artifacts/basic_suite_master.sh-fc24/exported-artifacts/
>> ___
>> Devel mailing list
>> de...@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/devel
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: building engine artificats from a posted patch?

2016-10-26 Thread Nadav Goldin
Hi,
first you can run it locally quite easily using mock[1], the command
should be(after jenkins repo is cloned and mock installed) something
like:
../jenkins/mock_configs/mock_runner.sh --mock-confs-dir
../jenkins/mock_configs/ --build-only -s fc24
After running successfully the artifacts will be under
exported-artifacts directory.

It is possible to do it from Jenkins too, the problem is that the
current _build_artifacts job also deploy the created RPMs to
resources.ovirt.org's experimental repo, which is later consumed by
OST.
If needed, we can clone the needed job and remove the deploy part(and
add -manual suffix), then you can pass the gerrit refspec in the build
parameters. If so, tell me which job.

Nadav.


[1] 
http://ovirt-infra-docs.readthedocs.io/en/latest/CI/Build_and_test_standards.html#testing-the-scripts-locally

On Wed, Oct 26, 2016 at 5:47 PM, Dan Kenigsberg  wrote:
> Hi,
>
> Pardon my ignorance, but how can I trigger build-artifacts.sh after
> posting a patch to gerrit?
>
> I hope there's an easy way togenerate RPMs to be tested by third
> parties prior to merging the patch.
>
> Regards,
> Dan.
> ___
> Infra mailing list
> Infra@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/infra
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-devel] Failure on start VM in ovirt-system-tests from patches merged to master on the 25/10/2016

2016-10-26 Thread Nadav Goldin
Its running now:
http://jenkins.ovirt.org/job/ovirt-engine_master_build-artifacts-fc24-x86_64/1037/,
when it finishes it will trigger the deploy job which will trigger the
experimental job. The build_artifacts job is triggered approximately
every 2 hours(depends on the load).
Inside the build-artifacts job you can see the last
commit(https://gerrit.ovirt.org/65776, which is already after your
commit)


On Wed, Oct 26, 2016 at 9:26 PM, Allon Mureinik  wrote:
> @Infra - the last experimental job I saw was from ~17:30 local Israel time.
> Any idea why another one isn't being triggered (or am I just being daft)?
>
> On Wed, Oct 26, 2016 at 6:27 PM, Allon Mureinik  wrote:
>>
>> Yipes.
>> [1] should fix that, waiting for the CI to run to merge.
>>
>> [1] https://gerrit.ovirt.org/#/c/65768/
>>
>> On Wed, Oct 26, 2016 at 3:42 PM, Nadav Goldin  wrote:
>>>
>>> Unfortunately it is still failing, see[1], the repository used was
>>> built from commit [2]. If you want to check the logs same links
>>> apply(just replace build number 2759->2782)
>>>
>>>
>>>
>>> [1] http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_master/2782
>>> [2] https://gerrit.ovirt.org/#/c/65740/
>>>
>>> On Wed, Oct 26, 2016 at 11:33 AM, Tal Nisan  wrote:
>>> > Should be fixed now on master as those patches were just merged:
>>> >
>>> > https://gerrit.ovirt.org/#/c/65738/ - "Move InjectorRule to VdsBroker":
>>> > aligns the InjectorRule's package with the Injector so it can
>>> > essentially be
>>> > used wherever the Injector is used [+12, -4]
>>> > https://gerrit.ovirt.org/#/c/65739 - "InjectorRule: override injector
>>> > anyway": Fixes a bug in InjectorRule where the Injector is only mocked
>>> > if
>>> > you call InjectorRule.bind [+6, -1]
>>> > https://gerrit.ovirt.org/#/c/65740 - "core: InjectorRule for injecting
>>> > members": Fixes a bug in InjectorRule to allow using
>>> > Injector.injectMembers
>>> > when using it [+18, -5]
>>> > https://gerrit.ovirt.org/#/c/65725 - "core: Fix AuditLogging": The
>>> > actual
>>> > fix. Basically, goes over all the places that create an
>>> > AuditLoggableBase
>>> > that needs injecting and take care of it [+155, -160]
>>> >
>>> >
>>> > On Wed, Oct 26, 2016 at 10:04 AM, Nadav Goldin 
>>> > wrote:
>>> >>
>>> >> Hi,
>>> >> We have new failure on OST from patches merged to master yesterday,
>>> >> the failure started after the merge of [1], but as there were quite a
>>> >> few patches merged quickly I can't make sure it is the one causing
>>> >> it(OST aren't ran per-patch).
>>> >> The test that fails is [2] when attempting to start the VM.
>>> >>
>>> >> The error from the API side:
>>> >>
>>> >> RequestError:
>>> >> status: 500
>>> >> reason: Internal Server Error
>>> >> detail: javax.ejb.EJBException: java.lang.NullPointerException
>>> >> at
>>> >>
>>> >> org.jboss.as.ejb3.tx.CMTTxInterceptor.handleExceptionInNoTx(CMTTxInterceptor.java:213)
>>> >> at
>>> >>
>>> >> org.jboss.as.ejb3.tx.CMTTxInterceptor.invokeInNoTx(CMTTxInterceptor.java:265)
>>> >> at
>>> >>
>>> >> org.jboss.as.ejb3.tx.CMTTxInterceptor.supports(CMTTxInterceptor.java:374)
>>> >> at
>>> >>
>>> >> org.jboss.as.ejb3.tx.CMTTxInterceptor.processInvocation(CMTTxInterceptor.java:243)
>>> >> at
>>> >>
>>> >> org.jboss.invocation.InterceptorContext.proceed(InterceptorContext.java:340)
>>> >> 
>>> >>
>>> >> In the engine logs there are a few 'java.lang.NullPointerException'
>>> >> errors:
>>> >>
>>> >> 2016-10-25 11:53:52,845 INFO
>>> >> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogableBase]
>>> >> (org.ovirt.thread.pool-6-thread-2) [5e6a88be] Failed to get vds
>>> >> 'd60db21f-95f0-487b-9f17-44861e2610a7', error: null
>>> >> 2016-10-25 11:53:52,864 DEBUG
>>> >> [org.ovirt.engine.core.utils.timer.FixedDelayJobListener]
>>> >> (DefaultQuartzScheduler5) [] Rescheduling

Re: building engine artificats from a posted patch?

2016-10-27 Thread Nadav Goldin
On Thu, Oct 27, 2016 at 1:32 PM, Martin Mucha  wrote:
> 1. if I opted to run these tests locally, what are expected hw specification? 
> I mean devel build is already more than laptop can handle. If this has 
> enabled all translations, I'd have to take a pto to run it. So is this even 
> possible to be ran on laptop with only 12G ram?
I didn't do it on ovirt-engine myself for a while, but last time I
checked, it was possible to run the build with 12gb of ram. The slave
VMs in Jenkins usually have less than that(8GB), though you are right
it would take more than a hour.

>2. Since I probably should be coding instead of waiting for build on 
>irresponsible laptop (which it is even for devel build), would it be possible 
>to have jenkins build, which prepares rpms as described above without need to 
>deploy them to some repo, but allowing to download them instead?
As I wrote in the previous message, sure, just tell me which
_build_artifacts job you need and I'll clone them, removing the
deployment stage. To create the RPMs you'd trigger the job, and when
completed download the exported-artifacts from there.

On a side note, you will need to make sure whoever you pass the RPMs
to has the needed repositories with the dependencies(besides the
created RPMs..), probably exploring OST dependencies, for example[1]
will help.

Nadav.

[1] 
https://github.com/oVirt/ovirt-system-tests/blob/master/common/yum-repos/ovirt-4.0.repo

On Thu, Oct 27, 2016 at 1:54 PM, Roy Golan  wrote:
>
>
> On 27 October 2016 at 13:32, Martin Mucha  wrote:
>>
>> Hi,
>>
>> let me step back a little and explain what we want to achieve. We have
>> patch pushed to gerrit, not merged to master. We want to build rpms from it
>> and pass it (via no official way) to some tester so that he can test it.
>>
>> I read provided documentation, but I do not have sufficient background to
>> understand it fully.
>> Questions:
>>
>> 1. if I opted to run these tests locally, what are expected hw
>> specification? I mean devel build is already more than laptop can handle. If
>> this has enabled all translations, I'd have to take a pto to run it. So is
>> this even possible to be ran on laptop with only 12G ram?
>>
> PTO isn't needed if your tester doesn't need all locales + permutations
>
>>
>> 2. Since I probably should be coding instead of waiting for build on
>> irresponsible laptop (which it is even for devel build), would it be
>> possible to have jenkins build, which prepares rpms as described above
>> without need to deploy them to some repo, but allowing to download them
>> instead?
>>
>
> You can already explore the CI jenkins jobs on how to do it or have a look
> at this jenkins job
> http://jenkins-dev.eng.lab.tlv.redhat.com/job/ovirt-engine-create-rpms/
>
>>
>> thanks,
>> M.
>>
>> - Original Message -
>> > Hi,
>> > first you can run it locally quite easily using mock[1], the command
>> > should be(after jenkins repo is cloned and mock installed) something
>> > like:
>> > ../jenkins/mock_configs/mock_runner.sh --mock-confs-dir
>> > ../jenkins/mock_configs/ --build-only -s fc24
>> > After running successfully the artifacts will be under
>> > exported-artifacts directory.
>> >
>> > It is possible to do it from Jenkins too, the problem is that the
>> > current _build_artifacts job also deploy the created RPMs to
>> > resources.ovirt.org's experimental repo, which is later consumed by
>> > OST.
>> > If needed, we can clone the needed job and remove the deploy part(and
>> > add -manual suffix), then you can pass the gerrit refspec in the build
>> > parameters. If so, tell me which job.
>> >
>> > Nadav.
>> >
>> >
>> > [1]
>> >
>> > http://ovirt-infra-docs.readthedocs.io/en/latest/CI/Build_and_test_standards.html#testing-the-scripts-locally
>> >
>> > On Wed, Oct 26, 2016 at 5:47 PM, Dan Kenigsberg 
>> > wrote:
>> > > Hi,
>> > >
>> > > Pardon my ignorance, but how can I trigger build-artifacts.sh after
>> > > posting a patch to gerrit?
>> > >
>> > > I hope there's an easy way togenerate RPMs to be tested by third
>> > > parties prior to merging the patch.
>> > >
>> > > Regards,
>> > > Dan.
>> > > ___
>> > > Infra mailing list
>> > > Infra@ovirt.org
>> > > http://lists.ovirt.org/mailman/listinfo/infra
>> >
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: building engine artificats from a posted patch?

2016-10-27 Thread Nadav Goldin
The options are combination of the distro and branch, I added the
job[1](el7, 4.0), to use it press 'build' from Jenkins's GUI and fill
in the GERRIT_REFSPEC to the needed patch.
As the build-artifacts job run on the limited resources of bare metal
hypervisors I restricted it for now to one concurrent job.

On infra side, I guess this can be yamlized later if it becomes
useful(just copy-paste of the current build-artifacts removing some
stages).


[1] 
http://jenkins.ovirt.org/job/ovirt-engine_4.0_build-artifacts-el7-x86_64-manual/


On Thu, Oct 27, 2016 at 3:18 PM, Dan Kenigsberg  wrote:
> On Thu, Oct 27, 2016 at 06:32:47AM -0400, Martin Mucha wrote:
>> Hi,
>>
>> let me step back a little and explain what we want to achieve. We have patch 
>> pushed to gerrit, not merged to master. We want to build rpms from it and 
>> pass it (via no official way) to some tester so that he can test it.
>>
>> I read provided documentation, but I do not have sufficient background to 
>> understand it fully.
>> Questions:
>>
>> 1. if I opted to run these tests locally, what are expected hw 
>> specification? I mean devel build is already more than laptop can handle. If 
>> this has enabled all translations, I'd have to take a pto to run it. So is 
>> this even possible to be ran on laptop with only 12G ram?
>>
>> 2. Since I probably should be coding instead of waiting for build on 
>> irresponsible laptop (which it is even for devel build), would it be 
>> possible to have jenkins build, which prepares rpms as described above 
>> without need to deploy them to some repo, but allowing to download them 
>> instead?
>>
>> thanks,
>> M.
>>
>> - Original Message -
>> > Hi,
>> > first you can run it locally quite easily using mock[1], the command
>> > should be(after jenkins repo is cloned and mock installed) something
>> > like:
>> > ../jenkins/mock_configs/mock_runner.sh --mock-confs-dir
>> > ../jenkins/mock_configs/ --build-only -s fc24
>> > After running successfully the artifacts will be under
>> > exported-artifacts directory.
>> >
>> > It is possible to do it from Jenkins too, the problem is that the
>> > current _build_artifacts job also deploy the created RPMs to
>> > resources.ovirt.org's experimental repo, which is later consumed by
>> > OST.
>> > If needed, we can clone the needed job and remove the deploy part(and
>> > add -manual suffix), then you can pass the gerrit refspec in the build
>> > parameters. If so, tell me which job.
>
> Adding to Matin's explaination: He posted https://gerrit.ovirt.org/65793
> and would like it be tested. It would be wonderful if you can add a job
> that makes it possible to build el7 rpms from that patch, to bet
> executed by QE.
>
> So yes, I'd appreciate if you can add such a -manual job for building
> ovirt-engine. I'm not sure I know "which job" precisely is that, though.
> Can you tell me which are the options?
>
> Regrards,
> Dan.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [ovirt-devel] Failure on start VM in ovirt-system-tests from patches merged to master on the 25/10/2016

2016-10-29 Thread Nadav Goldin
;>>> [weld-core-impl-2.3.5.Final.jar:2.3.5.Final]
>>>>>>> at
>>>>>>> org.jboss.weld.bean.proxy.CombinedInterceptorAndDecoratorStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:79)
>>>>>>> [weld-core-impl-2.3.5.Final.jar:2.3.5.Final]
>>>>>>> at
>>>>>>> org.jboss.weld.bean.proxy.CombinedInterceptorAndDecoratorStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:68)
>>>>>>> [weld-core-impl-2.3.5.Final.jar:2.3.5.Final]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.scheduling.SchedulingManager$Proxy$_$$_WeldSubclass.schedule(Unknown
>>>>>>> Source) [bll.jar:]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.getVdsToRunOn(RunVmCommand.java:818)
>>>>>>> [bll.jar:]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.runVm(RunVmCommand.java:231)
>>>>>>> [bll.jar:]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.perform(RunVmCommand.java:414)
>>>>>>> [bll.jar:]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.RunVmCommand.executeVmCommand(RunVmCommand.java:339)
>>>>>>> [bll.jar:]
>>>>>>> at
>>>>>>> org.ovirt.engine.core.bll.VmCommand.executeCommand(VmCommand.java:106)
>>>>>>> [bll.jar:]
>>>>>>>
>>>>>>>
>>>>>>> On Thu, Oct 27, 2016 at 5:12 AM, Allon Mureinik 
>>>>>>> wrote:
>>>>>>>>
>>>>>>>> Seems like we cleared up the engine issues related to the recent
>>>>>>>> injection changes.
>>>>>>>>
>>>>>>>> I am now seeing stop errors, e.g.:
>>>>>>>>
>>>>>>>> {"jsonrpc": "2.0", "id": "ea0c564f-bc17-4fc2-8f1b-67c4d28257c6",
>>>>>>>> "result": {"cpuStatistics": {"1": {"cpuUser": "3.07", "nodeIndex": 0,
>>>>>>>> "cpuSys": "3.00", "cpuIdle": "93.93"}, "0": {"cpuUser": "1.67", 
>>>>>>>> "nodeIndex":
>>>>>>>> 0, "cpuSys": "2.07", "cpuIdle": "96.26"}}, "numaNodeMemFree": {"0":
>>>>>>>> {"memPercent": 83, "memFree": "359"}}, "memShared": 0, "thpState": 
>>>>>>>> "always",
>>>>>>>> "ksmMergeAcrossNodes": true, "vmCount": 0, "memUsed": "20",
>>>>>>>> "storageDomains": {"b2bb3220-1eb3-426a-90c2-5e236aefbe1a": {"code": 0,
>>>>>>>> "actual": true, "version": 0, "acquired": true, "delay": "0.000840117",
>>>>>>>> "lastCheck": "7.1", "valid": true}, 
>>>>>>>> "3130195a-73f9-4490-b554-98a9205cead6":
>>>>>>>> {"code": 0, "actual": true, "version": 4, "acquired": true, "delay":
>>>>>>>> "0.00150771", "lastCheck": "7.5", "valid": true},
>>>>>>>> "1a9e202b-83b7-4bdc-9b0c-e76b83676068": {"code": 0, "actual": true,
>>>>>>>> "version": 4, "acquired": true, "delay": "0.000590956",
>>>>>>>> 2016-10-26 21:51:09,878 DEBUG
>>>>>>>> [org.ovirt.engine.core.utils.timer.FixedDelayJobListener]
>>>>>>>> (DefaultQuartzScheduler7) [6d206bd1] Rescheduling
>>>>>>>> DEFAULT.org.ovirt.engine.core.bll.tasks.CommandCallbacksPoller.invokeCallbackMethods#-9223372036854775783
>>>>>>>> as there is no unfired trigger.
>>>>>>>> 2016-10-26 21:51:28,705 DEBUG
>>>>>>>> [org.ovirt.vdsm.jsonrpc.client.reactors.ReactorClient] (SSL Stomp 
>>>>>>>> Reactor)
>>>>>>>> [383dd6a0] Heartbeat exceeded. C

Re: Build failed in Jenkins: ovirt_4.0_he-system-tests #464

2016-11-02 Thread Nadav Goldin
Hi Lev,
This is due to an on-going process to remove underscores in the host
name, and is addressed in [1].

[1] https://gerrit.ovirt.org/#/c/64465/

On Wed, Nov 2, 2016 at 10:53 AM, Lev Veyde  wrote:
> Hi Gal,
>
> It seems that someone changed the translation algo. between the host name 
> defined in the conf. and one that will be used.
>
> Note:
>
> 07:47:04 Deploying on host0
> 07:47:04 + lago copy-to-vm lago-he-basic-suite-4-0-host0 
> /home/jenkins/workspace/ovirt_4.0_he-system-tests/ovirt-system-tests/he-basic-suite-4.0/answers.conf.in
>  /root/hosted-engine-deploy-answers-file.conf.in
> 07:47:05 current session does not belong to lago group.
> 07:47:05 Unable to find VM lago-he-basic-suite-4-0-host0
> 07:47:05 Available VMs:
> 07:47:05lago_he-basic-suite-4_0-engine
> 07:47:05lago-he-basic-suite-4_0-host0
> 07:47:05lago-he-basic-suite-4_0-storage
> 07:47:05lago-he-basic-suite-4_0-host1
> 07:47:05 Error occured, aborting
> 07:47:05 Traceback (most recent call last):
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/cmd.py", line 691, in 
> main
> 07:47:05 cli_plugins[args.verb].do_run(args)
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/plugins/cli.py", line 
> 180, in do_run
> 07:47:05 self._do_run(**vars(args))
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/utils.py", line 488, 
> in wrapper
> 07:47:05 return func(*args, **kwargs)
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/utils.py", line 499, 
> in wrapper
> 07:47:05 return func(*args, prefix=prefix, **kwargs)
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/cmd.py", line 539, in 
> do_copy_to_vm
> 07:47:05 host = prefix.virt_env.get_vm(host)
> 07:47:05   File "/usr/lib/python2.7/site-packages/lago/virt.py", line 229, in 
> get_vm
> 07:47:05 return self._vms[name]
> 07:47:05 KeyError: 'lago-he-basic-suite-4-0-host0'
>
> Thanks in advance,
> Lev Veyde.
>
> - Original Message -
> From: jenk...@jenkins.phx.ovirt.org
> To: sbona...@redhat.com, infra@ovirt.org, lve...@redhat.com
> Sent: Wednesday, November 2, 2016 10:05:34 AM
> Subject: Build failed in Jenkins: ovirt_4.0_he-system-tests #464
>
> See 
>
> Changes:
>
> [Gal Ben Haim] Switched the mock automation scripts to symlinks
>
> [Ryan Barry] cockpit-ovirt: add 4.1 jobs
>
> --
> [...truncated 966 lines...]
> ##  rc = 1
> ##
> ##! ERROR v
> ##! Last 20 log entries: 
> logs/mocker-fedora-23-x86_64.fc23.he_basic_suite_4.0.sh/he_basic_suite_4.0.sh.log
> ##!
> + env_cleanup
> + echo '#'
> #
> + local res=0
> + local uuid
> + echo ' Cleaning up'
>  Cleaning up
> + [[ -e 
> 
>  ]]
> + echo '--- Cleaning with lago'
> --- Cleaning with lago
> + lago --workdir 
> 
>  destroy --yes --all-prefixes
> + echo '--- Cleaning with lago done'
> --- Cleaning with lago done
> + [[ 0 != \0 ]]
> + echo ' Cleanup done'
>  Cleanup done
> + exit 0
> + exit
> Took 1616 seconds
> ===
> ##!
> ##! ERROR ^^
> ##!
> ##
> Build step 'Execute shell' marked build as failure
> Performing Post build task...
> Match found for :.* : True
> Logical operation result is TRUE
> Running script  : #!/bin/bash -xe
> echo 'shell_scripts/system_tests.collect_logs.sh'
>
> #
> # Required jjb vars:
> #version
> #
> VERSION=4.0
> SUITE_TYPE=
>
> WORKSPACE="$PWD"
> OVIRT_SUITE="$SUITE_TYPE_suite_$VERSION"
> TESTS_LOGS="$WORKSPACE/ovirt-system-tests/exported-artifacts"
>
> rm -rf "$WORKSPACE/exported-artifacts"
> mkdir -p "$WORKSPACE/exported-artifacts"
>
> if [[ -d "$TESTS_LOGS" ]]; then
> mv "$TESTS_LOGS/"* "$WORKSPACE/exported-artifacts/"
> fi
>
> [ovirt_4.0_he-system-tests] $ /bin/bash -xe /tmp/hudson6985408007367819294.sh
> + echo shell_scripts/system_tests.collect_logs.sh
> shell_scripts/system_tests.collect_logs.sh
> + VERSION=4.0
> + SUITE_TYPE=
> + WORKSPACE=
> + OVIRT_SUITE=4.0
> + 
> TESTS_LOGS=
> + rm -rf 
> 
> + mkdir -p 
> 
> + [[ -d 
> 

OST 3.6 branch failing due to mismatch vdsm-jsonrpc-java version

2016-11-04 Thread Nadav Goldin
Hi Piotr,
OST on 3.6 branch is failing due to[1]:

03:50:01 Error: Package:
ovirt-engine-backend-3.6.10-0.0.master.20161104223702.git9d234e8.el6.noarch
(alocalsync)
03:50:01Requires: vdsm-jsonrpc-java >= 1.1.14
03:50:01Available: vdsm-jsonrpc-java-1.1.13-1.el6.noarch
(alocalsync)
03:50:01vdsm-jsonrpc-java = 1.1.13-1.el6

Has there been any package changes? maybe a caching issue?


Thanks,

Nadav.


[1] http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_3.6/3297/
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is CRITICAL **

2016-11-15 Thread Nadav Goldin
Hi,
the oVirt mirror site stopped synchronising again few days ago, can
you please check?

Thanks,
Nadav.


On Tue, Nov 15, 2016 at 11:56 AM, icinga  wrote:
> * Icinga *
>
> Notification Type: PROBLEM
>
> Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
> Host: ovirt-mirrorchecker
> Address: 66.187.230.105
> State: CRITICAL
>
> Date/Time: Tue Nov 15 09:56:58 UTC 2016
>
> Additional Info:
>
> CRITICAL - 713972 seconds since last sync, which are 198.3256 hours.
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is CRITICAL **

2016-11-20 Thread Nadav Goldin
Hi Neil,
I've opened a request, will update once the IP is added.

Thanks,
Nadav.


On Tue, Nov 15, 2016 at 7:05 PM, Neil Bright  wrote:
> Hi Nadav,
>
> It looks like I’ve run afowl of some upgrades on my end.  Would you mind 
> adding zaphod2 to the ACL I’m allowed to sync from?  Meanwhile, I’ve kicked 
> off a sync from zaphod, which should already be in the ACL.
>
> Thanks!
>
>
> zaphod2.gtlib.gatech.edu has address 128.61.111.20
> zaphod.gtlib.gatech.edu has address 128.61.111.12
>
>
>
>> On Nov 15, 2016, at 3:39 AM, Nadav Goldin  wrote:
>>
>> Hi,
>> the oVirt mirror site stopped synchronising again few days ago, can
>> you please check?
>>
>> Thanks,
>> Nadav.
>>
>>
>> On Tue, Nov 15, 2016 at 11:56 AM, icinga  wrote:
>>> * Icinga *
>>>
>>> Notification Type: PROBLEM
>>>
>>> Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
>>> Host: ovirt-mirrorchecker
>>> Address: 66.187.230.105
>>> State: CRITICAL
>>>
>>> Date/Time: Tue Nov 15 09:56:58 UTC 2016
>>>
>>> Additional Info:
>>>
>>> CRITICAL - 713972 seconds since last sync, which are 198.3256 hours.
>
> +==+
> Neil Bright (ncbri...@gatech.edu)  (404) 385-6954
> http://www.pace.gatech.edu
> 258 Fourth Street, Rich Bldg, Rm 321 / Atlanta, GA  30332-0700
>
>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


[FYI] - new network tests landed in OST

2016-11-21 Thread Nadav Goldin
Hi,
I just merged lgoldber's patch[1] to OST, this are new network tests
which will appear as '003_basic_networking.py' in the logs and would
run only on the master suite.

Thanks,
Nadav.

[1] https://gerrit.ovirt.org/#/c/66055/
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: [FYI] - new network tests landed in OST

2016-11-21 Thread Nadav Goldin
It ran yesterday here[1]

16:45:05 @ Run test: 003_basic_networking.py:
16:45:05 nose.config: INFO: Ignoring files matching ['^\\.', '^_',
'^setup\\.py$']
16:45:05   # attach_vlan_to_host:
16:45:14   # attach_vlan_to_host: Success (in 0:00:09)
16:45:14   # detach_vlan_from_host:
16:45:18   # detach_vlan_from_host: Success (in 0:00:04)
16:45:18   # Results located at
/home/jenkins/workspace/ovirt-system-tests_master_check-patch-fc24-x86_64/ovirt-system-tests/deployment-basic-suite-master/default/nosetests-003_basic_networking.py.xml
16:45:18 @ Run test: 003_basic_networking.py: Success (in 0:00:13)


[1] 
http://jenkins.ovirt.org/job/ovirt-system-tests_master_check-patch-fc24-x86_64/240/console


On Mon, Nov 21, 2016 at 11:36 AM, Eyal Edri  wrote:
> Did you verify it?
> I just run it on [1]. I think it didn't run CI because the suite names in
> check-patch.sh weren't updated until yesterday.
>
>
> [1]
> http://jenkins.ovirt.org/view/oVirt%20system%20tests/job/ovirt_master_system-tests/817/console
>
> On Mon, Nov 21, 2016 at 11:04 AM, Nadav Goldin  wrote:
>>
>> Hi,
>> I just merged lgoldber's patch[1] to OST, this are new network tests
>> which will appear as '003_basic_networking.py' in the logs and would
>> run only on the master suite.
>>
>> Thanks,
>> Nadav.
>>
>> [1] https://gerrit.ovirt.org/#/c/66055/
>> ___
>> Infra mailing list
>> Infra@ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/infra
>>
>>
>
>
>
> --
> Eyal Edri
> Associate Manager
> RHV DevOps
> EMEA ENG Virtualization R&D
> Red Hat Israel
>
> phone: +972-9-7692018
> irc: eedri (on #tlv #rhev-dev #rhev-integ)
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


Re: ** PROBLEM Service Alert: ovirt-mirrorchecker/www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync is CRITICAL **

2016-11-21 Thread Nadav Goldin
The IP was white-listed, please test and tell if it works.

Thanks,
Nadav.


On Mon, Nov 21, 2016 at 8:15 AM, Nadav Goldin  wrote:
> Hi Neil,
> I've opened a request, will update once the IP is added.
>
> Thanks,
> Nadav.
>
>
> On Tue, Nov 15, 2016 at 7:05 PM, Neil Bright  wrote:
>> Hi Nadav,
>>
>> It looks like I’ve run afowl of some upgrades on my end.  Would you mind 
>> adding zaphod2 to the ACL I’m allowed to sync from?  Meanwhile, I’ve kicked 
>> off a sync from zaphod, which should already be in the ACL.
>>
>> Thanks!
>>
>>
>> zaphod2.gtlib.gatech.edu has address 128.61.111.20
>> zaphod.gtlib.gatech.edu has address 128.61.111.12
>>
>>
>>
>>> On Nov 15, 2016, at 3:39 AM, Nadav Goldin  wrote:
>>>
>>> Hi,
>>> the oVirt mirror site stopped synchronising again few days ago, can
>>> you please check?
>>>
>>> Thanks,
>>> Nadav.
>>>
>>>
>>> On Tue, Nov 15, 2016 at 11:56 AM, icinga  
>>> wrote:
>>>> * Icinga *
>>>>
>>>> Notification Type: PROBLEM
>>>>
>>>> Service: www.gtlib.gatech.edu/pub/oVirt/pub mirror site last sync
>>>> Host: ovirt-mirrorchecker
>>>> Address: 66.187.230.105
>>>> State: CRITICAL
>>>>
>>>> Date/Time: Tue Nov 15 09:56:58 UTC 2016
>>>>
>>>> Additional Info:
>>>>
>>>> CRITICAL - 713972 seconds since last sync, which are 198.3256 hours.
>>
>> +==+
>> Neil Bright (ncbri...@gatech.edu)  (404) 385-6954
>> http://www.pace.gatech.edu
>> 258 Fourth Street, Rich Bldg, Rm 321 / Atlanta, GA  30332-0700
>>
>>
___
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra


  1   2   3   >