201-250 of 2902 results (10ms)
2015-04-09 §
17:11 <Krinkle> Depool integration-slave1402...integration-slave1405 [releng]
16:52 <Krinkle> Pool integration-slave-trusty-1011...integration-slave-trusty-1016 [releng]
16:00 <hashar> integration-slave-jessie-1001 recreated. Applying it role::ci::slave::labs which should also bring in the package builder role under /mnt/pbuilder [releng]
14:24 <hashar> deleting integration-slave-jessie-1001 extended disk is too small [releng]
14:24 <hashar> deleting integration-slave-jessie-1001 extended disk is too smal [releng]
13:14 <hashar> integration-zuul-packaged applied role::labs::lvm::srv [releng]
13:01 <hashar> integration-zuul-packaged applied zuul::merger and zuul::server [releng]
12:59 <Krinkle> Creating integration-slave-trusty-1011 - integration-slave-trusty-1016 [releng]
12:40 <hashar> spurts out <tt>Permission denied (publickey).</tt> [releng]
12:39 <hashar> https://integration.wikimedia.org/ci/job/beta-scap-eqiad/ is still broken :-( [releng]
12:31 <hashar> beta: reset hard of operations/puppet repo on the puppetmaster since it has been stalled for 9+days https://phabricator.wikimedia.org/T95539 [releng]
10:46 <hashar> repacked extensions in deployment-bastion staging area: <tt>find /mnt/srv/mediawiki-staging/php-master/extensions -maxdepth 2 -type f -name .git -exec bash -c 'cd `dirname {}` && pwd && git repack -Ad && git gc' \\;</tt> [releng]
10:31 <hashar> deployment-bastion has a lock file remaining /mnt/srv/mediawiki-staging/php-master/extensions/.git/refs/remotes/origin/master.lock [releng]
09:55 <hashar> restarted Zuul to clear out some stalled jobs [releng]
09:35 <Krinkle> Pooled integration-slave-trusty-1010 [releng]
08:59 <hashar> rebooted deployment-bastion and cleared some files under /var/ [releng]
08:51 <hashar> deployment-bastion is out of disk space on /var/ :( [releng]
08:50 <hashar> https://integration.wikimedia.org/ci/job/beta-code-update-eqiad/ timed out after 30 minutes while trying to git pull [releng]
08:50 <hashar> https://integration.wikimedia.org/ci/job/beta-update-databases-eqiad/ job stalled for some reason [releng]
06:15 <legoktm> deploying https://gerrit.wikimedia.org/r/202998 [releng]
06:02 <legoktm> deploying https://gerrit.wikimedia.org/r/202992 [releng]
05:11 <legoktm> deleted core dumps from integration-slave1002, /var had filled up [releng]
04:36 <legoktm> deploying https://gerrit.wikimedia.org/r/202938 [releng]
00:32 <legoktm> deploying https://gerrit.wikimedia.org/r/202279 [releng]
2015-04-08 §
21:56 <legoktm> deploying https://gerrit.wikimedia.org/r/202930 [releng]
21:15 <legoktm> deleting non-existent jobs' workspaces on labs slaves [releng]
19:09 <Krinkle> Re-establishing Gearman-Jenkins connection [releng]
19:00 <Krinkle> Restarting Jenkins [releng]
19:00 <Krinkle> Jenkins Master unable to re-establish Gearman connection [releng]
19:00 <Krinkle> Zuul queue is not being distributed properly. Many slaves are idling waiting to receive builds but not getting any. [releng]
18:29 <Krinkle> Another attempt at re-creating the Trusty slave pool (T94916) [releng]
18:07 <legoktm> deploying https://gerrit.wikimedia.org/r/202289 and https://gerrit.wikimedia.org/r/202445 [releng]
18:01 <Krinkle> Jobs for Precise slaves are not starting. Stuck in Zuul as 'queued'. Disconnected and restarted slave agent on them. Queue is back up now. [releng]
17:36 <legoktm> deployed https://gerrit.wikimedia.org/r/180418 [releng]
13:32 <hashar> Disabled Zuul install based on git clone / setup.py by cherry picking https://gerrit.wikimedia.org/r/#/c/202714/ . Installed the Zuul debian package on all slaves [releng]
13:31 <hashar> integration: running <tt>apt-get upgrade</tt> on Trusty slaves [releng]
13:30 <hashar> integration: upgrading python-gear and python-six on Trusty slaves [releng]
12:43 <hasharLunch> Zuul is back and it is nasty [releng]
12:24 <hasharLunch> killed zuul on gallium :/ [releng]
2015-04-07 §
16:26 <Krinkle> git-deploy: Deploying integration/slave-scripts 4c6f541 [releng]
12:57 <hashar> running apt-get upgrade on integration-slave-trusty* hosts [releng]
12:45 <hashar> recreating integration-slave-trusty-1005 [releng]
12:26 <hashar> deleting integration-slave-trusty-1005 has been provisioned with role::ci::website instead of role::ci::slave::labs [releng]
12:11 <hashar> retriggering a bunch of browser tests hitting beta.wmflabs.org [releng]
12:07 <hashar> Puppet being fixed, it is finishing the installation of integration-slave-trusty-*** hosts [releng]
12:03 <hashar> Browser tests against beta cluster were all failing due to an improper DNS resolver being applied on CI labs instances {{bug|T95273}}. Should be fixed now. [releng]
12:00 <hashar> running puppet on all integration machines and resigning puppet client certs [releng]
11:31 <hashar> integration-puppetmaster is back and operational with local puppet client working properly. [releng]
11:28 <hashar> restored /etc/puppet/fileserver.conf [releng]
11:08 <hashar> dishing out puppet SSL configuration on all integratio nodes. Can't figure out so lets restart from scratch [releng]