9851-9900 of 10000 results (49ms)
2016-01-11 §
01:11 <paravoid> deactivating eqiad<->GTT BGP peering, reported network issues (P2469) [production]
2016-01-10 §
22:00 <gwicke> restbase: 1005-1009 now on node 4.2 [production]
19:44 <paravoid> powercycling mw1004, mw1008, mw1012 [production]
19:38 <paravoid> restarting hhvm on jobrunners again [production]
12:40 <mwdeploy@tin> mwdeploy@tin sync-l10n completed (1.27.0-wmf.9) (duration: 626m 20s) [production]
10:13 <ori> disabled categoryMembershipChange on mw1165 too, then restart jobrunner / jobchron / hhvm on mw1165 and mw1164 [production]
08:55 <ori> mw1166 -- disabled puppet; disabled categoryMembershipChange jobs [production]
08:48 <ori> mw1167 -- disabled puppet; disabled deleteLinks and refreshLinks* jobs [production]
08:45 <ori> mw1168 -- disabled puppet; disabled restbase jobs [production]
08:41 <ori> mw1169 -- disables cirrus jobs. [production]
08:33 <ori> Attempting to isolate cause of T122069 by toggling job types on mw1169. Disabling Puppet to prevent it from clobbering config changes. [production]
08:29 <paravoid> restarting hhvm on jobrunners again [production]
04:58 <paravoid> powercycling mw1005, mw1008, mw1009 -- unresponsive due to OOM [production]
04:56 <paravoid> restarting HHVM on eqiad jobrunners, OOM, memleak faster than the 24h restarts [production]
2016-01-09 §
21:07 <valhallasw`cloud> moved tools-checker/208.80.155.229 back to tools-checker-01 [tools]
21:02 <andrewbogott> rebooting tools-checker-01 as it is unresponsive. [tools]
13:12 <valhallasw`cloud> tools-worker-1002. is unresponsive. Maybe that's where the other grrrit-wm is hiding? Rebooting. [tools]
13:01 <valhallasw`cloud> grrrit-wm1 is handled by k8s, and thus restarts nicely. I can't figure out where the other one is running, though... [tools.lolrrit-wm]
12:58 <valhallasw`cloud> I don't see it running anywhere else either. Odd. Let's kill the single one I see... [tools.lolrrit-wm]
12:54 <valhallasw`cloud> running on tools-worker-1005.tools.eqiad.wmflabs according to k8s (kubectl --user=lolrrit-wm --namespace=lolrrit-wm describe pod grrrit-sm8km) [tools.lolrrit-wm]
12:52 <valhallasw`cloud> two instances running, k8s is only aware of one of them [tools.lolrrit-wm]
02:33 <l10nupdate@tin> l10nupdate@tin ResourceLoader cache refresh completed at Sat Jan 9 02:33:40 UTC 2016 (duration 6m 57s) [production]
02:26 <mwdeploy@tin> mwdeploy@tin sync-l10n completed (1.27.0-wmf.9) (duration: 11m 19s) [production]
2016-01-08 §
23:49 <RobH> stalled puppet on carbon for now, messing with partman files [production]
19:46 <chasemp> couldn't get into tools-mail-01 at all and it seemed borked so I rebooted [tools]
17:23 <andrewbogott> killing tools.icelab as per https://wikitech.wikimedia.org/wiki/User_talk:Torin#Running_queries_on_tools-dev_.28tools-bastion-02.29 [tools]
02:31 <l10nupdate@tin> l10nupdate@tin ResourceLoader cache refresh completed at Fri Jan 8 02:31:46 UTC 2016 (duration 7m 0s) [production]
02:24 <mwdeploy@tin> mwdeploy@tin sync-l10n completed (1.27.0-wmf.9) (duration: 10m 15s) [production]
2016-01-07 §
23:24 <akosiaris> repooled scb1002 for mobileapps [production]
23:24 <akosiaris> enabled puppet,salt on scb1001 [production]
23:23 <mobrovac> mobileapps deploying 58b371a on scb1001 [production]
23:16 <legoktm> deleted /mnt/jenkins-workspace/workspace/mediawiki-extensions-qunit/src/extensions/PdfHandler/.git/refs/heads/wmf/1.26wmf16.lock on slave 1013 [releng]
23:09 <mobrovac> mobileapps deploying 58b371a on scb1002 [production]
23:01 <akosiaris> apt-mark hold nodejs on scb1001, etherpad1001 and maps-test200{1,2,3,4} [production]
22:58 <akosiaris> disable puppet and salt on scb1001 from nodejs 4.2 transition [production]
22:57 <akosiaris> depool scb1002 for mobileapps. Transition to nodejs 4.2 ongoing [production]
20:53 <wikibugs> Updated channels.yaml to: db26b7db94db89a49fac63df54d0189cf39ffc90 Send Labs* to `#wikimedia-labs` [tools.wikibugs]
19:21 <YuviPanda> started tools / maps backup on labstore1001 [production]
19:13 <YuviPanda> remove snapshots others20150815030010, others20150815030010, maps20151216040005 and maps20151028040004 that were all stale and should've been removed anyway (on labstore2001) [production]
19:13 <YuviPanda> remove snapshots others20150815030010, others20150815030010, maps20151216040005 and maps20151028040004 that were all stale and should've been removed anyway [production]
19:11 <YuviPanda> run sudo lvremove backup/tools20151216020005 on labstore2001 to clean up full snapshot [production]
19:11 <jynus> setting up watchdog process killing long running queries on db1051 [production]
18:54 <_joe_> also resetting the drac [production]
18:53 <_joe_> powercycling ms-be1013 [production]
06:32 <legoktm> deploying https://gerrit.wikimedia.org/r/262868 [releng]
02:32 <l10nupdate@tin> l10nupdate@tin ResourceLoader cache refresh completed at Thu Jan 7 02:32:04 UTC 2016 (duration 6m 54s) [production]
02:25 <mwdeploy@tin> mwdeploy@tin sync-l10n completed (1.27.0-wmf.9) (duration: 10m 33s) [production]
02:24 <legoktm> deploying https://gerrit.wikimedia.org/r/262855 [releng]
01:25 <jzerebecki> reloading zuul for b0a5335..c16368a [releng]
2016-01-06 §
23:03 <gwicke> switched restbase1009 to node 4.2 for testing, and restarted restbase; see https://phabricator.wikimedia.org/T107762 [production]