2015-06-08
§
|
23:11 |
<ebernhardson> |
Synchronized php-1.26wmf8/extensions/UploadWizard/: Bump UploadWizard in 1.26wmf8 for evening SWAT (duration: 01m 09s) |
[production] |
22:21 |
<bblack> |
depooled cp3030, cp1065 in pybal for ipsec |
[production] |
20:17 |
<subbu> |
deployed parsoid sha 131554ba |
[production] |
19:18 |
<jynus> |
RAID degradation (disk failure) on s5 master (db1058), no production impact, replacement on the way |
[production] |
17:13 |
<ottomata> |
restarted eventlogging services on eventlog1001 after disabling kafka pieces |
[production] |
16:13 |
<_joe_> |
powercycling tmh1001, console blank, unresponsive to pings |
[production] |
16:00 |
<thcipriani> |
Synchronized commonsuploads.dblist: SWAT: Revert Temporarily re-enable uploads on Marathi Wikipedia, for real [[gerrit:216719]] (duration: 01m 07s) |
[production] |
15:58 |
<thcipriani> |
Synchronized commonsuploads.dblist: SWAT: Revert Temporarily re-enable uploads on Marathi Wikipedia [[gerrit:216719]] (duration: 01m 08s) |
[production] |
15:40 |
<thcipriani> |
Synchronized php-1.26wmf8/extensions/Cite: SWAT: Revert Do all of Cite's real work during unstrip and followup [[gerrit:216715]] (duration: 01m 08s) |
[production] |
15:19 |
<Coren> |
T96063: process halted for now as store/backup is unmovable and on slice5 |
[production] |
15:17 |
<thcipriani> |
Synchronized w/static/images/project-logos/pflwiki.png: SWAT: Fix transparency of pflwiki logo [[gerrit:216595]] (duration: 01m 08s) |
[production] |
15:15 |
<akosiaris> |
disabled ircecho on neon for a while |
[production] |
14:53 |
<Coren> |
T96063: starting pvmove from slice5 to slice2 |
[production] |
14:48 |
<Coren> |
T96063: dropped volume slice1 from vg store |
[production] |
14:46 |
<Coren> |
T96063: dropped store/project |
[production] |
14:44 |
<Coren> |
starting https://phabricator.wikimedia.org/T96063 on labstore1001 |
[production] |
14:24 |
<jynus> |
Synchronized wmf-config/db-eqiad.php: depool es1005 (duration: 01m 08s) |
[production] |
14:23 |
<Coren> |
rsync in progress between labstore1001:store/backup and labstore1002:backup/backup (at ionice idle) |
[production] |
14:13 |
<Coren> |
created store/backup snapshot on labstore1001 for backup copy |
[production] |
13:03 |
<moritzm> |
added strongswan_5.3.0-1+wmf2 to jessie-wikimedia on carbon |
[production] |
11:42 |
<_joe_> |
purging squid cache on carbon |
[production] |
11:26 |
<moritzm> |
updated mc2* to 2:2.8.17-1+deb8u1 |
[production] |
10:55 |
<jynus> |
Synchronized wmf-config/db-eqiad.php: repool es1007 (duration: 01m 08s) |
[production] |
10:27 |
<akosiaris> |
disabled puppet on uranium, investigating ganglia problems |
[production] |
10:05 |
<akosiaris> |
ganglia gmetad problems |
[production] |
05:25 |
<LocalisationUpdate> |
ResourceLoader cache refresh completed at Mon Jun 8 05:24:08 UTC 2015 (duration 24m 7s) |
[production] |
02:26 |
<LocalisationUpdate> |
completed (1.26wmf8) at 2015-06-08 02:25:12+00:00 |
[production] |
02:21 |
<l10nupdate> |
Synchronized php-1.26wmf8/cache/l10n: (no message) (duration: 07m 07s) |
[production] |
2015-06-05
§
|
22:42 |
<godog> |
powercycle graphite2001, no console no ssh |
[production] |
22:06 |
<andrewbogott> |
restarted apache on virt1000 |
[production] |
20:49 |
<ori> |
Upgrading hhvm-fss on application servers to 1.1.7; expect brief 5xx spike. |
[production] |
20:14 |
<demon> |
Synchronized php-1.26wmf8: live hack (duration: 02m 32s) |
[production] |
20:10 |
<mutante> |
apt-get upgrade on terbium |
[production] |
19:52 |
<godog> |
bounce redis on rdb1001/rdb1003 to pick up new slave limits |
[production] |
19:51 |
<mutante> |
chown root:root / on terbium |
[production] |
19:50 |
<godog> |
bounce redis on rdb1002/rdb1004 to pick up new slave limits |
[production] |
19:29 |
<godog> |
bounce redis again on rdb1003 after increasing the slave limits more |
[production] |
19:17 |
<godog> |
bounce redis on rdb1003 after bumping slave limits |
[production] |
19:07 |
<godog> |
redis master logs shows periodic 'cmd=sync scheduled to be closed ASAP for overcoming of output buffer limits.' indicating the slave fails to sync |
[production] |
18:40 |
<godog> |
spike in redis network starting at ~15.00 UTC, correlates with ocg failures |
[production] |