2009-07-14
§
|
22:24 |
<brion> |
synchronized php-1.5/InitialiseSettings.php 'fix to confirmed group for en' |
[production] |
20:44 |
<brion> |
synchronized wmf-deployment/cache/trusted-xff.cdb |
[production] |
20:41 |
<brion> |
synchronized wmf-deployment/cache/trusted-xff.cdb |
[production] |
20:40 |
<brion> |
synchronized wmf-deployment/AdminSettings.php |
[production] |
20:22 |
<Fred> |
restarted a bunch of dead apaches |
[production] |
20:10 |
<brion> |
doing a sync-common-all w/ attempt to put test.wikipedia on wmf-deployment branch |
[production] |
19:50 |
<robh> |
synchronized php-1.5/InitialiseSettings.php |
[production] |
19:11 |
<robh> |
synchronized php-1.5/InitialiseSettings.php 'bug 19611 forgot one thing' |
[production] |
19:09 |
<robh> |
synchronized php-1.5/InitialiseSettings.php 'bug 19611' |
[production] |
19:08 |
<robh> |
synchronized php-1.5/InitialiseSettings.php 'bug 19611' |
[production] |
14:16 |
<domas> |
dropped all june snapshots on ms1, thus providing some relief |
[production] |
01:52 |
<river> |
patched ms4 in preperation for upload copy |
[production] |
2009-07-13
§
|
21:31 |
<Rob> |
pushing dns update to fix management ips for new apaches |
[production] |
19:05 |
<Fred> |
added storage3 to ganglia monitor. |
[production] |
18:50 |
<brion> |
synchronized php-1.5/abusefilter.php 'Disable dewiki missingsummary, mysteriously in abusefilter section. Per bug 19208' |
[production] |
16:30 |
<Fred> |
install wikimedia-nis-client to srv66 and mounted /home. |
[production] |
16:28 |
<brion> |
synchronized php-1.5/InitialiseSettings.php 'fixing wikispecies RC-IRC prefix to species.wikimedia' |
[production] |
16:27 |
<brion> |
test wiki was apparently moved from dead srv35 to srv66, which has new NFS-less config. thus fail since test runs from nfs |
[production] |
16:24 |
<brion> |
test wiki borked; reported down for several days now :) investigating |
[production] |
15:12 |
<midom> |
synchronized php-1.5/db.php 'db26 raid issues' |
[production] |
14:55 |
<midom> |
synchronized php-1.5/db.php 'db3 and db5 coming live as commons servers' |
[production] |
14:13 |
<domas> |
dropped few more snapshots, as %sys was increasing on ms1... |
[production] |
11:16 |
<domas> |
manually restarted plethora of failing apaches (direct segfaults and other possible APC corruptions, leading to php OOM errors) |
[production] |
09:50 |
<tstarling> |
synchronized php-1.5/includes/specials/SpecialBlockip.php |
[production] |
09:00 |
<Tim> |
restarted apache2 on image scalers |
[production] |
08:39 |
<tstarling> |
synchronized php-1.5/includes/Math.php 'statless render hack' |
[production] |
08:05 |
<Tim> |
killed all image scalers to see if that helps with ms1 load |
[production] |
08:00 |
<Tim> |
killed waiting apache processes |
[production] |
07:35 |
<midom> |
synchronized php-1.5/mc-pmtpa.php |
[production] |
07:24 |
<midom> |
synchronized php-1.5/mc-pmtpa.php 'swapping out srv81' |
[production] |
04:11 |
<Tim> |
fixed /opt/local/bin/zfs-replicate on ms1 to write the snapshot number before starting replication, to avoid permanent error "dataset already exists" after failure |
[production] |
02:16 |
<brion> |
-> https://bugzilla.wikimedia.org/show_bug.cgi?id=19683 |
[production] |
02:12 |
<brion> |
sync-common script doesn't work on nfs-free apaches; language lists etc not being updated. Deployment scripts need to be fixed? |
[production] |
02:03 |
<brion> |
srv159 is absurdly loaded/lagged wtf? |
[production] |
01:58 |
<brion> |
reports of servers with old config, seeing "doesn't exist" for new mhr.wikipedia. checking... |
[production] |
01:16 |
<brion> |
so far so good; CPU graphs on image scalers and ms1 look clean, and I can purge thumbs on commons ok |
[production] |
01:10 |
<brion> |
trying switching image scalers back in for a few, see if they go right back to old pattern or not |
[production] |
01:03 |
<brion> |
load on ms1 has fallen hugely; outgoing network is way up. looks like we're serving out http images fine... of course scaling's dead :P |
[production] |
00:59 |
<brion> |
stopping apache on image scaler boxes, see what that does |
[production] |
00:49 |
<brion> |
attempting to replicate domas's earlier temp success dropping oldest snapshot (last was 4/13): zfs destroy export/upload@weekly-2009-04-20_03:30:00 |
[production] |
00:45 |
<brion> |
restarting nfs server |
[production] |
00:44 |
<brion> |
stopping nfs server, restarting web server |
[production] |
00:40 |
<brion> |
restarting nfs server on ms1 |
[production] |
00:36 |
<brion> |
doesn't seem so far to have changed the NFS access delays on image scalers. |
[production] |
00:31 |
<brion> |
shutting down webserver7 on ms1 |
[production] |
00:23 |
<brion> |
investigating site problem reports. image server stack seems overloaded, so intermittent timeouts on nfs to apaches or http/squid to outside |
[production] |