3351-3400 of 10000 results (20ms)
2010-08-15 §
19:36 <jeluf> synchronized php-1.5/wmf-config/InitialiseSettings.php '24790 - Localize Wikipedia sitename in devanagari' [production]
19:33 <JeLuF> fixed zero byte thumbnail of commons:Shkval_head.jpg [production]
19:19 <jeluf> synchronized php-1.5/wmf-config/InitialiseSettings.php '24789 - Enable AbuseFilter for ja.wikipedia' [production]
19:18 <jeluf> synchronized php-1.5/wmf-config/abusefilter.php '24789 - Enable AbuseFilter for ja.wikipedia' [production]
19:14 <midom> synchronized php-1.5/wmf-config/db.php [production]
18:44 <midom> synchronized php-1.5/wmf-config/db.php [production]
18:24 <midom> synchronized php-1.5/wmf-config/db.php [production]
17:44 <midom> synchronized php-1.5/wmf-config/db.php [production]
16:23 <midom> synchronized php-1.5/wmf-config/db.php [production]
2010-08-14 §
20:15 <mark> Decommissioning srv150 [production]
19:56 <jeluf> synchronized php-1.5/wmf-config/InitialiseSettings.php '24789 - Enable AbuseFilter for ja.wikipedia' [production]
19:52 <jeluf> synchronized php-1.5/wmf-config/InitialiseSettings.php '24626 - Add an "autopatrolled" status for frwiktionary' [production]
15:37 <mark> dobson has failed RAID1 array member /dev/sda. Running long SMART self test on /dev/sda [production]
14:18 <mark> synchronized php-1.5/wmf-config/db.php 'Add ms2 and ms1 to clusters rc1 an cluster22' [production]
14:06 <mark> FLUSH TABLES WITH READ LOCK on ms1 for testing [production]
13:59 <mark> Stopping mysql on ms1 as monitoring test [production]
13:59 <mark> Granted SELECT on mysql.* to nagios on ms3 [production]
10:57 <mark> Removed oldest LVM snapshot on ixia [production]
09:43 <mark> Fixed apparmor profile /etc/apparmor.d/usr.sbin.mysqld on ms1, restarted mysql under apparmor [production]
09:39 <mark> START SLAVE on ms1, catching up with ms3 [production]
09:38 <mark> RESET SLAVE on db5 [production]
09:37 <mark> STOP SLAVE on db5 [production]
09:35 <mark> Stopped apparmor on ms1 [production]
08:41 <Andrew> Leaving as-is for now, hoping somebody with appropriate permissions can fix it later. [production]
08:40 <Andrew> STOP SLAVE on db5 gives me ERROR 1045 (00000): Access denied for user: 'wikiadmin@208.80.152.%' (Using password: NO) [production]
08:34 <Andrew> Slave is supposedly still running on db5. Assuming Roan didn't stop it when he switched masters a few days ago. Going to text somebody to confirm that stopping is correct course of action. [production]
08:24 <Andrew> db5 can't be lagged, it's the master ;-). Obviously something wrong with wfWaitForSlaves. [production]
08:19 <Andrew> db5 lagged 217904 seconds [production]
05:09 <Andrew> Ran thread_pending_relationship and thread_reaction schema changes on all LiquidThreads wikis [production]
05:06 <andrew> synchronizing Wikimedia installation... Revision: 70933 [production]
05:04 <Andrew> About to update LiquidThreads production version to the alpha. [production]
2010-08-13 §
22:03 <mark> API logins on commons (only) are reported broken [production]
21:45 <mark> Set correct $cluster variable for reinstalled knsq* squids [production]
21:03 <mark> Increased cache_mem from 1000 to 2500 on sq33, like the other API backend squids [production]
20:58 <mark> Stopping backend squid on sq33 [production]
20:50 <jeluf> synchronized php-1.5/wmf-config/InitialiseSettings.php '24769 - Import source addition for tpi.wikipedia.org' [production]
17:46 <Fred> and srv100 [production]
17:45 <Fred> restarted apache on srv219 and srv222 [production]
15:57 <mark> synchronized php-1.5/wmf-config/mc.php 'Remove some to-be-decommissioned from the down list' [production]
15:56 <mark> synchronized php-1.5/wmf-config/mc.php 'Remove some to-be-decommissioned hosts from the down list' [production]
15:53 <RobH> srv146 removed from puppet and nodelists, slated for wipe, decommissioned. [production]
15:47 <mark> Sent srv146 to death using echo b > /proc/sysrq-trigger. It had a read-only filesystem and is therefore decommissioned. [production]
15:38 <mark> Restarted backend squid on sq33 [production]
15:36 <mark> synchronized php-1.5/wmf-config/mc.php 'Remove some to-be-decommissioned hosts from the down list' [production]
15:25 <mark> Reinstalled sq32 with Lucid [production]
15:01 <mark> Removed sq86 and sq87 from API LVS pool [production]
14:55 <mark> sq80 had been down for a long time. Brought it back up and synced it [production]
14:54 <rainman-sr> all of the search cluster restored to pre-relocation configuration [production]
14:34 <robh> synchronized php-1.5/wmf-config/lucene.php 'reverting search13 to search11' [production]
13:55 <mark> /dev/sda on sq57 is busted [production]