4101-4150 of 10000 results (27ms)
2015-07-13 §
06:25 <ori> Experimenting with altering the localisation cache implementation for testwiki, operations/mediawiki-config on tin will have a local hack for a little bit [production]
05:07 <LocalisationUpdate> ResourceLoader cache refresh completed at Mon Jul 13 05:07:32 UTC 2015 (duration 7m 31s) [production]
02:26 <LocalisationUpdate> ResourceLoader cache refresh completed at Mon Jul 13 02:25:58 UTC 2015 (duration 25m 57s) [production]
02:23 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-13 02:23:43+00:00 [production]
02:20 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 06m 16s) [production]
02:10 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-13 02:10:25+00:00 [production]
02:10 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 00m 34s) [production]
01:47 <springle> restarted labsdb1002 mysqld while troubleshooting replication [production]
2015-07-12 §
14:59 <bblack> upgraded most packages on sodium [production]
14:48 <bblack> upgraded apache2 to 2.2.22-1ubuntu1.9 on: antimony argon caesium fluorine helium iodine logstash1001 logstash1003 magnesium neon netmon1001 rhodium stat1001 ytterbium [production]
04:49 <LocalisationUpdate> ResourceLoader cache refresh completed at Sun Jul 12 04:49:08 UTC 2015 (duration 49m 7s) [production]
02:26 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-12 02:26:52+00:00 [production]
02:25 <LocalisationUpdate> ResourceLoader cache refresh completed at Sun Jul 12 02:25:33 UTC 2015 (duration 25m 32s) [production]
02:23 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 06m 12s) [production]
02:10 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-12 02:10:00+00:00 [production]
02:09 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 00m 34s) [production]
2015-07-11 §
19:48 <jynus> stopping labsdb1002 after table corruption has been detected [production]
19:37 <urandom> from restbase1002, starting revision culling process (node thin_out_key_rev_value_data.js `hostname -i` local_group_wikimedia_T_parsoid_html 2>&1 | tee >(gzip -c > local_group_wikimedia_T_parsoid_html.log.`date +%s`.gz)) [production]
19:33 <urandom> restbase: setting gc_grace_seconds to 604800 (1 week) on local_group_wikipedia_T_parsoid_html.data [production]
04:56 <LocalisationUpdate> ResourceLoader cache refresh completed at Sat Jul 11 04:55:56 UTC 2015 (duration 55m 55s) [production]
04:21 <bd808> Logstash cluster upgrade complete! Kibana working again [production]
04:21 <bd808> Upgraded Elasticsearch to 1.6.0 on logstash1006 [production]
04:12 <bd808> rebooting logstash1006 [production]
04:07 <bd808> logstash1005 fully recovered all shards [production]
03:21 <mattflaschen> Synchronized php-1.26wmf13/extensions/Flow/includes/Parsoid/Utils.php: Bump Flow to encode page name when sending to Parsoid (duration: 00m 13s) [production]
02:28 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-11 02:28:18+00:00 [production]
02:25 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 06m 07s) [production]
02:25 <LocalisationUpdate> ResourceLoader cache refresh completed at Sat Jul 11 02:25:19 UTC 2015 (duration 25m 18s) [production]
02:09 <LocalisationUpdate> completed (1.26wmf13) at 2015-07-11 02:09:45+00:00 [production]
02:09 <l10nupdate> Synchronized php-1.26wmf13/cache/l10n: (no message) (duration: 00m 35s) [production]
00:46 <bd808> Upgraded Elasticsearch to 1.6.0 on logstash1005; replicas recovering now [production]
00:35 <bd808> rebooting logstash1005 [production]
00:30 <bd808> logstash1004 fully recovered all shards [production]
2015-07-10 §
22:51 <mutante> tendril: very short maintenance downtime [production]
20:10 <bd808> `service elasticsearch start` not starting on logstash1004; investigating [production]
20:07 <bd808> ran apt-get upgrade on logstash1004 [production]
19:52 <mutante> adminbot - built and imported 1.7.10 into APT repo [production]
19:43 <bd808> rebooting logstash1004 [production]
19:40 <bd808> Kibana seems to be broken by mixed 1.6.0/1.3.9 cluster [production]
19:32 <bd808> kibana not seeing indices after upgrading elasticsearch to 1.6.0; investigating [production]
19:26 <bd808> Upgraded logstash1003 to elasticsearch 1.6.0 [production]
19:22 <bd808> Upgraded logstash1002 to elasticsearch 1.6.0 [production]
19:19 <bd808> Upgraded logstash1001 to elasticsearch 1.6.0 [production]
19:10 <krenair> Synchronized php-1.26wmf13/extensions/VisualEditor/lib/ve/src/ce/nodes/ve.ce.TableNode.js: https://gerrit.wikimedia.org/r/#/c/224122/ (duration: 00m 12s) [production]
18:11 <gwicke> ansible -i production restbase -a 'nodetool setcompactionthroughput 120' [production]
18:00 <gwicke> ansible -i production restbase -a 'nodetool setcompactionthroughput 90' [production]
17:49 <gwicke> rolling restart of the cassandra cluster to apply https://gerrit.wikimedia.org/r/#/c/224114/ [production]
17:32 <demon> Synchronized wmf-config/CommonSettings.php: prevent race condition on writing settings (duration: 00m 13s) [production]
17:26 <moritzm> installed python security updates on mc* [production]
17:25 <Coren> rebooting labstore2001 (experiments with the new raid setup caused the mapper table to fill) [production]