2009-10-27
§
|
21:08 |
<RoanKattouw> |
Last scap deployed r58216 |
[production] |
21:06 |
<tstarling> |
ran sync-common-all |
[production] |
20:59 |
<RoanKattouw> |
Running LocalisationUpdate/update.php on test |
[production] |
20:58 |
<RoanKattouw> |
running svn up on test |
[production] |
20:54 |
<Tim> |
deployed r58215 |
[production] |
20:46 |
<catrope> |
synchronized php-1.5/wmf-config/CommonSettings.php 'Fix config for strategywiki and usabilitywiki so the enhanced toolbar is enabled by default again' |
[production] |
20:18 |
<midom> |
synchronized php-1.5/includes/memcached-client.php 'bumping up timeout to 0.1s up from 0.05s - as we do have some megabyte sized objects....' |
[production] |
19:32 |
<mark> |
Ran "deluser catrope" across the cluster to prompt puppet to recreate |
[production] |
19:30 |
<mark> |
Fixed admins.pp in puppet, "managehome" attribute had disappeared |
[production] |
17:16 |
<midom> |
synchronized php-1.5/languages/LanguageConverter.php |
[production] |
10:14 |
<midom> |
synchronized php-1.5/StartProfiler.php |
[production] |
10:08 |
<midom> |
synchronized php-1.5/languages/LanguageConverter.php 'oops, this is not entirely right, livehacking for now' |
[production] |
09:58 |
<midom> |
synchronized php-1.5/languages/LanguageConverter.php 'push locking change live' |
[production] |
07:45 |
<domas> |
rolled live memcached changes, read/write timeouts down from 1s to 50ms, connect timeouts from 3x10ms with backoff to 2x10ms with no backoff, and fixed some host blacklist bug. |
[production] |
07:43 |
<midom> |
synchronized php-1.5/includes/memcached-client.php 'HERE WE GO MEMCACHED FIXES' |
[production] |
06:05 |
<domas> |
fixed perms in survey.wikimedia.org's /srv/org/wikimedia/survey/tmp/ , as well as set display_errors to off, in case there's more incompetence around ;-) |
[production] |
01:39 |
<rainman-sr> |
turned back on highlighting on en/de/fr, turned off interwiki search on smaller wikis ... we need more servers to cope with increase in traffic on large wikis |
[production] |
01:11 |
<atglenn> |
disabled search2 from lvs3 pybal config at rainman's request (it had load 21) |
[production] |
01:01 |
<rainman-sr> |
could someone please remove search2 from lsv3 search group ASAP |
[production] |
00:15 |
<andrew> |
synchronized php-1.5/extensions/LiquidThreads/pages/SpecialNewMessages.php 'Deploy r58176' |
[production] |
2009-10-26
§
|
20:45 |
<Andrew> |
scapping to update LiquidThreads |
[production] |
20:16 |
<Andrew> |
Going to update LiquidThreads to trunk state in a few minutes |
[production] |
16:08 |
<rainman-sr> |
overloads all around, turned off en/de/fr wiki highlighting so that searchs don't time out |
[production] |
11:10 |
<hcatlin> |
reworked mobile1's config so that its more standardized and more of the config is in the repo |
[production] |
08:53 |
<domas> |
updated nagios to reflect changed server roles |
[production] |
08:43 |
<domas> |
dewiki is now separate cluster, s5, replication switch over done at http://p.defau.lt/?kfvvlNOc4TkJ_6SCAVe6mg |
[production] |
08:42 |
<midom> |
synchronized php-1.5/wmf-config/CommonSettings.php 'dewiki readwrite' |
[production] |
08:40 |
<midom> |
synchronized php-1.5/wmf-config/db.php 'restructuring s2dewiki into s5' |
[production] |
08:38 |
<midom> |
synchronized php-1.5/wmf-config/CommonSettings.php 'dewiki read-only' |
[production] |
07:57 |
<midom> |
synchronized php-1.5/wmf-config/db.php 'entirely separating dewiki slaves' |
[production] |
06:54 |
<midom> |
synchronized php-1.5/wmf-config/db.php 'taking out db4 for copy to db23' |
[production] |
05:45 |
<midom> |
synchronized php-1.5/wmf-config/db.php |
[production] |
2009-10-25
§
|
15:23 |
<domas> |
converting usability initiative tables to InnoDB... |
[production] |
13:23 |
<domas> |
set up snapshot rotation on db10 |
[production] |
12:36 |
<hcatlin> |
mobile1: created init.d/cluster to correct USR1 sig problem, fully updated sys ops on wikitech |
[production] |
12:03 |
<domas> |
Mark, I'm sure you'll like that! ;-p~ |
[production] |
12:02 |
<domas> |
started sq43 without /dev/sdd COSS store (manual conf hack) |
[production] |
11:54 |
<domas> |
removed ns3 from nagios, added ns1 |
[production] |
11:45 |
<domas> |
bounced ns1 too, was affected by selective-answer leak (same number as ns0, btw, 507!) ages ago, just not noticed by nagios. this seem to resolve some slowness I noticed few times. |
[production] |
11:41 |
<domas> |
bounced pdns on ns0, was affected by selective-answer leak |
[production] |