2018-03-28
ยง
|
20:44 |
<bsitzmann@tin> |
Started deploy [mobileapps/deploy@6a0d877]: Update mobileapps to a5833a0 |
[production] |
20:12 |
<mlitn@tin> |
Finished deploy [3d2png/deploy@c447488]: Updating 3d2png (duration: 02m 26s) |
[production] |
20:09 |
<mlitn@tin> |
Started deploy [3d2png/deploy@c447488]: Updating 3d2png |
[production] |
20:02 |
<Hauskatze> |
Restarted SULWatcher after maintenance |
[tools.stewardbots] |
19:57 |
<Hauskatze> |
Ran DELETE FROM logging WHERE l_timestamp < 20180201000000; -- Query OK, 14678 rows affected (6.64 sec) |
[tools.stewardbots] |
19:54 |
<mutante> |
deploy1001 - schedule downtime for reinstall with jessie, reinstalling (T175288) |
[production] |
19:39 |
<ottomata> |
bouncing main -> jumbo mirrormaker to apply increase in consumer num.streams |
[analytics] |
19:24 |
<twentyafterfour@tin> |
Synchronized php: group1 wikis to 1.31.0-wmf.26 (duration: 01m 17s) |
[production] |
19:22 |
<twentyafterfour@tin> |
rebuilt and synchronized wikiversions files: group1 wikis to 1.31.0-wmf.26 |
[production] |
19:21 |
<milimetric> |
synced refinery to hdfs (only python changes but just so we have latest) |
[analytics] |
19:20 |
<joal> |
Start Geowiki jobs (monthly and druid) starting 2018-01 |
[analytics] |
19:20 |
<twentyafterfour> |
Rolling back to wmf.26 due to increase in fatals: "Replication wait failed: lost connection to MySQL server during query" |
[production] |
19:13 |
<legoktm> |
killed stuck docker container on 1003 to free up root partition, and then deleted old/all images to free up the rest of the space |
[releng] |
19:12 |
<milimetric@tin> |
Finished deploy [analytics/refinery@c22fd1e]: Fixing python import bug (duration: 02m 48s) |
[production] |
19:09 |
<milimetric@tin> |
Started deploy [analytics/refinery@c22fd1e]: Fixing python import bug |
[production] |
19:09 |
<milimetric@tin> |
Started deploy [analytics/refinery@c22fd1e]: (no justification provided) |
[production] |
19:07 |
<legoktm> |
legoktm@integration-slave-docker-1003:/srv/jenkins-workspace/workspace$ sudo rm -rf * # full disk |
[releng] |
19:06 |
<twentyafterfour@tin> |
Synchronized php: group1 wikis to 1.31.0-wmf.27 (duration: 01m 17s) |
[production] |
19:05 |
<twentyafterfour@tin> |
rebuilt and synchronized wikiversions files: group1 wikis to 1.31.0-wmf.27 |
[production] |
19:02 |
<ebernhardson> |
restore elasticsearch eqiad disk high/low watermarks to 75/80% with all large reindexes complete |
[production] |
18:52 |
<urandom> |
upgrading restbase-dev1005-{a,b} to cassandra 3.11.2 -- T178905 |
[production] |
18:36 |
<joal> |
Making hdfs://analytics-hadoop/wmf/data/wmf/mediawiki_private accessible only by analytics-privatedata-users group (and hdfs obviously) |
[analytics] |
18:35 |
<andrewbogott> |
rebooting ws-web to recover from a full / |
[wikistream] |
18:18 |
<zhuyifei1999_> |
depooling encoding0[1-3]. systemd will restart them |
[video] |
18:17 |
<urandom> |
upgrading restbase-dev1004-b to cassandra 3.11.2 (canary) -- T178905 |
[production] |
18:15 |
<zhuyifei1999_> |
upgrading youtube_dl from 2017.8.27 to 2018.3.26.1 on encoding0[1-3] |
[video] |
18:12 |
<twentyafterfour@tin> |
rebuilt and synchronized wikiversions files: group0 wikis to 1.31.0-wmf.27 |
[production] |
18:12 |
<urandom> |
upgrading restbase-dev1004-a to cassandra 3.11.2 (canary) -- T178905 |
[production] |
18:03 |
<twentyafterfour> |
deploying 1.31.0-wmf.27 to group0. group1 in an hour. See T183966 for blockers. |
[production] |
18:02 |
<joal> |
Kill-Restart mobile_apps-session_metrics (bundle killed, coord started) |
[analytics] |
18:00 |
<joal> |
Kill-Restart mediawiki-history-reduced-coord after deploy |
[analytics] |
17:44 |
<joal> |
Deploying refinery onto hadoop |
[analytics] |
17:38 |
<joal@tin> |
Finished deploy [analytics/refinery@7135d44]: Regular weekly analytics deploy - Scheduled hadoop jobs updates (duration: 05m 21s) |
[production] |
17:32 |
<joal@tin> |
Started deploy [analytics/refinery@7135d44]: Regular weekly analytics deploy - Scheduled hadoop jobs updates |
[production] |
17:29 |
<joal> |
Deploy refinery using scap |
[analytics] |
16:37 |
<akosiaris> |
T189075 upload lttoolbox_3.4.0~r84331-1+wmf1 to apt.wikimedia.org/jessie-wikimedia/main |
[production] |
16:32 |
<ottomata> |
bouncing main -> jumbo mirror makers to increase heap size to 2G |
[analytics] |
16:04 |
<hasharAway> |
nodepool: deleting 4 instances that are no more used but that Nodepool failed to detect as no omre used (due to some reboots in the openstack infra) |
[releng] |
15:37 |
<catrope@tin> |
Synchronized wmf-config/InitialiseSettings.php: Enable oversampling for IN, GU, MP in preparation for eqsin (T189252) (duration: 01m 18s) |
[production] |
15:13 |
<andrewbogott> |
restarting nodepool on labnodepool1001 (cleanup from T189115) |
[production] |
15:08 |
<andrewbogott> |
restarting nova-fullstack on labnet1001 |
[production] |
15:07 |
<andrewbogott> |
restarting nova-network on labnet1001 in case it's upset by the rabbit outage |
[production] |
15:02 |
<andrewbogott> |
rebooting labservices1001 and labcontrol1001 for T189115 |
[production] |
15:00 |
<andrewbogott> |
stopping nova-fullstack on labnet1001 for T189115 |
[production] |
15:00 |
<andrewbogott> |
stopping nodepool on labnodepool1001 |
[production] |
14:58 |
<mobrovac@tin> |
Synchronized wmf-config/jobqueue.php: Disable redis queue for cirrusSearch jobs for test wikis, file 2/2 - T189137 (duration: 01m 17s) |
[production] |
14:56 |
<mobrovac@tin> |
Synchronized wmf-config/InitialiseSettings.php: Disable redis queue for cirrusSearch jobs for test wikis, file 1/2 - T189137 (duration: 01m 17s) |
[production] |
14:54 |
<ppchelko@tin> |
Finished deploy [cpjobqueue/deploy@c84880a]: Switch CirrusSearch jobs to kafka for test wikis (duration: 00m 44s) |
[production] |
14:54 |
<ppchelko@tin> |
Started deploy [cpjobqueue/deploy@c84880a]: Switch CirrusSearch jobs to kafka for test wikis |
[production] |
14:16 |
<ottomata> |
re-enabling replication of mediawiki job topics from main -> jumbo |
[analytics] |