4751-4800 of 10000 results (42ms)
2018-03-28 ยง
20:12 <mlitn@tin> Finished deploy [3d2png/deploy@c447488]: Updating 3d2png (duration: 02m 26s) [production]
20:09 <mlitn@tin> Started deploy [3d2png/deploy@c447488]: Updating 3d2png [production]
20:02 <Hauskatze> Restarted SULWatcher after maintenance [tools.stewardbots]
19:57 <Hauskatze> Ran DELETE FROM logging WHERE l_timestamp < 20180201000000; -- Query OK, 14678 rows affected (6.64 sec) [tools.stewardbots]
19:54 <mutante> deploy1001 - schedule downtime for reinstall with jessie, reinstalling (T175288) [production]
19:39 <ottomata> bouncing main -> jumbo mirrormaker to apply increase in consumer num.streams [analytics]
19:24 <twentyafterfour@tin> Synchronized php: group1 wikis to 1.31.0-wmf.26 (duration: 01m 17s) [production]
19:22 <twentyafterfour@tin> rebuilt and synchronized wikiversions files: group1 wikis to 1.31.0-wmf.26 [production]
19:21 <milimetric> synced refinery to hdfs (only python changes but just so we have latest) [analytics]
19:20 <joal> Start Geowiki jobs (monthly and druid) starting 2018-01 [analytics]
19:20 <twentyafterfour> Rolling back to wmf.26 due to increase in fatals: "Replication wait failed: lost connection to MySQL server during query" [production]
19:13 <legoktm> killed stuck docker container on 1003 to free up root partition, and then deleted old/all images to free up the rest of the space [releng]
19:12 <milimetric@tin> Finished deploy [analytics/refinery@c22fd1e]: Fixing python import bug (duration: 02m 48s) [production]
19:09 <milimetric@tin> Started deploy [analytics/refinery@c22fd1e]: Fixing python import bug [production]
19:09 <milimetric@tin> Started deploy [analytics/refinery@c22fd1e]: (no justification provided) [production]
19:07 <legoktm> legoktm@integration-slave-docker-1003:/srv/jenkins-workspace/workspace$ sudo rm -rf * # full disk [releng]
19:06 <twentyafterfour@tin> Synchronized php: group1 wikis to 1.31.0-wmf.27 (duration: 01m 17s) [production]
19:05 <twentyafterfour@tin> rebuilt and synchronized wikiversions files: group1 wikis to 1.31.0-wmf.27 [production]
19:02 <ebernhardson> restore elasticsearch eqiad disk high/low watermarks to 75/80% with all large reindexes complete [production]
18:52 <urandom> upgrading restbase-dev1005-{a,b} to cassandra 3.11.2 -- T178905 [production]
18:36 <joal> Making hdfs://analytics-hadoop/wmf/data/wmf/mediawiki_private accessible only by analytics-privatedata-users group (and hdfs obviously) [analytics]
18:35 <andrewbogott> rebooting ws-web to recover from a full / [wikistream]
18:18 <zhuyifei1999_> depooling encoding0[1-3]. systemd will restart them [video]
18:17 <urandom> upgrading restbase-dev1004-b to cassandra 3.11.2 (canary) -- T178905 [production]
18:15 <zhuyifei1999_> upgrading youtube_dl from 2017.8.27 to 2018.3.26.1 on encoding0[1-3] [video]
18:12 <twentyafterfour@tin> rebuilt and synchronized wikiversions files: group0 wikis to 1.31.0-wmf.27 [production]
18:12 <urandom> upgrading restbase-dev1004-a to cassandra 3.11.2 (canary) -- T178905 [production]
18:03 <twentyafterfour> deploying 1.31.0-wmf.27 to group0. group1 in an hour. See T183966 for blockers. [production]
18:02 <joal> Kill-Restart mobile_apps-session_metrics (bundle killed, coord started) [analytics]
18:00 <joal> Kill-Restart mediawiki-history-reduced-coord after deploy [analytics]
17:44 <joal> Deploying refinery onto hadoop [analytics]
17:38 <joal@tin> Finished deploy [analytics/refinery@7135d44]: Regular weekly analytics deploy - Scheduled hadoop jobs updates (duration: 05m 21s) [production]
17:32 <joal@tin> Started deploy [analytics/refinery@7135d44]: Regular weekly analytics deploy - Scheduled hadoop jobs updates [production]
17:29 <joal> Deploy refinery using scap [analytics]
16:37 <akosiaris> T189075 upload lttoolbox_3.4.0~r84331-1+wmf1 to apt.wikimedia.org/jessie-wikimedia/main [production]
16:32 <ottomata> bouncing main -> jumbo mirror makers to increase heap size to 2G [analytics]
16:04 <hasharAway> nodepool: deleting 4 instances that are no more used but that Nodepool failed to detect as no omre used (due to some reboots in the openstack infra) [releng]
15:37 <catrope@tin> Synchronized wmf-config/InitialiseSettings.php: Enable oversampling for IN, GU, MP in preparation for eqsin (T189252) (duration: 01m 18s) [production]
15:13 <andrewbogott> restarting nodepool on labnodepool1001 (cleanup from T189115) [production]
15:08 <andrewbogott> restarting nova-fullstack on labnet1001 [production]
15:07 <andrewbogott> restarting nova-network on labnet1001 in case it's upset by the rabbit outage [production]
15:02 <andrewbogott> rebooting labservices1001 and labcontrol1001 for T189115 [production]
15:00 <andrewbogott> stopping nova-fullstack on labnet1001 for T189115 [production]
15:00 <andrewbogott> stopping nodepool on labnodepool1001 [production]
14:58 <mobrovac@tin> Synchronized wmf-config/jobqueue.php: Disable redis queue for cirrusSearch jobs for test wikis, file 2/2 - T189137 (duration: 01m 17s) [production]
14:56 <mobrovac@tin> Synchronized wmf-config/InitialiseSettings.php: Disable redis queue for cirrusSearch jobs for test wikis, file 1/2 - T189137 (duration: 01m 17s) [production]
14:54 <ppchelko@tin> Finished deploy [cpjobqueue/deploy@c84880a]: Switch CirrusSearch jobs to kafka for test wikis (duration: 00m 44s) [production]
14:54 <ppchelko@tin> Started deploy [cpjobqueue/deploy@c84880a]: Switch CirrusSearch jobs to kafka for test wikis [production]
14:16 <ottomata> re-enabling replication of mediawiki job topics from main -> jumbo [analytics]
13:51 <elukey> reduced number of jobrunner runners on the videoscalers after the last burst of jobs that maxed out the cluster [production]