551-600 of 2789 results (21ms)
2019-10-23 §
18:21 <mforns> deploying refinery with scap up to 1110d59c3983bcff4986bce1baf885f05ee06ba5 [analytics]
2019-10-22 §
15:47 <fdans> start backfilling of mediarequests per file from 2015-01-02 to 2019-05-17 after ok vetting of 2015-01-01 [analytics]
2019-10-18 §
14:45 <fdans> backfilling 2015-1-1 for mediarequests per file, proceeding with all days until 2019-05-17 if successful [analytics]
2019-10-17 §
18:01 <elukey> update librdkafka on eventlog1002 and restart eventlogging [analytics]
10:26 <elukey> rollback eventlogging back to Python 2, some errors (unseen in tests) logged by the processors [analytics]
10:18 <elukey> move eventlogging to python 3 [analytics]
2019-10-16 §
20:27 <ottomata> upgrading to spark 2.4.4 in analytics test cluster [analytics]
20:20 <joal> Kill-restart mediawiki-history-dumps-coord to pick up changes [analytics]
20:16 <joal> Deployed refinery onto HDFS [analytics]
20:08 <joal> Deployed refinery using scap [analytics]
19:45 <joal> Refinery-source v0.0.103 released to refinery [analytics]
19:29 <joal> Ask jenkins to release refinery-source v0.0.103 to archiva [analytics]
19:19 <joal> AQS deployed with mediarequest-top endpoint [analytics]
18:45 <joal> Manually create mediarequest-top cassandra keyspace and tables, and add fake test data into it [analytics]
2019-10-15 §
13:15 <elukey> re-enable timers on an-coord1001 [analytics]
12:57 <fdans> resumed backfilling of mediarequests per referer daily [analytics]
12:46 <elukey> moved hadoop cluster to new zookeeper cluster [analytics]
11:25 <elukey> stop all systemd timers on an-coord1001 as prep step for hadoop maintenance [analytics]
10:42 <fdans> backfilling January 1st 2015 for mediarequests per referer daily, proceeding with all days until May 2019 if successful [analytics]
2019-10-14 §
18:13 <joal> Manually add ban.wikipedia.org to pageview whitelist (T234768) [analytics]
14:28 <elukey> matomo upgraded to 3.11 on matomo1001 [analytics]
2019-10-11 §
12:51 <elukey> deployed eventlogging python3 version in deployment-prep [analytics]
07:09 <elukey> drop test_wmf_netflow fro druid analytics and restart turnilo [analytics]
06:24 <elukey> remove /tmp/hive-staging_hive_(2017|2018)* data from HDFS instead of /tmp/* to avoid causing hive failures (it needs to write temporary data for the current running jobs) [analytics]
06:04 <elukey> delete content of /tmp/* on HDFS [analytics]
2019-10-10 §
09:13 <joal> rerun failed pageview hour after manual job killing (pageview-hourly-wf-2019-10-9-19) [analytics]
09:13 <joal> Kill stuck oozie launcher in yarn (application_1569878150519_43184) [analytics]
2019-10-09 §
20:52 <milimetric> deploy of refinery and refinery-source 0.0.102 finally seems to have finished [analytics]
19:55 <milimetric> refinery ... probably? deployed with errors like "No such file or directory (2)\nrsync error" [analytics]
17:11 <elukey> restart druid-broker on druid100[5-6] - not serving data correctly [analytics]
2019-10-08 §
09:22 <elukey> delete druid old test datasource from the analytics cluster - test_kafka_event_centralnoticeimpression [analytics]
2019-10-07 §
17:46 <ottomata> powercycling stat1007 [analytics]
06:08 <elukey> upgrade python-kafka on eventlog1002 to 1.4.7-1 (manually via dpkg -i) [analytics]
2019-10-05 §
18:18 <elukey> kill/restart mediawiki-history-reduced oozie coord to pick up the new druid_loader.py version on HDFS [analytics]
06:49 <elukey> force umount/remount of /mnt/hdfs on an-coord1001 - processes stuck in D state, fuser proc consuming a ton of memory [analytics]
2019-10-04 §
16:27 <ottomata> manually rsyncing mediawiki_history 2019-08 snapshot to labstore1006 [analytics]
2019-10-03 §
14:17 <elukey> stop the Hadoop test cluster to migrate it to the new kerberos cluster [analytics]
13:26 <elukey> re-run refinery-download-project-namespace-map (modified with recent fixes for encoding and python3) [analytics]
09:48 <elukey> ran apt-get autoremove -y on all Hadoop workers to remove old Python 2 deps [analytics]
08:43 <elukey> apply 5% threshold to the HDFS balancer - T231828 [analytics]
07:48 <elukey> restart druid-broker on druid1003 (used by superset) [analytics]
07:47 <elukey> restart superset to test if a stale status might cause data not to be shown [analytics]
2019-10-02 §
21:21 <nuria> restarting superset [analytics]
16:18 <elukey> kill duplicate of oozie pageview-druid-hourly coord and start the wrongly killed oozie pageview-hourly-coord (causing jobs to wait for data) [analytics]
13:12 <elukey> remove python-request from all the hadoop workers (shouldn't be needed anymore) [analytics]
13:08 <elukey> kill/start oozie webrequest druid daily/hourly coords to pick up new druid_loader.py version [analytics]
13:04 <elukey> kill/start oozie virtualpageview druid daily/monthly coords to pick up new druid_loader.py version [analytics]
12:54 <elukey> kill/start oozie unique devices per family druid daily/daily_agg_mon/monthly coords to pick up new druid_loader.py version [analytics]
10:24 <elukey> restart unique dev per domain druid daily_agg_monthly/daily/montly coords to pick up new hdfs version of druid_loader.py [analytics]
10:15 <elukey> re-run unique devices druid daily 28/09/2019 - failed but possibly no alert was fired to analytics-alerts@ [analytics]