101-150 of 2773 results (20ms)
2020-05-13 §
13:46 <elukey> upgrade spark2 on all stat100x hosts - T250161 [analytics]
06:47 <elukey> upgrade spark2 on stat1004 - canary host - T250161 [analytics]
2020-05-11 §
10:17 <elukey> re-run webrequest-load-wf-text-2020-5-11-9 [analytics]
06:06 <elukey> restart wikimedia-discovery-golden on stat1007 - apparenlty killed by no memory left to allocate on the system [analytics]
05:14 <elukey> force re-run of monitor_refine_event_failure_flags after fixing a refine failed hour [analytics]
2020-05-10 §
07:44 <joal> Rerun webrequest-load-wf-upload-2020-5-10-1 [analytics]
2020-05-08 §
21:06 <ottomata> running prefered replica election for kafka-jumbo to get preferred leaders back after reboot of broker earlier today - T252203 [analytics]
15:36 <ottomata> starting kafka broker on kafka-jumbo1006, same issue on other brokers when they are leaders of offending partitions - T252203 [analytics]
15:27 <ottomata> stopping kafka broker on kafka-jumbo1006 to investigate camus import failures - T252203 [analytics]
15:16 <ottomata> restarted turnilo after applying nuria and mforns changes [analytics]
2020-05-07 §
17:39 <ottomata> deploying fix to refinery bin/camus CamusPartitionChecker when using dynamic stream configs [analytics]
16:49 <joal> Restart and babysit mediawiki-history-denormalize-wf-2020-04 [analytics]
16:37 <elukey> roll restart of all the nodemanagers on the hadoop cluster to pick up new jvm settings [analytics]
13:53 <elukey> move stat1007 to role::statistics::explorer (adding jupyterhub) [analytics]
11:00 <joal> Moving application_1583418280867_334532 to the nice queue [analytics]
10:58 <joal> Rerun wikidata-articleplaceholder_metrics-wf-2020-5-6 [analytics]
07:45 <elukey> re-run mediawiki-history-denormalize [analytics]
07:43 <elukey> kill application_1583418280867_333560 after a chat with David, the job is consuming ~2TB of RAM [analytics]
07:32 <elukey> re-run mediawiki history load [analytics]
07:18 <elukey> execute yarn application -movetoqueue application_1583418280867_332862 -queue root.nice [analytics]
07:06 <elukey> restart mediawiki-history-load via hue [analytics]
06:41 <elukey> restart oozie on an-coord1001 [analytics]
05:46 <elukey> re-run mediarequest-hourly-wf-2020-5-6-19 [analytics]
05:35 <elukey> re-run two failed hours for webrequest load text (07/05T05) and upload (06/05T23) [analytics]
05:33 <elukey> restart hadoop yarn nodemanager on analytics1071 [analytics]
2020-05-06 §
12:49 <elukey> restart oozie on an-coord1001 to pick up the new shlib retention changes [analytics]
12:28 <mforns> re-run pageview-druid-hourly-coord for 2020-05-06T06:00:00 after oozie shared lib update [analytics]
11:30 <elukey> use /run/user as kerberos credential cache for stat1005 [analytics]
09:25 <elukey> re-run projectview coordinator for 2020-5-6-5 after oozie shared lib update [analytics]
09:24 <elukey> re-run virtualpageview coordinator for 2020-5-6-5 after oozie shared lib update [analytics]
09:13 <elukey> re-run apis coordinator for 2020-5-6-7 after oozie shared lib update [analytics]
09:11 <elukey> re-run learning features actor coordinator for 2020-5-6-7 after oozie shared lib update [analytics]
09:10 <elukey> re-run aqs-hourly coordinator for 2020-5-6-7 after oozie shared lib update [analytics]
09:09 <elukey> re-run mediacounts coordinator for 2020-5-6-7 after oozie shared lib update [analytics]
09:08 <elukey> re-run mediarequest coordinator for 2020-5-6-7 after oozie shared lib update [analytics]
09:07 <elukey> re-run data quality coordinators for 2020-5-6-5/6 after oozie shared lib update [analytics]
09:05 <elukey> re-run pageview-hourly coordinator 2020-5-6-6 after oozie shared lib update [analytics]
09:04 <elukey> execute oozie admin -sharelibupdate on an-coord1001 [analytics]
06:05 <elukey> execute hdfs dfs -chown -R analytics-search:analytics-search-users /wmf/data/discovery/search_satisfaction/daily/year=2019 [analytics]
2020-05-05 §
19:49 <mforns> Finished re-deploying refinery using scap, then re-deploying onto hdfs [analytics]
18:47 <mforns> Finished deploying refinery using scap, then deploying onto hdfs [analytics]
18:13 <mforns> Deploying refinery using scap, then deploying onto hdfs [analytics]
18:02 <mforns> Deployed refinery-source using the awesome new jenkins jobs :] [analytics]
13:15 <joal> Dropping unavailable mediawiki-history-reduced datasources from superset [analytics]
2020-05-04 §
17:08 <joal> Restart refinery-sqoop-mediawiki-private.service on an-launcher1001 [analytics]
17:03 <elukey> restart refinery-drop-webrequest-refined-partitions after manual chown [analytics]
17:03 <joal> Restart refinery-sqoop-whole-mediawiki.service on an-launcher1001 [analytics]
17:02 <elukey> chown analytics (was: hdfs) /wmf/data/wmf/webrequest/webrequest_source=text/year=2019/month=12/day=14/hour={13,18} [analytics]
16:44 <joal> Deploy refinery again using scap (trying to fox sqoop) [analytics]
15:39 <joal> restart refinery-sqoop-whole-mediawiki.service [analytics]