651-700 of 5815 results (29ms)
2023-05-16 §
11:04 <btullis> depooled schema2004 for T335042 [analytics]
2023-05-15 §
07:16 <joal> Rerun failed refine_eventlogging_legacy job for universallanguageselector [analytics]
07:02 <joal> Rerun failed refine_event job for content_translation_event [analytics]
2023-05-12 §
16:05 <mforns> dropped mobile_apps_* hive tables because of https://phabricator.wikimedia.org/T329310 [analytics]
2023-05-11 §
14:55 <xcollazo> replaced /user/spark/share/lib/spark-3.1.2-assembly.jar in HDFS with new version that includes Iceberg. [analytics]
2023-05-10 §
20:37 <milimetric> deployed refinery (except to an-airflow1001) [analytics]
16:50 <stevemunene> deploy conda-analytics v0.0.13 T335721 [analytics]
16:36 <btullis> installing airflow 2.6.0 on an-test-client1001 for T336286 [analytics]
15:51 <mforns> stopped Airflow DAG mobile_app_session_metrics_weekly because of https://phabricator.wikimedia.org/T329310 [analytics]
15:50 <mforns> killed oozie job mobile_apps-uniques-monthly-coord because of https://phabricator.wikimedia.org/T329310 [analytics]
2023-05-09 §
13:02 <btullis> rebooting an-worker1088 after firmware upgrade for T336077 [analytics]
12:59 <btullis> upgrading SAS RAID controller firmware on an-worker1088 for T336077 [analytics]
12:27 <btullis> rebooting eventlog1003 for T325132 [analytics]
2023-05-08 §
21:22 <mforns> deployed airflow analytics for a quick fix [analytics]
2023-05-05 §
15:44 <mforns> re-ran projectview_hourly DAG for 2023-05-05T13 [analytics]
15:06 <mforns> deployed airflow analytics [analytics]
14:26 <btullis> roll-rebooting presto workers for T335835 [analytics]
2023-05-04 §
20:12 <btullis> executed `sudo apt clean` on stat1005 to free up some space. [analytics]
20:09 <btullis> restarting hive-server2 and hive-metastore on an-coord1002 [analytics]
14:07 <btullis> failing back hive service to an-coord1001 [analytics]
2023-05-03 §
21:43 <milimetric> deployed refinery-source and refinery to prepare for launching new airflow druid jobs [analytics]
2023-05-02 §
18:20 <milimetric> deployed refinery as part of weekly train [analytics]
16:24 <btullis> roll-restarting AQS [analytics]
13:49 <btullis> deploying updated mediawiki history snapshot to aqs [analytics]
09:33 <btullis> depooled schema2003 for T334049 [analytics]
2023-04-26 §
11:12 <btullis> restart refine_netflow service on an-launcher1002. [analytics]
10:55 <btullis> deploying refinery to hdfs [analytics]
09:12 <btullis> deploying refinery [analytics]
2023-04-25 §
13:47 <btullis> rebooting an-test-worker1002 T335358 [analytics]
13:07 <btullis> restarted the gobblin-eventlogging_legacy_test on an-test-coord1001 [analytics]
13:06 <btullis> killed the gobblin-eventlogging_legacy_test on an-test-coord1001 [analytics]
2023-04-24 §
09:40 <btullis> upgrading RAID controller firmware an an-worker1110 T334832 [analytics]
2023-04-20 §
16:25 <SandraEbele> Deployed refinery using scap, then deployed onto hdfs as part of weekly deployment train. [analytics]
15:44 <SandraEbele> deploying weekly deployment train for analytics refinery. [analytics]
2023-04-18 §
15:49 <btullis> restarting refinery-drop-raw-netflow-event.service refinery-drop-webrequest-raw-partitions.service refinery-drop-webrequest-refined-partitions.service on an-launchger1002 [analytics]
15:48 <btullis> restart refinery-drop-raw-event.service on an-launcher1002 [analytics]
15:45 <btullis> restart refinery-drop-pageview-actor-hourly-partitions.service on an-launcher1002 [analytics]
15:44 <btullis> restart refinery-drop-eventlogging-legacy-raw-partitions.service on an-launcher1002 [analytics]
15:42 <btullis> restart drop-webrequest-actor-metrics-rollup-hourly.service on an-launcher1002 [analytics]
15:40 <btullis> restart drop-webrequest-actor-metrics-hourly.service on an-launcher1002 [analytics]
14:51 <btullis> restart drop-webrequest-actor-label-hourly.service on an-launcher1002 [analytics]
13:56 <btullis> re-enabling gobblin timers [analytics]
13:52 <btullis> pooled schema1004 [analytics]
13:51 <btullis> pooled aqs10[14,15,19] [analytics]
13:49 <btullis> re-enabling YARN queues [analytics]
13:43 <btullis> leaving HDFS safe mode on an-master1001 [analytics]
11:55 <btullis> entering safe mode for prod hadoop HDFS [analytics]
11:48 <btullis> depooled aqs10[14,15,19] [analytics]
11:45 <btullis> depooled schema1004 T333377 [analytics]
11:41 <btullis> refreshed yarn queues with `sudo cumin '(A:hadoop-master or A:hadoop-standby)' 'kerberos-run-command yarn /usr/bin/yarn rmadmin -refreshQueues'` [analytics]