2023-11-15
§
|
20:44 |
<xcollazo> |
Ran 'sudo -u analytics hdfs dfs -rm -r -skipTrash /user/hive/warehouse/wmf_dumps.db/wikitext_raw_rc1' to delete HDFS data of old release candidate table |
[analytics] |
20:43 |
<xcollazo> |
Ran 'sudo -u analytics hdfs dfs -rm -r -skipTrash /wmf/data/wmf_dumps/wikitext_raw_rc0' to delete HDFS data of old release candidate table |
[analytics] |
20:42 |
<xcollazo> |
Ran 'DROP TABLE wmf_dumps.wikitext_raw_rc0' and 'DROP TABLE wmf_dumps.wikitext_raw_rc1' to delete older release candidate tables. |
[analytics] |
14:51 |
<ottomata> |
deployed refine using refinery-job 0.2.26 JsonSchemaConverter from wikimedia-event-utilities - https://phabricator.wikimedia.org/T321854 |
[analytics] |
14:33 |
<joal> |
Deploy refinery onto HDFS (unique-devices hotfix) |
[analytics] |
13:44 |
<joal> |
Deploying refinery for unique-devices hotfix |
[analytics] |
11:22 |
<btullis> |
exiting safe mode |
[analytics] |
11:06 |
<btullis> |
merged all config files changes replacing an-coord1001 with an-mariadb1001 |
[analytics] |
11:04 |
<btullis> |
position confirmed, resetting all slaves on an-mariadb1001 for T284150 |
[analytics] |
11:02 |
<btullis> |
set an-coord1001 mysql to read_only |
[analytics] |
11:01 |
<btullis> |
entering HDFS safe mode |
[analytics] |
11:01 |
<btullis> |
proceeding with the implementation plan here: https://phabricator.wikimedia.org/T284150#9330525 |
[analytics] |
10:43 |
<btullis> |
temporarily disabled production jobs that write to HDFS |
[analytics] |
2023-11-07
§
|
20:48 |
<xcollazo> |
Ran 'kerberos-run-command hdfs hdfs dfs -chmod -R g+w /wmf/data/wmf_dumps/wikitext_raw_rc2' to ease experimentation on this release candidate table. |
[analytics] |
15:52 |
<btullis> |
restart airflow-sheduler and airflow-webserver services on an-test-client1002 |
[analytics] |
15:50 |
<btullis> |
restart mariadb service on an-test-coord1001 |
[analytics] |
15:50 |
<btullis> |
restart mariadb service on an-test-coord100 |
[analytics] |
15:49 |
<btullis> |
restart presto-server service on an-test-coord1001 and an-test-presto1001 to pick up new puppet 7 CA settings |
[analytics] |
15:48 |
<btullis> |
restart hive-server2 and hive-metastore services on an-test-coord1001 to pick up new puppet 7 CA settings. |
[analytics] |
15:35 |
<btullis> |
roll-restarting hadoop workers in test, to test new puppet 7 CA settings. |
[analytics] |
14:52 |
<btullis> |
roll-restarting hadoop masters on the test cluster, after upgrading to puppet 7 |
[analytics] |
12:05 |
<btullis> |
deploying datahub to prod for the pki certificates. |
[analytics] |
11:36 |
<btullis> |
deploying datahub to staging to start using pki certificates - https://gerrit.wikimedia.org/r/c/operations/deployment-charts/+/969345/ |
[analytics] |
10:40 |
<btullis> |
re-running the kafka_jumbo_ingestion in analytics airflow |
[analytics] |