|
2020-10-01
§
|
| 19:07 |
<fdans> |
deploying wikistats |
[analytics] |
| 19:06 |
<fdans> |
restarted banner_activity-druid-daily-coord from Sep 26 |
[analytics] |
| 18:59 |
<fdans> |
restarting mediawiki-history-load-coord |
[analytics] |
| 18:57 |
<fdans> |
creating hive table wmf_raw.mediawiki_page_props |
[analytics] |
| 18:56 |
<fdans> |
creating hive table wmf_raw.mediawiki_user_properties |
[analytics] |
| 17:40 |
<elukey> |
remove + re-create /srv/deployment/analytics/refinery* on stat100[46] (perm issues after reimage) |
[analytics] |
| 17:32 |
<elukey> |
remove + re-create /srv/deployment/analytics/refinery on stat1007 (perm issues after reimage) |
[analytics] |
| 17:18 |
<fdans> |
deploying refinery |
[analytics] |
| 14:51 |
<elukey> |
bootstrap an-worker109[8-9] as hadoop workers (with GPU) |
[analytics] |
| 13:35 |
<elukey> |
bootstrap an-worker1097 (GPU node) as hadoop worker |
[analytics] |
| 13:15 |
<elukey> |
restart performance-asoranking on stat1007 |
[analytics] |
| 13:15 |
<elukey> |
execute "sudo chown analytics-privatedata:analytics-privatedata-users /srv/published-datasets/performance/autonomoussystems/*" on stat1007 to fix a perm issue after reimage |
[analytics] |
| 10:30 |
<elukey> |
add an-worker1103 to the hadoop cluster |
[analytics] |
| 07:15 |
<elukey> |
restart hdfs namenodes on an-master100[1,2] to pick up new hadoop workers settings |
[analytics] |
| 06:04 |
<elukey> |
execyte "sudo chown -R analytics-privatedata:analytics-privatedata-users /srv/geoip/archive" on stat1007 - T264152 |
[analytics] |
| 05:58 |
<elukey> |
execute "sudo -u hdfs kerberos-run-command hdfs hdfs dfs -chown -R analytics-privatedata /wmf/data/archive/geoip" - T264152 |
[analytics] |
|
2020-09-24
§
|
| 13:24 |
<elukey> |
moved the hadoop cluster to puppet TLS certificates |
[analytics] |
| 13:20 |
<elukey> |
re-enable timers on an-launcher1002 after maintenance |
[analytics] |
| 09:51 |
<elukey> |
stop all timers on an-launcher1002 to ease maintenance |
[analytics] |
| 09:41 |
<elukey> |
force re-creation of jupyterhub's default venv on stat1006 after reimage |
[analytics] |
| 07:29 |
<klausman> |
Starting reimaging of stat1006 |
[analytics] |
| 06:48 |
<elukey> |
on an-launcher1002: sudo -u hdfs kerberos-run-command hdfs hdfs dfs -rm -r -skipTrash /var/log/hadoop-yarn/apps/mirrys/logs/* |
[analytics] |
| 06:45 |
<elukey> |
on an-launcher1002: sudo -u hdfs kerberos-run-command hdfs hdfs dfs -rm -r -skipTrash /var/log/hadoop-yarn/apps/analytics-privatedata/logs/* |
[analytics] |
| 06:39 |
<elukey> |
manually ran "/usr/bin/find /srv/backup/hadoop/namenode -mtime +15 -delete" on an-master1002 to free some space in the backup partition |
[analytics] |