2020-10-01
§
|
14:51 |
<elukey> |
bootstrap an-worker109[8-9] as hadoop workers (with GPU) |
[analytics] |
13:35 |
<elukey> |
bootstrap an-worker1097 (GPU node) as hadoop worker |
[analytics] |
13:15 |
<elukey> |
restart performance-asoranking on stat1007 |
[analytics] |
13:15 |
<elukey> |
execute "sudo chown analytics-privatedata:analytics-privatedata-users /srv/published-datasets/performance/autonomoussystems/*" on stat1007 to fix a perm issue after reimage |
[analytics] |
10:30 |
<elukey> |
add an-worker1103 to the hadoop cluster |
[analytics] |
07:15 |
<elukey> |
restart hdfs namenodes on an-master100[1,2] to pick up new hadoop workers settings |
[analytics] |
06:04 |
<elukey> |
execyte "sudo chown -R analytics-privatedata:analytics-privatedata-users /srv/geoip/archive" on stat1007 - T264152 |
[analytics] |
05:58 |
<elukey> |
execute "sudo -u hdfs kerberos-run-command hdfs hdfs dfs -chown -R analytics-privatedata /wmf/data/archive/geoip" - T264152 |
[analytics] |
2020-09-24
§
|
13:24 |
<elukey> |
moved the hadoop cluster to puppet TLS certificates |
[analytics] |
13:20 |
<elukey> |
re-enable timers on an-launcher1002 after maintenance |
[analytics] |
09:51 |
<elukey> |
stop all timers on an-launcher1002 to ease maintenance |
[analytics] |
09:41 |
<elukey> |
force re-creation of jupyterhub's default venv on stat1006 after reimage |
[analytics] |
07:29 |
<klausman> |
Starting reimaging of stat1006 |
[analytics] |
06:48 |
<elukey> |
on an-launcher1002: sudo -u hdfs kerberos-run-command hdfs hdfs dfs -rm -r -skipTrash /var/log/hadoop-yarn/apps/mirrys/logs/* |
[analytics] |
06:45 |
<elukey> |
on an-launcher1002: sudo -u hdfs kerberos-run-command hdfs hdfs dfs -rm -r -skipTrash /var/log/hadoop-yarn/apps/analytics-privatedata/logs/* |
[analytics] |
06:39 |
<elukey> |
manually ran "/usr/bin/find /srv/backup/hadoop/namenode -mtime +15 -delete" on an-master1002 to free some space in the backup partition |
[analytics] |