2025-06-05
ยง
|
12:56 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2042 (re)pooling @ 40%: Repooling', diff saved to https://phabricator.wikimedia.org/P77161 and previous config saved to /var/cache/conftool/dbconfig/20250605-125641-root.json |
[production] |
12:56 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2045 (re)pooling @ 40%: Repooling', diff saved to https://phabricator.wikimedia.org/P77160 and previous config saved to /var/cache/conftool/dbconfig/20250605-125637-root.json |
[production] |
12:55 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'db2151 (re)pooling @ 10%: Repooling', diff saved to https://phabricator.wikimedia.org/P77159 and previous config saved to /var/cache/conftool/dbconfig/20250605-125540-root.json |
[production] |
12:54 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.ganeti.drain-node (exit_code=0) for draining ganeti node ganeti2035.codfw.wmnet |
[production] |
12:54 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host ganeti2035.codfw.wmnet |
[production] |
12:51 |
<marostegui@cumin1002> |
DONE (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1:00:00 on db2151.codfw.wmnet with reason: Maintenance |
[production] |
12:50 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'Depool db2151 T395989', diff saved to https://phabricator.wikimedia.org/P77158 and previous config saved to /var/cache/conftool/dbconfig/20250605-125057-marostegui.json |
[production] |
12:49 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2210 (T395241)', diff saved to https://phabricator.wikimedia.org/P77157 and previous config saved to /var/cache/conftool/dbconfig/20250605-124912-fceratto.json |
[production] |
12:48 |
<jmm@cumin1003> |
START - Cookbook sre.hosts.reboot-single for host ganeti2035.codfw.wmnet |
[production] |
12:43 |
<jmm@cumin1003> |
START - Cookbook sre.ganeti.drain-node for draining ganeti node ganeti2035.codfw.wmnet |
[production] |
12:41 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2042 (re)pooling @ 20%: Repooling', diff saved to https://phabricator.wikimedia.org/P77156 and previous config saved to /var/cache/conftool/dbconfig/20250605-124136-root.json |
[production] |
12:41 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2045 (re)pooling @ 20%: Repooling', diff saved to https://phabricator.wikimedia.org/P77155 and previous config saved to /var/cache/conftool/dbconfig/20250605-124131-root.json |
[production] |
12:41 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Depooling db2210 (T395241)', diff saved to https://phabricator.wikimedia.org/P77154 and previous config saved to /var/cache/conftool/dbconfig/20250605-124110-fceratto.json |
[production] |
12:41 |
<fceratto@cumin1002> |
DONE (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1 day, 0:00:00 on db2210.codfw.wmnet with reason: Maintenance |
[production] |
12:40 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2206 (T395241)', diff saved to https://phabricator.wikimedia.org/P77153 and previous config saved to /var/cache/conftool/dbconfig/20250605-124043-fceratto.json |
[production] |
12:32 |
<jakob@deploy1003> |
helmfile [eqiad] DONE helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:32 |
<jakob@deploy1003> |
helmfile [eqiad] START helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:31 |
<jakob@deploy1003> |
helmfile [codfw] DONE helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:30 |
<jakob@deploy1003> |
helmfile [codfw] START helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:30 |
<jakob@deploy1003> |
helmfile [staging] DONE helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:30 |
<jakob@deploy1003> |
helmfile [staging] START helmfile.d/services/wikidata-query-gui: apply |
[production] |
12:26 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2042 (re)pooling @ 10%: Repooling', diff saved to https://phabricator.wikimedia.org/P77151 and previous config saved to /var/cache/conftool/dbconfig/20250605-122631-root.json |
[production] |
12:26 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2045 (re)pooling @ 10%: Repooling', diff saved to https://phabricator.wikimedia.org/P77150 and previous config saved to /var/cache/conftool/dbconfig/20250605-122625-root.json |
[production] |
12:25 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2206', diff saved to https://phabricator.wikimedia.org/P77149 and previous config saved to /var/cache/conftool/dbconfig/20250605-122537-fceratto.json |
[production] |
12:21 |
<marostegui@cumin1002> |
DONE (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1:00:00 on es[2042,2045].codfw.wmnet with reason: Maintenance |
[production] |
12:20 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'Depool es2042 es2045 T395241', diff saved to https://phabricator.wikimedia.org/P77147 and previous config saved to /var/cache/conftool/dbconfig/20250605-122035-marostegui.json |
[production] |
12:10 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2206', diff saved to https://phabricator.wikimedia.org/P77146 and previous config saved to /var/cache/conftool/dbconfig/20250605-121029-fceratto.json |
[production] |
11:55 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2206 (T395241)', diff saved to https://phabricator.wikimedia.org/P77145 and previous config saved to /var/cache/conftool/dbconfig/20250605-115522-fceratto.json |
[production] |
11:49 |
<fceratto@deploy1003> |
helmfile [aux-k8s-eqiad] 'sync' command on namespace 'zarcillo' for release 'main' . |
[production] |
11:48 |
<fceratto@deploy1003> |
helmfile [aux-k8s-eqiad] 'sync' command on namespace 'zarcillo' for release 'main' . |
[production] |
11:47 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Depooling db2206 (T395241)', diff saved to https://phabricator.wikimedia.org/P77144 and previous config saved to /var/cache/conftool/dbconfig/20250605-114711-fceratto.json |
[production] |
11:47 |
<fceratto@cumin1002> |
DONE (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1 day, 0:00:00 on db2206.codfw.wmnet with reason: Maintenance |
[production] |
11:42 |
<fceratto@cumin1002> |
DONE (PASS) - Cookbook sre.hosts.downtime (exit_code=0) for 1 day, 0:00:00 on db2199.codfw.wmnet with reason: Maintenance |
[production] |
11:42 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179 (T395241)', diff saved to https://phabricator.wikimedia.org/P77143 and previous config saved to /var/cache/conftool/dbconfig/20250605-114213-fceratto.json |
[production] |
11:35 |
<moritzm> |
installing Linux 5.10.237 on Bullseye hosts |
[production] |
11:27 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179', diff saved to https://phabricator.wikimedia.org/P77142 and previous config saved to /var/cache/conftool/dbconfig/20250605-112706-fceratto.json |
[production] |
11:26 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.ganeti.drain-node (exit_code=0) for draining ganeti node ganeti2032.codfw.wmnet |
[production] |
11:25 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host ganeti2032.codfw.wmnet |
[production] |
11:25 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2046 (re)pooling @ 100%: Repooling', diff saved to https://phabricator.wikimedia.org/P77141 and previous config saved to /var/cache/conftool/dbconfig/20250605-112518-root.json |
[production] |
11:25 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2043 (re)pooling @ 100%: Repooling', diff saved to https://phabricator.wikimedia.org/P77140 and previous config saved to /var/cache/conftool/dbconfig/20250605-112511-root.json |
[production] |
11:19 |
<jmm@cumin1003> |
START - Cookbook sre.hosts.reboot-single for host ganeti2032.codfw.wmnet |
[production] |
11:11 |
<fceratto@cumin1002> |
dbctl commit (dc=all): 'Repooling after maintenance db2179', diff saved to https://phabricator.wikimedia.org/P77139 and previous config saved to /var/cache/conftool/dbconfig/20250605-111158-fceratto.json |
[production] |
11:10 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2046 (re)pooling @ 75%: Repooling', diff saved to https://phabricator.wikimedia.org/P77138 and previous config saved to /var/cache/conftool/dbconfig/20250605-111013-root.json |
[production] |
11:10 |
<marostegui@cumin1002> |
dbctl commit (dc=all): 'es2043 (re)pooling @ 75%: Repooling', diff saved to https://phabricator.wikimedia.org/P77137 and previous config saved to /var/cache/conftool/dbconfig/20250605-111005-root.json |
[production] |
11:09 |
<jmm@cumin1003> |
START - Cookbook sre.ganeti.drain-node for draining ganeti node ganeti2032.codfw.wmnet |
[production] |
11:04 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.ganeti.drain-node (exit_code=0) for draining ganeti node ganeti2031.codfw.wmnet |
[production] |
11:03 |
<ryankemper@cumin2002> |
END (FAIL) - Cookbook sre.wdqs.data-reload (exit_code=99) reloading scholarly_articles on wdqs1023.eqiad.wmnet from DumpsSource.HDFS (hdfs:///wmf/data/discovery/wikidata/munged_n3_dump/wikidata/scholarly/20250526/ using stat1011.eqiad.wmnet) |
[production] |
11:03 |
<jmm@cumin1003> |
END (PASS) - Cookbook sre.hosts.reboot-single (exit_code=0) for host ganeti2031.codfw.wmnet |
[production] |
11:02 |
<gehel> |
restarting Blazegraph on wdqs1023 to address allocator decreasing alert |
[production] |
10:57 |
<jmm@cumin1003> |
START - Cookbook sre.hosts.reboot-single for host ganeti2031.codfw.wmnet |
[production] |