Earlier this year I replaced all my drives with higher capacity ones (Advice needed on All-Drives replacement), and followed the approach of adding one new one to have enough capacity overhead and then one after the other remove and replace my other drives (with all the steps required in between using the WebUI).
While investigating another issue 9 months later I found that one of the btrfs tasks for removing a device keeps on attempting to do so (and obviously failing because the device has not been there in a long time).
Looking at ztask.log I can see this:
When I look at the ztask.log I am seeing something curious (have not looked in that log I think, ever) …
2020-10-09 07:57:48,306 - ztaskd - INFO - Calling fs.btrfs.start_resize_pool 2020-10-09 07:57:48,308 - ztaskd - ERROR - Error calling fs.btrfs.start_resize_pool. Details: Error running a command. cmd = /usr/sbin/btrfs device delete /dev/disk/by-id/ata-HGST_HDN724040ALE640_PK1334PEHZB3ZS /dev/disk/by-id/ata-HGST_HDN724040ALE640_PK1334PEHZNA2S /mnt2/4xRAID5. rc = 1. stdout = ['']. stderr = ['ERROR: not a block device: /dev/disk/by-id/ata-HGST_HDN724040ALE640_PK1334PEHZB3ZS', 'ERROR: not a block device: /dev/disk/by-id/ata-HGST_HDN724040ALE640_PK1334PEHZNA2S', '']
I have not triggered or scheduled a pool-resizing, since I replaced all my Hard drives earlier in the year. I don’t even have these HGST drives physically in the system anymore (I replaced those with WDs). In the WebUI I have not seen any errors surfaced related to this, and obviously this is not a regularly scheduled task one would want to set up (if it were possible).
@phillxnet you had already mentioned in another topic that you will look into it, just not now with everything going on for the rebase. I can alternatively create an issue on GitHub, but if I’m the only one experiencing this then I don’t want to clutter up the issue lists with this.