phillxnet
(Philip Guyton)
August 30, 2020, 1:03pm
5
@NoirXIII Re:
OK, that’s a shame.
NoirXIII:
… would updating to the latest stable version, which I guess it isn’t updating automatically anymore, help with this process as it sounds like I may need to just pull a drive and replace it with a new one which is what I was hoping to avoid.
Definitely worth updating as we added some major fixes since 3.9.2-28, see:
3.9.2-57
Merged 19th April 2020
Released 25th April 2020
We are now starting to move back to our older, hopefully more robust model of releasing first into our testing channel updates and then into stable if no ‘worries’ were reported. But during our transition as part of the ‘Built on openSUSE’ endeavour we only have an openSUSE based testing channel. But as we move closer to achieving feature parity with our current stable updates channel, which is based on CentOS, we can hopefully soon desi…
and specifically to your circumstance of drive removal there was a very significant, and long awaited fix in 3.9.2-49:
3.9.2-49
Merged end September 2019
Released 2nd October 2019
I am chuffed, and somewhat relieved, to finally release what is one of our largest and longest awaited releases for some time now. I present the 81st Rockstor.
In this release we welcome and thank first time Rockstor contributor @Psykar for a 2 in one fix for long standing issues of table sorting; nice:
Fix size sorting on snapshots / shares pages. Fixes #1368 #1878 @Psykar on-Github
Thanks also to @Flox for a wide range of non …
notable this error bug referenced in the above:
opened 06:19PM - 01 Jun 17 UTC
closed 02:10PM - 09 Jul 19 UTC
Thanks to forum member Noggin for highlighting this behaviour. Occasionally when… removing a disk from a pool there can be a UI time out directly after the last dialog entitled "Resize Pool / Change RAID level for ..." which acts as last confirmation of the configured operation:

There is then no UI 'balance' indicated while the removal is in progress, yet the UI indicates that a balance is in progress when a balance is attempted (only attempted by Noggin as I did not attempt to execute a balance whilst the removal was in progress).
```
btrfs balance status /mnt2/time_machine_pool/
No balance found on '/mnt2/time_machine_pool/'
```
The pool resize is however indicated by the requested disk's having their size 'demoted' to zero and showing a reduced usage with subsequent executions of **btrfs fi show**:
```
Label: 'time_machine_pool' uuid: 8f363c7d-2546-4655-b81b-744e06336b07
Total devices 4 FS bytes used 31.57GiB
devid 3 size 149.05GiB used 17.03GiB path /dev/sdd
devid 4 size 0.00B used 5.00GiB path /dev/sda
devid 5 size 149.05GiB used 23.03GiB path /dev/mapper/luks-d36d39ea-c0b3-4355-b0c5-bd3248e6bbfe
devid 6 size 149.05GiB used 23.00GiB path /dev/mapper/luks-d7524e90-4d9e-4772-932f-d1407b6b5fe7
```
and then later on:
```
Label: 'time_machine_pool' uuid: 8f363c7d-2546-4655-b81b-744e06336b07
Total devices 4 FS bytes used 32.57GiB
devid 3 size 149.05GiB used 18.03GiB path /dev/sdd
devid 4 size 0.00B used 2.00GiB path /dev/sda
devid 5 size 149.05GiB used 24.03GiB path /dev/mapper/luks-d36d39ea-c0b3-4355-b0c5-bd3248e6bbfe
devid 6 size 149.05GiB used 24.00GiB path /dev/mapper/luks-d7524e90-4d9e-4772-932f-d1407b6b5fe7
```
As can be seen devid 4 is having it's pool usage reduced (from 5 to 3 GB) between runs. In the above example the disk removal completed successfully however there was never an UI indication of it's 'in progress' nature or any record of a balance having taken place at that time.
Reference to Noggins's forum thread suspected as indicating the same as my observations in final testing of pr #1716 which lead also to this issue creation (details of the precedence steps available in that pr):
https://forum.rockstor.com/t/cant-remove-failed-drive-from-pool-rebalance-in-progress/3319
where a 3.8.16-16 (3.9.0 iso install) version exhibited the same behaviour (pre #1716 merge).
If you system is not offering you 3.9.2-57 then let me know via PM here on the forum, along with your Appliance ID and order email and I’ll look into why this is, and we can get you sorted one way or another. A subset of our longer term subscribers were affected by a migration bug when we moved to Appman (https://appman.rockstor.com/ ) that unfortunately need an admin to fix. We note this in the following forum post:
Just a friendly reminder that our last and now old CentOS based testing channel had a rather deceitful bug.
If you updated to Stable from 3.9.1-16 (last available CentOS based testing release), you have likely been affected by this issue. But if you subscribed to Stable directly from our last iso install, or a prior Stable channel install then you are likely not affected.
Essentially this bug would cause available version to be displayed as installed version within the Web-UI.
If you have sub…
That would at least get you the important Rockstor package fixes but you might also want to consider re-installing to get the new Rockstor 4 and the newer btrfs which is known to be much faster. But as mentioned you will have to build you own installer currently which may not be to your taste / available time. Either way best make sure you aren’t affected by this known update bug.
Agreed it’s bad to put a pool into a degraded state, if that is what you were considering. But if there is room you can remove a disk via the Web-UI (given the above update) and once done add the new one.
But if not room then best to explore other temporary means of adding a sata port. Does the machine have any pci slots?
Hope that helps.