Running inside proxmox, drives passed through, after reading the problems with RAID5 I set a balance task going via the webUI to change to RAID1 it was running ok but at somepoint within the last 24 hours it looks like the VM crashed or paniced (It was sitting at the login prompt but unressponive on the console and network)
After restarting the drives appear to be totally blank, fdisk shows no partions, gdisk run on the proxmox host also shows no partitions.
Also any possiblity of data recovery?
Thankfully I backed up pretty much everything I really cared about but there’s a stil a load of other data that would be nice to recover as it would take a while to find again.
@Dragon2611 Firstly I would say not to worry about not seeing any partitions on your data disks as Rockstor only uses whole disk (ie non partitioned) btrfs configurations for it’s data disks. In this configuration btrfs deals with the disk ‘raw’ hence no partitions. Secondly I would say not to panic as on the forum there are probably a few who have run into similar problems, however I haven’t seen this one myself. So I’ll leave recovery advise to those who are more experienced with this one.
But as a start it would be good to have more info so could your first past the result of the following command into this thread to help others help you (executed as the root user):
btrfs fi show
And any notes on how that output differs from what you expect.
Also what does the Rockstor Web-UI have to say on the state of the balance / pools / drives etc.
And what does a pain mount command return?:
Well done on having backups by the way but I’d say it’s definitely worth dabbling to see if you can bring this one back, I expect something triggered a kernel panic or memory exhaustion and left you halfway way along on the raid conversion but it may very well be possible to recover from this. Note however that you may only get one chance to mount read-write in this scenario in order to do a repair so tread carefully with any manual commands, especially initially. There after your options for repair are reduced, if I understand it correctly. However there are still options to mount degraded and read only, that may well allow you to retrieve the data prior to starting a-fresh on the raid1 pool.