I opened the fstab and added compress=zstd:1 to both / and /home as thats how its done in Fedora 34.
Updated fstab and also rebooted the machine to ensure that daemons reload.
Everything is working nicely until now but now to convert existing things into btrfs compression, How to test tells us to btrfs filesystem defrag -czstd -r
so do I have to sudo btrfs filesystem defrag -czstd -r / and also sudo btrfs filesystem defrag -czstd -r /home
Ya I figures this our too but just had some doubts, thanks ! I got into a small issue, my system is showing 0 in available space, all tho total and used is great and I can install everything so its just a cosmetic issue but still any idea ? @nickavem
Apparently same for me, solved with df -H or duf. But this “100% used” has not been indicated with the compression but with a Fed34 install in Gnome-Boxes (Machines). I suppose it is a bug in gnome-monitor. Ouups sorry, correct, but I have performed the step compress/decompress as indicated and it is ok now.
du, df, and other utilities can be somewhat funky with the compression. The best utility to use is compsize which I mentioned at the end of the other post.
Hello Jatin (and Chris), yes it is ok now. I thought gnome-system-monitor had a bug, as it indicated 100 % used. But after application of the other Jatin’s post (sudo btrfs filesystem defrag -czstd -rv / /home/jcvminou), same results as df -H are now indicated. Thank you.
A backup drive can be either compressed or not. I have two backup drives which are compressed. And one backup drive that isn’t, mainly because it has a ton of free space on it, and most files are formats that are already compressed.
OK; it´s done. I followed all steps and applied compression to both drives and all looks fine. In the case of the backup drive and as anticipated, I applied a full reformat to BTRFS. Then, I used Deja-Dup as backup tool, instead of Timeshift. Once backup was completed, modified fstab. So far, so good. Thanks for your support.
Also you said you did (sudo btrfs filesystem defrag -czstd -rv / /home/jcvminou) , Thats what I did to convert existing system. Is it fine to redo this command (I think it will just compress again) but if someone can confirm.
gnome-system-monitor should change the backend from libgtop to something used by baobab (disk usage analyzer).
It’s just a verbose switch if you want to see what it’s doing.
Also you said you did (sudo btrfs filesystem defrag -czstd -rv / /home/jcvminou) , Thats what I did to convert existing system. Is it fine to redo this command (I think it will just compress again) but if someone can confirm.
It’ll submit everything for defrag/compression again.
It’s not strictly necessary to do this step. You can just leave things uncompressed, and change /etc/fstab to enable compression for all new writes.
@chrismurphy I was kind of able to re produce the issue, I did defrag and compression again and it solved my gnome system monitor issue. And I saw that live in action as I kept my system monitor on and it fixed it in between. However on my other system I did defrag and compress and kept my eye on the system monitor, it created the same bug but during the same run it fixed it as well. So maybe re iterating defrag and compression on some systems can reproduce this bug. On my laptop I was not able to reproduce the bug tho I did defrag and compression 5 times and couldn’t reproduce the issue.
Pass this info to the same bugzilla issue. Currently the solution is just defrag and compress again
Thanks for this post. However we were discussing issue with gnome system monitor (specifically libgtop) showing 100 percent disk usage after enabling compression on some systems. We solved this by doing defrag/compress again. This also happens on few newer installations as well. Just a cosmetic issue. But we have an issue on on bugzilla and gitlab. @nickavem
I still can’t reproduce it. If it happens again, the best thing to do is try to get some debugging information out of system-monitor/libgtop rather than stomping on things with a heavy weight operation like defragmenting again. That just makes it harder to find someone who has the problem to get us the debugging information we need.