I noticed that I only had 5 GiB of free space left today. After quickly deleting some cached files, I tried to figure out what was causing this, but a lot was missing. Every tool gives a different amount of remaining storage space. System Monitor says I’m using 892.2 GiB/2.8 TiB (I don’t even have 2.8 TiB of storage though???). Filelight shows 32.4 GiB in total when scanning root, but 594.9 GiB when scanning my home folder.

Meanwhile, ncdu (another tool to view disk usage) shows 2.1 TiB with an apparent size of 130 TiB of disk space!

    1.3 TiB [#############################################] /.snapshots
  578.8 GiB [####################                         ] /home
  204.0 GiB [#######                                      ] /var
   42.5 GiB [#                                            ] /usr
   14.1 GiB [                                             ] /nix
    1.3 GiB [                                             ] /opt
. 434.6 MiB [                                             ] /tmp
  350.4 MiB [                                             ] /boot
   80.8 MiB [                                             ] /root
   23.3 MiB [                                             ] /etc
.   5.5 MiB [                                             ] /run
   88.0 KiB [                                             ] /dev
@   4.0 KiB [                                             ]  lib64
@   4.0 KiB [                                             ]  sbin
@   4.0 KiB [                                             ]  lib
@   4.0 KiB [                                             ]  bin
.   0.0   B [                                             ] /proc
    0.0   B [                                             ] /sys
    0.0   B [                                             ] /srv
    0.0   B [                                             ] /mnt

I assume the /.snapshots folder isn’t really that big, and it’s just counting it wrong. However, I’m wondering whether this could cause issues with other programs thinking they don’t have enough storage space. Steam also seems to follow the inflated amount and refuses to install any games.

I haven’t encountered this issue before, I still had about 100 GiB of free space last time I booted my system. Does anyone know what could cause this issue and how to resolve it?

EDIT 2024-04-06:

snapper ls only shows 12 snapshots, 10 of them taken in the past 2 days before and after zypper transactions. There aren’t any older snapshots, so I assume they get cleaned up automatically. It seems like snapshots aren’t the culprit.

I also ran btrfs balance start --full-balance --bg / and that netted me an additional 30 GiB’s of free space, and it’s only at 25% yet.

EDIT 2024-04-07: It seems like Docker is the problem.

I ran the docker system prune command and it reclaimed 167 GB!

    • qaz@lemmy.worldOP
      link
      fedilink
      arrow-up
      11
      arrow-down
      1
      ·
      3 months ago

      I’m using BTRFS with compression, so that might also explain the numbers to some extent.

      I ran df -h but I’m not exactly sure how to interpret this. There are multiple file systems which seem to use all the space on the disk.

      Filesystem               Size  Used Avail Use% Mounted on
      /dev/mapper/system-root  923G  875G   29G  97% /
      devtmpfs                 4.0M  8.0K  4.0M   1% /dev
      tmpfs                     16G   86M   16G   1% /dev/shm
      efivarfs                 128K   46K   78K  37% /sys/firmware/efi/efivars
      tmpfs                    6.3G  3.0M  6.3G   1% /run
      tmpfs                     16G  442M   16G   3% /tmp
      /dev/mapper/system-root  923G  875G   29G  97% /.snapshots
      /dev/mapper/system-root  923G  875G   29G  97% /boot/grub2/i386-pc
      /dev/mapper/system-root  923G  875G   29G  97% /boot/grub2/x86_64-efi
      /dev/mapper/system-root  923G  875G   29G  97% /home
      /dev/mapper/system-root  923G  875G   29G  97% /opt
      /dev/mapper/system-root  923G  875G   29G  97% /srv
      /dev/mapper/system-root  923G  875G   29G  97% /root
      /dev/mapper/system-root  923G  875G   29G  97% /var
      /dev/mapper/system-root  923G  875G   29G  97% /usr/local
      /dev/nvme1n1p1           511M  226M  286M  45% /boot/efi
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/f307539e15a1a33ca416c757e267c389450275eec9e7f945ef0d8680d162eac2/merged
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/8e4898a8e32696e94dd6bb5c00d02893c0b629efda7f4a8c37da2d213fe1ffab/merged
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/db20cdcf8192f6a6597a3ad8330273f0435db9d4acfa8e20ad65524ab075697f/merged
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/92ce05516bde97ae9ff6d3c6b079e7c49b6691ebcfc60b850637cab20a921ebe/merged
      tmpfs                    3.2G   17M  3.2G   1% /run/user/1000
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/5a00d8c61b23c26c87fcb3be721bc1224db7de3c9a53ae4f9bc2b922ebe40c83/merged
      overlay                  923G  875G   29G  97% /var/lib/docker/overlay2/4f20dcdebc64c2603b5b5f6ad71e116b52e8e20af2a3fe53f9ca653421f871db/merged
      
      • Corngood@lemmy.ml
        link
        fedilink
        arrow-up
        8
        ·
        3 months ago

        Try using btdu. I’m not sure how it works with compression, but it at least understands snapshots, as long as they are named in a sane way.

        • qaz@lemmy.worldOP
          link
          fedilink
          arrow-up
          2
          ·
          edit-2
          3 months ago

          Thanks for the suggestion. The repository says it is able to deal with BTRFS compression.

          I do have some issues using the application. The instructions say to run it with the filesystem you want to check as argument. However, I get an error when using it with the root filesystem from df -h --output=source,target. Running sudo btdu /dev/mapper/system-root gives the following error: Fatal error: /dev/mapper/system-root is not a btrfs filesystem. /etc/fstab shows /dev/system/root as being mounted on /, but it gives the same error.

          Do you happen to know which path I should be using (or how I can find out)?

          EDIT 2024-04-07: It seems like Docker is the problem.

          I ran the docker system prune command and it reclaimed 167 GB!

          • EinfachUnersetzlich@lemm.ee
            link
            fedilink
            arrow-up
            3
            ·
            3 months ago

            You need to point it at a directory that has the btrfs root subvolume mounted on it (subvolid=5) although I thought it gave a different error if that was your problem.

          • Corngood@lemmy.ml
            link
            fedilink
            arrow-up
            2
            ·
            3 months ago

            As the other user suggested, you probably just need to mount the root subvolume somewhere and run it on that.

      • Björn Tantau@swg-empire.de
        link
        fedilink
        arrow-up
        4
        ·
        3 months ago

        Unless you have multiple partitions or disks just concentrate on the one for /. So you have 29 GiB available.

        Everything else is sharing the same drive for different purposes.

        The beauty of BTRFS is that you can partition your disk into different parts but still actually use the whole disk for every “partition”. That makes management of snapshots easier. I think it would even enable you to combine multiple physical disks into one.

        • bitfucker@programming.dev
          link
          fedilink
          arrow-up
          2
          ·
          3 months ago

          … combine multiple physical disks into one.

          Isn’t that RAID 0 and generally a bad idea? Since one disk failure can bring down the whole system.

          • rotopenguin@infosec.pub
            link
            fedilink
            English
            arrow-up
            2
            ·
            3 months ago

            You can do “zfs style raid things” with btrfs, but there are way too many reports of it ending badly for my tastes. Something-something about “write hole”.

            • EinfachUnersetzlich@lemm.ee
              link
              fedilink
              arrow-up
              3
              ·
              3 months ago

              You can set the metadata and data independently as RAID0, RAID1 or other levels depending on the number of disks and your desired level of data loss risk.