So, I’m trying to clone an SSD to an NVME drive and I’m bumping into this “dev-disk-by” error when I boot from the NVME (the SSD is unplugged).

I can’t find anyone talking about this in this context. It seems like what I’ve done here should be fine and should work, but there’s clearly something I and the arch wiki are missing.

  • y0din@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    ·
    3 months ago

    probably the disk UUID has changed because of the path to the NVMe vs SSD. If you use partition UUID, they will be exactly the same, but the UUID of the physical disk is not cloned, as it is a identifier of the physical device and not it’s content.

    change it to partition UUID and it will boot.

    • gansheim@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      3 months ago

      Definitely second this. If you’re using LVM, it uses the physical UUID for the pv. You have to update that on the new drive so it knows where the vg and lvs are being mounted to.

      • Dark Arc@social.packetloss.ggOP
        link
        fedilink
        English
        arrow-up
        1
        ·
        3 months ago

        There wasn’t any LVM involved, it’s AFAIK pretty rare outside of MBR installs (as GPT typically lets you have more than enough partitions).

        • gansheim@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          3 months ago

          LVM is actually super common. Most Linux distros default to LVM unless you do custom partitioning. It’s not just about the max number of partitions supported by the table. LVM provides a TON more flexibility and ease of management of partitions.

          • Dark Arc@social.packetloss.ggOP
            link
            fedilink
            English
            arrow-up
            3
            ·
            edit-2
            3 months ago

            I haven’t seen LVM in any recent Fedora (very high confidence), Debian (high confidence), or OpenSUSE (fairly confident) installations (just using the default options) on any system that’s using GPT partition tables.

            For RAID, I’ve only ever seen mdadm or ZFS (though I see LVM is an option for doing this as well per the arch wiki). Snapshotting I normally see done either at the file system level with something like rsnapshot, kopia, restic, etc or using a file system that supports snapshots like btrfs or ZFS.

            If you’re still using MBR and/or completely disabling EFI via the “legacy boot loader” option or similar, then yeah they will use LVM … but I wouldn’t say that’s the norm.

            • gansheim@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              3 months ago

              That’s fair, I should have clarified that on most Enterprise Linux distros LVM is definitely the norm. I know Fedora switched to btrfs a few releases back and you may be right about Suse Tumbleweed but pretty sure Suse Leap uses LVM. CentOS, RHEL, Alma, etc. all still default to LVM, as the idea of keeping everything on a single partition is a bad idea and managing multiple partitions is significantly easier with LVM. More than likely that’ll change when btrfs has a little more mileage on it and is trusted as “enterprise ready” but for now LVM is the way they go. MBR vs GPT and EFI vs non-EFI don’t have a lot to do with it though, it’s more about the ease of managing multiple partitions (or subvolumes if you’re used to btrfs), as having a single partition for root, var, and home is bad idea jeans.

              • Dark Arc@social.packetloss.ggOP
                link
                fedilink
                English
                arrow-up
                2
                ·
                3 months ago

                That’s fair, I did just check my Rocky Linux install and it does indeed use LVM.

                So much stuff in this space has moved to hosted/cloud I didn’t think about that.

    • Dark Arc@social.packetloss.ggOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      3 months ago

      So I fixed this by using clonezilla (which seemed to fix things up automatically), but for my edification, how do you get the UUID of the device itself? The only UUIDs I was seeing seemingly were the partition UUIDs.

      • y0din@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        3 months ago

        sorry for the late reply, the command ‘lsblk’ can output it:

        “sudo lsblk -o +uuid,name”

        check “man lsblk” to see all possible combinations if needed.

        there is also ‘blkid’ but I’m unsure whether that package is installed by default on all Linux releases, so that’s why I chose ‘lsblk’

        if ‘blkid’ is installed, the syntax would be:

        “sudo blkid /dev/sda1 -s UUID -o value”

        glad you got it fixe, and hope this answers your question

        (edit pga big thumbs and autocorrect… )

        • y0din@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          3 months ago

          also, remember that the old drive now share the UUID with the NVMe drive (which is why I recommended using partition UUID and not disk UUID), so you will have to create a new GPT signature on the old drive to avoid boot issues if both drives are connected at the same time during boot, otherwise you might run into boot issues or booting from the wrong drive.