Curious to know the coolest things you achieved by configuring your kernel. I know kernel config can be boring, but I’m hoping someone will have an impressive answer.

For me I have a very lightweight kernel that runs wayland on nvidia without any issues to date.

  • ExtremeDullard@lemmy.sdf.org
    link
    fedilink
    arrow-up
    133
    ·
    8 months ago

    As a Linux user of almost 30 years, compiling hundreds of kernels over the years has given me a great appreciation of pre-build kernels, and a profound gratitude for those who package them up into convenient distros that work out of the box and let me get on with the rest of my life.

    • limelight79@lemm.ee
      link
      fedilink
      arrow-up
      19
      ·
      8 months ago

      Well said. I originally compiled my own kernels because I thought it was something you just did to use Linux. I also compiled hundreds of them, probably. Now it’s stock kernel all the way. Not worth the effort and time and headache.

      • ExtremeDullard@lemmy.sdf.org
        link
        fedilink
        arrow-up
        10
        ·
        8 months ago

        Absolutely! If you’re doing it to learn something, by all means compile your own kernel. Every Linux user should do that at least once in my opinion. But once the learning is done, the novelty wears off fast and it just becomes tedious.

    • ikidd@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      edit-2
      8 months ago

      I used Linux for about a decade from the mid nineties then took a break for a few years. When I came back, every distro kernel was precompiled, it was glorious. There was never a day I said to myself “damn, I miss compiling a kernel”.

  • ChojinDSL@discuss.tchncs.de
    link
    fedilink
    arrow-up
    43
    ·
    8 months ago

    Back when I was still using Gentoo, configuring your own kernel was a rite of passage. It was kind of fun to try and configure it as minimalist as possible to cut down on the kernel compile time. Also, understanding all the different options and possibilities. And thanks to use flags, you had access to all these different patch sets for the kernel, which took a lot of the pain out of trying things like experimental schedulers or filesystems.

  • thejml@lemm.ee
    link
    fedilink
    English
    arrow-up
    20
    ·
    8 months ago

    Years ago (2006-ish), I ran Gentoo on a 300mhz ultra low power system I used for an irc & web server. I gained LOTS of speed and lowered power draw even further while also enabling the hardware acceleration the board had for ssl encryption and video encoding. The whole thing would pull <5 watts and be super stable. It was well worth it.

    But now days a Pi zero would trounce it in both low power draw and speed with stock kernels and I don’t really care enough to try to squeeze more out.

    • CarrotsHaveEars@lemmy.ml
      link
      fedilink
      arrow-up
      3
      ·
      8 months ago

      Customising the kernel just means something works properly in rare hardware configurations like you described. It’s something which he who uses the general hardware (like an X86 desktop) can’t easily see or understand because the ‘stock’ kernel is already working properly.

  • nyan
    link
    fedilink
    arrow-up
    15
    ·
    8 months ago

    Well, I can still boot my system without an initram (although that isn’t just due to the kernel config)—does that count?

    Other than that, custom kernels free up a small amount of disk space that would otherwise be taken up by modules for driving things like CANbus, and taught me a whole lot about the existence of hardware and protocols that I will never use.

      • Rikj000@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        19
        ·
        8 months ago

        Amazing, basically native speeds,
        currently playing Horizon Forbidden West with maxed out graphics and DRS disabled at a steady 60-80 FPS.

        Previously I also played Horizon Zero Dawn in it, also maxed out graphics, steady locked 100 FPS,
        below is a benchmark comparison of HZD in the Linux host OS and the Windows KVM guest OS:
        workstation-gaming-linux-vs-windows

        • ikidd@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          8 months ago

          Has this gotten any easier to do? I set it up a few years ago, it was painful to do and maintain so I let it slide. You were writing all sorts of scripts to specify the passthrough devices and then they’d stop working so you had to track down what was failing and update. Then there was iommu so you had to be careful which groups you added devices to.

          • Rikj000@discuss.tchncs.de
            link
            fedilink
            English
            arrow-up
            4
            ·
            8 months ago

            Gotta admit, it was very hard to setup initially.
            However it’s been working perfectly ever since I did.
            Been using it for about a year or 2 now.

            Also when I linked the Arch wiki,
            I noticed in it’s article that there’s now a gpu-passthrough-manager,
            which will likely make the process of setting up a little bit easier.

    • Pantherina@feddit.de
      link
      fedilink
      arrow-up
      1
      arrow-down
      1
      ·
      8 months ago

      Root Waydroid lol, thats basically hell.

      Waydroid without SELinux already removes all the Android sandboxing. Now its rooted!

      • Rikj000@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 months ago

        Root on Android is a necessity for me.
        I’ve been rooting all droids I use for the past 10 years or so.

        Imagine using Linux as a power user,
        without being able to use sudo/su.

        Also, Magisk does not just allow any application to access root, you have to manually allow apps to make use of it.

        Just like administrator rights on any other OS,
        things only go wrong if you don’t know what you’re doing, and then grant rights to something malicious.

  • xycu@programming.dev
    link
    fedilink
    English
    arrow-up
    14
    ·
    edit-2
    8 months ago

    I do it because I can… I read release notes on every update and once you’ve configured a kernel for a particular machine you really don’t need to touch the config, barring major changes like when PATA and SATA merged. Or of course if I’m adding a new piece of hardware.

    I remove everything I don’t need and compiling the kernel only takes a couple minutes. I use Gentoo and approach everything on my system the same way - remove the things I don’t need to make it as minimal as possible.

    Compiling your own kernel also makes it easier when you need to do a git bisect to determine when a bug was introduced to report it or try to fix it. I’ve also included kernel patches in my build years ago, but haven’t needed to do that in a long time.

    I used to compile a custom kernel for my phone to enable modules/drivers that weren’t included by default by the maintainer.

    It’s not about performance for me, it’s about control.

      • ExtremeDullard@lemmy.sdf.org
        link
        fedilink
        arrow-up
        8
        arrow-down
        3
        ·
        edit-2
        8 months ago

        Compiling kernels makes no sense anymore.

        Back in the days - Linux versions 2 and below - the kernel was much less modular, and resources wasn’t as plentiful. So it often made sense to build kernels with the stuff you needed statically compiled for speed, and the rest left out fo save memory and shorten boot time. Not to mention, Lilo (the thing we used before Grub) had limitations with respect to kernel size.

        Nowadays, Grub can load a kernel of any size from anywhere on the disk. There’s plenty enough memory and CPU to leave the kernel core slightly bloated with stuff almost nobody needs with zero practical impact on boot time and memory usage, and most everything else is compiled as modules and loaded as needed - again with next to no boot time or running speed impact.

        If you custom-build a kernel today, you’ll boot a tiny bit faster and it’ll run a tiny bit faster, and you’ll have a tiny bit more free memory - all of which you will never notice. What you will notice however is that kernel updates are a PITA on a regular basis.

        • taladar
          link
          fedilink
          arrow-up
          3
          ·
          8 months ago

          Kernel updates are extremely easy when custom compiling, not sure what you are doing to make them a pain. Custom compiling is a great way to sort of passively absorb knowledge about kernel changes and new features or features you didn’t know about as they change and make oldconfig brings up questions about them.

  • galoisghost@aussie.zone
    link
    fedilink
    arrow-up
    13
    ·
    8 months ago

    Mostly just understanding what was there, what was necessary for my machine at the time and what was optional.

  • zarenki@lemmy.ml
    link
    fedilink
    arrow-up
    8
    ·
    8 months ago

    I have configured custom Android kernel builds to enable more USB drivers, enable module support, and tweak various other things. For one tangible example of the result: I could plug in a USB Wi-Fi adapter and use it to simultaneously connect to another Wi-Fi network with the internal NIC while also sharing my own AP over USB. On an Android device of all things. I have also adjusted kernel builds for SBCs (like Pi clones) to get things working at all.

    I have never seen any reason to configure a custom kernel for my own desktop/laptop systems. Default builds for the distros I’ve used have been fine for me; if I’m ever dissatisfied with anything it’s the version number rather than the defconfig. The RHEL/Rocky kernel omits a few features I want (like btrfs) but I’d rather stick to other distros on personal systems than tweak a distro that isn’t even meant for tweaking.

  • pastermil
    link
    fedilink
    arrow-up
    7
    ·
    8 months ago

    I’m playing around with coreboot and that gives me ability to embed Linux kernel. The problem is we’re limited by the amount of ROM chip which is between 4MiB to 16MiB depending on the specific device. The one I’m working on got 12MiB, about 3 is taken in order to boot normally, leaving me with 9 to play around.

    Enter buildroot, (arguably) a Linux distro that allows you to have kernel, busybox, minimum libc, along with whatever software you’d choose.

    While it’s easy to include only what’s needed to have a working system (busybox provides working shell as well as the coreutils), you’d need to get rid of stuff you don’t need, such as drivers for hardware you wouldn’t have.

    Aside from that, you’d end up with better running kernel in general if you know what you’re doing. I run Gentoo and have kept a working config that I tweak from time to time (especially on version upgrade).

  • atzanteol
    link
    fedilink
    English
    arrow-up
    6
    ·
    edit-2
    8 months ago

    I haven’t custom compiled a kernel in ages - does anyone still do this?

    It used to be sorta-kinda-necessary back when memory sizes were measured in MB instead of GB. Kernels had to be under a certain size, the module system was a bit slow, memory was at a premium, hardware support was very spotty, etc. I remember applying some guy’s patches to a 2.2 kernel to get full-duplex sound on the crappy sound card my Pentium 120 had (Linux has always had garbage audio support).

    I think the last time I purposefully created a custom kernel was to enable some experimental scheduler code I hoped would give a performance boost. Was many years ago though.

    These days you only do it if you want to learn the process or performance test a system. Or if you’re running something like Gentoo - but even then you likely just use the ‘default’ configuration provided by Gentoo.