Wow, really? So, basically, since 1999 or so, I could have had a built up career because I mastered the Linux OS. I have built up a career in something else totally unrelated. Do you think I’d be richer and famouser, too? Maybe I should have just thrown myself at the technology labor market and taken control of it, like I do with the terminal app. snortreapplies tape to broken glassessnort snortreadjusts pocket protectorprefers platform games with a penguin over a guy with a moustachesnort snort
In software, it seems incredibly common for companies to give developers MacBooks and then have their software deployed on a linux VM in AWS.
It’s just one of the lower friction corporate options for software companies. The last time I used an institutionally managed linux computer was college.
There’s definitely tech jobs where you need to know linux. But there’s also a ton of jobs where you don’t have to know much of anything about it beyond common unix stuff, and where OS X specific knowledge is more useful.
When time is money, businesses give 0 shits about your Arch install, to be blunt, OSX and Apple are there to do work… Thay being said, I loves me some Unix Porn 😅
Sorry for the spicy reply. ❤️
Companies generally want something they control, so they can lock your computer and wipe it remotely when they lay you off.
They care about your arch install because they don’t want it any more than your OS X install. Their arch install would be fine, but their JAMF controlled OS X install is probably much cheaper for them to manage, practically speaking.
+1 for the corporate part. That said I’ve been at two large corpos where Ubuntu LTS was an official, blessed, IT-managed workstation OS option. In fact some development projects have it as the default OS, because some software simply doesn’t build on non-Linux OSes and requires Linux VMs on Windows or macOS. For example AOSP.
Disagree on macOS being a really good development OS. It may be so for iOS/macOS development but for many other use cases, package managed Linux OSes are superior. Case in point, macOS has no built-in package manager that can manipulate what toolchains, runtimes and libraries are installed on the system and available for software development. You have to resort to Homebrew or Macports, both of which are inferior options than say apt in Debian-based Linux OSes. And then there’s the fact that macOS doesn’t support Linux containers without virtualization. Given how useful and how widely used containerization is for software development, it really puts the nails in macOS as a great development OS. Yes you can use containerization but at the expense of significant resource overhead. That’s not great at all. I had a 6-container stack to run for development on a MacBook Pro a few years back. It was nearly unusable on that hardware due the RAM overhead and slow macOS<->container IO. The same stack was flying on equivalent hardware running Ubuntu LTS. Beyond that, you can distribute whole consistent development environments fairly easily using Linux OSes and trivially using containers. Having used Windows, Linux and macOS for professional software development, I think macOS is easier to live with for companies without IT departments, but not necessarily as easy to live with for developers. That said people who are used to work with it might find that easier than learning something else even if that something is easier in absolute terms. Which is fine.
Sure, but Linux doesn’t champion good package management. You either resort to the package system of the distro you installed, or slowly switch to flatpak/snap, and then likely the package manager of the language you are developing in. It’s no different to using brew on macOS.
I would argue basing your entire OS on the package managers makes dev add more friction, not less. Perhaps the only Linux example trying to escape that is nixos, which is nice when you get the hang of it, but it’s a little restrictive. For straight Linux development of course Linux will be the best choice. For straight macOS development of course macOS will be the best choice. For straight windows development of course windows will be the best choice. Distributing docker containers is mostly OS agnostic now, and though you mentioned it, the performance payoff for virtualization on apple M hardware is minuscule.
Most IT departments at big software companies will opt for macOS for their software devs nowadays, unless they explicitly request Linux. That’s not exclusive, bigger tech companies will let you choose because they have the bandwidth to support multiple OS’s, but they do that because supporting one *nix based OS is much easier than supporting every Linux based distro, with a different package system.
I’ve used arch. The community repo is amazing, but flawed in its own way. I’ve used Ubuntu, figuring out the PPA system sucks. I’ve used Fedora/Redhat, trying to get deb packages to work when they aren’t in the rpm library is a ride. I’ve used SUSE, Yast is great, and again become frustrating when the package you need isn’t there.
There are far more variables working with Linux distros, and usually those variables, when you’re just trying to write code, make things more difficult, not easier
At the end of the day it sounds like we’re tackling very different versions of the dev world anyway. You are trying to emulate Linux servers on apple hardware. I’m writing native apps for iOS and Android that use web views to render our components. Of course macOS wouldn’t be the best for Linux container infrastructure work, but that’s a small subset of the dev work that exists, just as my work is a small subset, however I’d still argue that on a larger scale, macOS is a better dev OS.
The technology labor market disagrees. Careers are built on mastering the Linux OS.
Wow, really? So, basically, since 1999 or so, I could have had a built up career because I mastered the Linux OS. I have built up a career in something else totally unrelated. Do you think I’d be richer and famouser, too? Maybe I should have just thrown myself at the technology labor market and taken control of it, like I do with the terminal app. snort reapplies tape to broken glasses snort snort readjusts pocket protector prefers platform games with a penguin over a guy with a moustache snort snort
In software, it seems incredibly common for companies to give developers MacBooks and then have their software deployed on a linux VM in AWS.
It’s just one of the lower friction corporate options for software companies. The last time I used an institutionally managed linux computer was college.
There’s definitely tech jobs where you need to know linux. But there’s also a ton of jobs where you don’t have to know much of anything about it beyond common unix stuff, and where OS X specific knowledge is more useful.
When time is money, businesses give 0 shits about your Arch install, to be blunt, OSX and Apple are there to do work… Thay being said, I loves me some Unix Porn 😅 Sorry for the spicy reply. ❤️
Companies generally want something they control, so they can lock your computer and wipe it remotely when they lay you off.
They care about your arch install because they don’t want it any more than your OS X install. Their arch install would be fine, but their JAMF controlled OS X install is probably much cheaper for them to manage, practically speaking.
Because Linux is really good at being a server, and macOS is really good at being a development OS, despite the hate it’s getting in this thread
I honestly think it has very little to do with the OS itself.
I think it’s more about practicalities and inertia - ordering laptops with the OS preinstalled, administering them, corporate VPN software, etc.
Both are great development OSs, but OS X is a better corporate OS.
+1 for the corporate part. That said I’ve been at two large corpos where Ubuntu LTS was an official, blessed, IT-managed workstation OS option. In fact some development projects have it as the default OS, because some software simply doesn’t build on non-Linux OSes and requires Linux VMs on Windows or macOS. For example AOSP.
Disagree on macOS being a really good development OS. It may be so for iOS/macOS development but for many other use cases, package managed Linux OSes are superior. Case in point, macOS has no built-in package manager that can manipulate what toolchains, runtimes and libraries are installed on the system and available for software development. You have to resort to Homebrew or Macports, both of which are inferior options than say apt in Debian-based Linux OSes. And then there’s the fact that macOS doesn’t support Linux containers without virtualization. Given how useful and how widely used containerization is for software development, it really puts the nails in macOS as a great development OS. Yes you can use containerization but at the expense of significant resource overhead. That’s not great at all. I had a 6-container stack to run for development on a MacBook Pro a few years back. It was nearly unusable on that hardware due the RAM overhead and slow macOS<->container IO. The same stack was flying on equivalent hardware running Ubuntu LTS. Beyond that, you can distribute whole consistent development environments fairly easily using Linux OSes and trivially using containers. Having used Windows, Linux and macOS for professional software development, I think macOS is easier to live with for companies without IT departments, but not necessarily as easy to live with for developers. That said people who are used to work with it might find that easier than learning something else even if that something is easier in absolute terms. Which is fine.
Sure, but Linux doesn’t champion good package management. You either resort to the package system of the distro you installed, or slowly switch to flatpak/snap, and then likely the package manager of the language you are developing in. It’s no different to using brew on macOS.
I would argue basing your entire OS on the package managers makes dev add more friction, not less. Perhaps the only Linux example trying to escape that is nixos, which is nice when you get the hang of it, but it’s a little restrictive. For straight Linux development of course Linux will be the best choice. For straight macOS development of course macOS will be the best choice. For straight windows development of course windows will be the best choice. Distributing docker containers is mostly OS agnostic now, and though you mentioned it, the performance payoff for virtualization on apple M hardware is minuscule.
Most IT departments at big software companies will opt for macOS for their software devs nowadays, unless they explicitly request Linux. That’s not exclusive, bigger tech companies will let you choose because they have the bandwidth to support multiple OS’s, but they do that because supporting one *nix based OS is much easier than supporting every Linux based distro, with a different package system.
I’ve used arch. The community repo is amazing, but flawed in its own way. I’ve used Ubuntu, figuring out the PPA system sucks. I’ve used Fedora/Redhat, trying to get deb packages to work when they aren’t in the rpm library is a ride. I’ve used SUSE, Yast is great, and again become frustrating when the package you need isn’t there.
There are far more variables working with Linux distros, and usually those variables, when you’re just trying to write code, make things more difficult, not easier
At the end of the day it sounds like we’re tackling very different versions of the dev world anyway. You are trying to emulate Linux servers on apple hardware. I’m writing native apps for iOS and Android that use web views to render our components. Of course macOS wouldn’t be the best for Linux container infrastructure work, but that’s a small subset of the dev work that exists, just as my work is a small subset, however I’d still argue that on a larger scale, macOS is a better dev OS.
Precisely one niche where mastering the Linux OS provides bread.
I mean, imagine running legacy apps… In Tomcat. I’ve seen into the void…
Don’t have to imagine it. 🥲