For those veteran linux people, what was it like back in 90s? I did see and hear of Unix systems being available for use but I did not see much apart from old versions of Debian in use.
Were they prominent in education like universities? Was it mainly a hobbyist thing at the time compared to the business needs of 98, 95 and classic mac?
I ask this because I found out that some PC games I owned were apparently also on Linux even in CD format from a firm named Loki.
Ah, yes, Linux around the turn of the century. Let’s see…
GPU acceleration? In your dreams. Only some cards had drivers, and there were more than 2 GPU manufacturers back then, too… We had ATi, nVidia, 3dfx, Cirrus, Matrox, Via, Intel… and almost everyone held their driver source cards close to their chest.
Modems? Not if they were “winmodems”, which had no hardware controller, the CPU and the Windows driver (which was always super proprietary) did all the hard work.
Sound? AC’97 software audio was out of the question. See above. You had to find a sound blaster card if you wanted to get audio to work right.
So, you know how modern linux has software packages? Well, back then, we had Slackware, and it compiled everything gentoo style back then. In addition, everyone had a hardon for " compiling from source is better"… so your single core Pentium II had to take its time compiling on a UDMA66-connected hard drive, constrained with 32 or 64 MB RAM. Updating was an overnight procedure.
RedHat and Debian were godsends for people who didn’t want to waste their time compiling… which unfortinately was more common even so, because a lot of software was source only.
Oh, and then MP3 support was ripped out of RedHat in Version 9 iirc, the last version before they split it into RHEL and Fedora. RIP music.
As for Linux on a Mac, there was Yellowdog, which supported the PPC iMacs and such. It was decently good, but I had to write my own x11 monitor settings file (which I still have on a server somewhere, shockingly, I should throw it on github or somewhere) to get the screen to line up and work right.
Basically, be glad Linux has gone from the “spend a considerable amount of time and have programming / underhood linux knowledge to get it working” to “insert stick, install os, start using it” we have now.
Driver support was so dicey. If you had anything even remotely not mainstream, you would be compiling your own video driver, or network driver, or basically left to figure it out for any other peripheral. So many devices like scanners and very early webcams just claimed zero Linux support at all, but you could at times find someone else’s project that might work.
I tried to switch to Linux as a desktop system several times in the late 90s but kept going back to windows because hardware support just wasn’t there yet.
You have several long and comprehensive answers so please allow me to add an emotional one:
Fucking compile error in hour six of what you estimated to be a four hour compile job because of a mistake you made that you found within 5 seconds after the error!!
Fucking why doesn’t this compilation start I can’t find my mistake for hours?!
Where does this module come from?! What do you mean “root kit”? Learning was fun!
It all was fun! :)
Everything was harder back then - even when using Windows. But you had to be a real masochist to run Linux.
Computers were still quite new that most people had no real use for them beyond “work things”. Only nerds really used them for anything else. “Do you have an email address” isn’t a question you ask today.
“Kids these days” don’t realize how easy they have it when it comes to just general comparability. There weren’t a lot of standards yet and vendors had proprietary drivers and offered no support AT ALL for “lye nux”. You had to do a ton of research and fiddling to figure out if there was any support for your specific version of a specific chip used by any peripheral you used. And then to discover that you had to patch your kernel to add a driver that somebody had bodged together. So now you were running your own fun custom-kernel so you could get full-duplex rather than simplex audio! But it works!
Like - lets say today you want to buy an external IDE drive controller to plug in some old drives to for backup. You to to Amazon, search “USB external IDE enclosure” and buy the cheapest one you find. It probably works unless it’s defective. In '95 USB and Firewire were in their infancy so you would probably buy a serial or parallel port device. You would need to find whether Linux supported the specific version of the thing you wanted to buy, what tools there would be for it, etc. There was no standard “bulk storage device” driver that you could rely on or hope the vendor would implement. Even if you were an early adopter and got a USB or Firewire device it might have some “basic” functionality that works with OSS drivers but you couldn’t use all of it.
Vendors back then also shipped their own software with things, not just drivers. It was always just the absolute worst crap that was buggy as shit. But it would do a lot of the heavy lifting in working with their device. Like any Creative Labs audio player you wanted to get working. Sure it used USB but it didn’t just mount as storage device, you needed to use the worst GUI ever put before mankind to use it (under Windows). Under Linux you had to find a specific tool that would support pushing/pulling media from it. These days it would just mount as a drive automatically and you’d use standard desktop tools to interact with it.
Even with DOS/Windows you’d buy a game and as you came home from the store with it in a box wonder “will this work on my computer and how long will I need to mess with it?” I had to configure a specific CD-ROM driver to be used by DOS just to run Tie Fighter vs. X-Wing for example. Had a special boot floppy just for that game since that driver didn’t work with literally anything else I had.
Hardware just generally didn’t “auto configure”. “Plug 'n Play” was still very much in its infancy and you often had to manually configure hardware and install special drivers just for a particular card or peripheral.
IRQ 7 DMA 220. I probably just triggered some folks. If you were setting up a “Sound Blaster or compatible” then you had to know what interrupt it used (7) and what address it was on on the direct-memory bus (220). And you hoped there wasn’t a conflict with something else. If there was then there would be a DIP switch you could use to change the base memory address or IRQ from the default. But you were telling your software where to find the card.
USB was a f’ing game changer for peripherals. Serial and parallel ports were so slow and obnoxious to use. Before that there was no real way to “probe” the bus to discover what was there unless you knew exactly what you were looking for (there’s no
lsusb
for serial ports). So you just guessed at the driver you need and “modprob foo” hoping it worked.It’s amazing what 20ish years of just developing standards has done.
If you want a taste of that world I highly recommend LGR on YouTube. He’s mostly Windows focused but look for videos where he tries out “oddware” to see how often he has trouble getting things to work on period hardware using the vendor-supplied software even. Then multiply that by 100x for Linux. :-)
Not a veteran, but… During the 90s, while still in primary school, a friend of mine bought a Chip magazine with a CD attached and instructions inside the magazine how to install a mysterious thing called “Linux” from said CD. It was supposed to be something like Windows 95, but new, better and it had a Penguin on it, so we decided to try it.
We followed magazine’s installation guide to the letter (or at least we thought so) until the installation stuck at error saying
KERNEL PANIC!!!
and wouldn’t let us finish. We didn’t understand English much back then, but we found the panicking kernel hilarious. Anyway, we figured it’s been enough h4Ck!nG for that day and got back to playing Diablo 1.Well, XFree86 ( before Xorg and before KMS ) was an adventure. I spent hours guessing the vertical and horizontal frequencies of my monitor trying to get decent resolutions.
Other than that, Linux was way more work but “felt” powerful relative to OS options of the time. Windows was still crashy. The five of us that used OS/2 hated that it still had a lot of 16 bit under the hood. Linux was pure 32 bit.
Later in the 90’s, you could run a handful of Windows apps on Linux and they seemed to run better on Linux. For example, file system operations were dramatically faster.
And Linux was improving incredibly rapidly so it felt inevitable that it would outpace everything else.
The reality though was that it was super limited and a pain in the ass. “Normal” people would never have put up with it. It did not run anything you wanted it to ( if you had apps you liked on Mac, Windows, OS/2, Amiga, NeXTstep, BeOS, or whatever else you were using ( there were of potential options at the time ). But even for the pure UNIX and POSIX stuff, it was hard.
Obviously installation was technical and complex. And everything was a hodge-podge of independently developed software. “Usability” was not a thing. Ubuntu was not release until 2004.
Linux back then was a lot of hitting FTP sites to download software that you would build yourself from source. Stuff could be anywhere on the Internet and your connection was probably slow. And it was dependency hell so you would be building a lot of software just to be able to build the software you want. And there was a decent chance that applications would disagree about what dependencies they needed ( like versions ). Or the config files would be expected in a different location. Or the build system could not find the required libraries because they were not where the Makefile was looking for them.
Linux in the 90’s had no package management. This is maybe the biggest difference between Linux then and Linux now. When package management finally arrived, it came in two stages. First, came packages but you were still grabbing them individually from FTP. Second came the package manager which handled dependencies and retrieval.
The most popular Linux in the mid to late 90’s was Red Hat. This was before RHEL and before Fedora. There was just “Red Hat Linux”. Red Hat featured RPMs ( packages ) but you were still installing them and any required dependencies yourself at the command line. YUM ( precursor to NRF ) was not added until Fedora Core 1 was release in 2003!
APT ( apt-get ) was not added to Debian until 1998.
And all of this meant that every Linux system ( not distro — individual computer ) was a unique snowflake. No two were alike. So bundling binary software to work on “Linux” was a real horror-show. People like Loki gave it a good run but I cannot imagine the pain they went through. To make matters worse, the Linux “community” was almost entirely people that had self-selected to give up pre-packaged software and to trade sweat-equity for paying for stuff. Getting large number of people to give you money for software was hard. I mean, as far as we have come, that is still harder on Linux than on Windows or macOS.
You can download early Debian or Red Hat distros today if you want to experience it for yourself. That said, even the world of hardware has changed. You will probably not be wrestling IRQs to get sound or networking running on modern hardware or in a VM. Your BIOS will probably not be buggy. You will have VESA at least and not be stuck on VGA. But all of that was just “computing” in the 90’s and the Windows crowd had the same problems.
One 90s hardware quirk was “Windows” printers or modems though where the firmware was half implanted in Windows drivers. This was because the hardware was too limited or too dumb to work on its own and to save money your computer would do some of the work. Good luck having Linux support for those though.
Even without trying old distros, just try to go a few days on you current Linux distro without using apt, nrf, pacman, zypper, the GUI App Store, or what have you. Imagine never being able to use those tools again. What would that be like?
Finally, on my much, much slower 90’s PC, I compiled my own kernel all the time. Honestly multiple times per month I would guess. Compiling new kernels was a significant fraction of where my computing resources went at the time. I cannot remember the last time I compiled a kernel.
It was a different world.
When Linus from LTT tried Linux not that long ago ( he wanted to game ), he commented that he felt like he was playing “with” his computer instead of playing “on” his computer. That comment still describes Linux to some extent but it really, really captures Linux in the 90’s.
This was me, you’re talking about me. 😂 In the 90’s Linux was barely getting started but slackware was probably the main distro everyone was focused on. That was the first one I ran across. This was probably late 90’s, I don’t remember when slack first came about though.
By the time the 2000’s came around, it was basically a normal thing for people in college to have used or at least tried. Linux was in the vernacular, text books had references to it, and the famous lawsuit from SCO v IBM was in full swing. There were distro choices for days, including Gentoo which I spent literally a week getting everything compiled on an old Pentium only for it to not support some of the hardware and refuse to boot.
There was a company I believe called VA Linux that declared that year to be the year of the Linux desktop. My memory might be faulty on this one.
Loki gaming was a company that specialized in porting games to Linux, and they did a good job at it but couldn’t make money. I remember being super excited about them and did buy a few games. I was broke too so that was a real splurge for me. I feel like they launched in the 90’s (late) and crashed in the early 2000’s.
Ah, the golden age. I dunno, I was running FreeBSD. It was awesome. It still is.