I really wish that I was born early so I’ve could witness the early years of Linux. What was it like being there when a kernel was released that would power multiple OSes and, best of all, for free?
I want know about everything: software, hardware, games, early community, etc.
All I know is I wish I would’ve stuck with it when I first installed but…alas…I was lazy and too dependent on GUI
The first time I ever used Linux was in high school around 2001-2002. I don’t remember what the distro was but it had drawing issues, clearly some kind of driver issue that I couldn’t figure out, on my PC so I switched back to Windows 98SE.
Not what op asked for, but it kept away from Linux at home until 2007. I started using Linux regularly in university around 2004.
I didn’t have a Pentium processor in my computer, the internet was young, information wasn’t as ready or available, and the mindset wasn’t that you could check everything. I don’t remember how many floppy disks it took to install Slackware, but at least one read error was definitely on the way. I had a 56k modem at home, so I had printed out the installation instructions from work. Compiling everything wasn’t a problem, because I learned to code back in 1983. When I tried to figure out the refresh rate of my screen, I was afraid I would blow it up and go blind. The feeling of freedom was when you were the one who could choose everything for the first time in your virtual life.
Relevant xkcd’s
Hmm my first linux distro was Suse 5.x that came on 5 CDs (i think it was 1998) … can’t say I used it much, I had weird German ISDN Internet at the time and the PPPoverWhatever (forgot the exact name) just didn’t wanna work. Making music wasn’t really feasible at the time. It mostly lay dormant. I slowly climbed the learning curve and switched to Linux full-time in the mid-2000s, when a lot more things were possible …
Games: xbill, koules, and quake1 prerelease test(8 or 16 player multi)
Crafting XFree86 config lines to get a monitor working(no auto-detect for resolution modes)
Sharing tips, on how to solve all these issues, with others at Linux User Groups(LUGs)
Contrary to other OSes, the information about it was mainly on the internet, no books or magazines. With only one computer at most homes, and no other internet-connected devices, that posed a problem when something didn’t work.
It took me weeks to write a working X11 config on my computer, finding all the hsync/vsync values that worked by rebooting back and forth. And the result was very underwhelming, just a terminal in an immovable window. I think I figured out how to install a window manager but lost all patience before getting to a working DE. Days and days of fiddling and learning.
Lol! 'Member Afterstep?
The desktop stretched across 4 screens was enough to hook me for life.
Xeyes… so many terminals… the artwork was artwork… wtf is transparency?! 😁 It was an amazing time to be a geek.
I didn’t get that far. And I only had an Amiga at that time, which made things more difficult to set up. I wonder how fluent transparency would be with AGA, haha. My next attempt was woth a PC around 2003 with KDE3 and it got me hooked.
I got tired of compiling the kernel taking a day on my Pentium pc. So I got a pile of 486s the uni was throwing out, built a Beowulf cluster out of them and soon I was able to compile the kernel in two and half days.
The only OS that was solid as a desktop OS back then, with good usability, was BeOS. Both MacOS and Windows had stability problems (although NT/2000 were much better, but lacked app/game compatibility), and Linux was a nightmare to update and run (lots of compiling too). So the OS of choice back then for me, was BeOS. I could do everything I needed with it too.
If you wanted to run Unix, your main choices were workstations (Sun, Silicon Graphics, Apollo, IBM RS/6000), or servers (DEC, IBM) They all ran different flavors of BSD or System-V unix and weren’t compatible with each other. Third-party software packages had to be ported and compiled for each one.
On x86 machines, you mainly had commercial SCO, Xenix, and Novell’s UnixWare. Their main advantage was that they ran on slightly cheaper hardware (< $10K, instead of $30-50K), but they only worked on very specifically configured hardware.
Then along came Minix, which showed a clean non-AT&T version of Unix was doable. It was 16-bit, though, and mainly ended up as a learning tool. But it really goosed the idea of an open-source OS not beholden to System V. AT&T had sued BSD which scared off a lot of startup adoption and limited Unix to those with deep pockets. Once AT&T lost the case, things opened up.
Shortly after that Linux came out. It ran on 32-bit 386es, was a clean-room build, and fully open source, so AT&T couldn’t lay claim to it. FSF was also working on their own open-source version of unix called GNU Hurd, but Linux caught fire and that was that.
The thing about running on PCs was that there were so many variations on hardware (disk controllers, display cards, sound cards, networking boards, even serial interfaces).
Windows was trying to corral all this crazy variety into a uniform driver interface, but you still needed a custom driver, delivered on a floppy, that you had to install after mounting the board. And if the driver didn’t match your DOS or Windows OS version, tough luck.
Along came Linux, eventually having a way to support pluggable device drivers. I remember having to rebuild the OS from scratch with every little change. Eventually, a lot of settings moved into config files instead of #defines (which would require a rebuild). And once there was dynamic library loading, you didn’t even have to reboot to update drivers.
The number of people who would write and post up device drivers just exploded, so you could put together a decent machine with cheaper, commodity components. Some enlightened hardware vendors started releasing with both Windows and Linux drivers (I had friends who made a good living writing those Linux drivers).
Later, with Apache web server and databases like MySql and Postgres, Linux started getting adopted in data centers. But on the desktop, it was mostly for people comfortable in terminal. X was ported, but it wasn’t until RedHat came around that I remember doing much with UIs. And those looked pretty janky compared to what you saw on NeXTStep or SGI.
Eventually, people got Linux working on brand name hardware like Dell and HPs, so you didn’t have to learn how to assemble PCs from scratch. But Microsoft tied these vendors so if you bought their hardware, you also had to pay for a copy of Windows, even if you didn’t want to run it. It took a government case against Microsoft before hardware makers were allowed to offer systems with Linux preloaded and without the Windows tax. That’s when things really took off.
It’s been amazing watching things grow, and software like LibreOffice, Wayland, and SNAP help move things into the mainstream. If it wasn’t for Linux virtualization, we wouldn’t have cloud computing. And now, with Steam Deck, you have a new generation of people learning about Linux.
PS, this is all from memory. If I got any of it wrong, hopefully somebody will correct it.
That’s great bit of history
It may be useful for people reading if you could add headers about when each decade starts, since you have many of them there
Stuff needed tweaking more wine worked almost never even for basically window’s programs. Configuring Xfree86 was black magic. Running Startx at the terminal prompt was like rolling the dice. Distro choice was smaller and it was really a choice. Since the child distros were less of a thing. You had Debian , Redhat, Slackware, and SUSE. All were very different at a fundamental level with packaging and philosophy. Also it was way more common to buy boxed copies of Linux distros with big thick manuals that helped you get it installed and take your first steps with Linux. It reminded me of when I first got my TI 83 calculator an it had that massive manual with it.
Also Lugs and spending a lot of time on IRC getting and helping people on freenode (don’t go there now) was a must.
it was garbage.
servers already worked well for the time, but desktop was rough.
You had to really want it, but once you got it, it was more than worth it.
you also had to have precisely the right hardware.
I tried to use Slackware in the mid 90s. After two solid days trying to get my ppp dialup connection to come up after a reboot - it would come up in the first boot after install, run for days like that, but after any reboot it was dead and gone and nobody on the internet seemed to know how to deal with it. “Real men” didn’t use dialup, and people on dialup (self included) had no clue.
I declared it “not ready for prime time” due to that, and issues with sound drivers, and ignored it until 2003. In 2003 I tried some Cygwin and was impressed with its performance, so very close to “bare metal” Ubuntu. In 2004-5 the 64 bit AMD chips were coming out and I used Gentoo to build a true 64 bit system addressing 8GB of RAM - there wasn’t really any other option.
I got tired of compiling every little part of the system from source for days on end and migrated to various flavors of Ubuntu / Debian, which by 2006-7 was becoming a viable desktop alternative. Before that you ALWAYS had to have a Windows machine for something, usually several somethings. At this point I only use my company issued Windows laptop when I need to connect to the company VPN, which can be months between needs depending on what I’m doing. My wife and I use Ubuntu full time now.
Honestly, it sucked. Like most computing at the time. Everything came on a ton of floppy disks, it was impossible to update online unless you had a good connection (which nobody did), and you had to do everything by hand, including compiling a lot of stuff which took forever. I mean, I’m glad I got the experience, but I would never wanna go back to that. It sucked.
Remember when packages like RPM were first introduced, and it was like, “cool, I don’t have to compile everything!” Then you were introduced to Red Hat’s version of DLL-Hell when the RPM couldn’t find some obsure library! Before YUM, rpmfind.net was sooo useful!
I still use pkgs.org pretty frequently when I need to find versions of packages and their dependencies across different distros and versions of distros. I had to use that to sneakernet something to fix a system just this past week.
Shit like that was the last straw for me and I ended up bailing on Linux for, like, 10 years until I got back into it around 2006.
Poor Annie.
Remember the slow internet had to wait overnight for 40 megabyte game and finally finding out it didn’t work.
Up all night, and all you got to see was a boob
Remember the Internet at these speeds, Moss? Up all night and you’d see three women.
Half of it because random disconnect happened in the middle and download did not resume.
In glorious 256 colors !
jad
Nope, i had 10Mb fiber in 1995.
I started using Linux right in the late 90’s. The small things I recall that might be amusing.
- The installation process was easier than installing Arch (before Arch got an installer)
- I don’t recall doing any regular updates after things were working except for when a new major release came out.
- You needed to buy a modem to get online since none of the “winmodems” ever worked.
- Dependency hell was real. When you were trying to install an RPM from Fresh Meat and then it would fail with all the missing libraries.
- GNOME and KDE felt sincerely bloated. They seemed to always run painfully slow on modern computers. Moving a lot of people to Window Managers.
- it was hard to have a good web browser. Before Firefox came out you struggled along with Netscape. I recall having to use a statically compiled ancient (even for the time) version of Netscape as that was the only thing available at the time for OpenBSD.
- Configuring XFree86 (pre-cursor to X.org) was excruciating. I think I still have an old book that cautioned if you configured your refresh rates and monitor settings incorrectly your monitor could catch on fire.
- As a follow on to the last statement. I once went about 6 months without any sort of GUI because I couldn’t get X working correctly.
- Before PulseAudio you’d have to go into every application that used sound and pick from a giant drop down list of your current sound card drivers (ALSA and OSS) combined with whatever mixer you were using. You’d hope the combo you were using was supported.
- Everyone cheered when you no longer had to fight to get flash working to get a decent web browsing experience.
I don’t recall doing any regular updates
You needed to buy a modem to get online
If you stay offline, you don’t need upgrading to prevent virus or hacking. That’s the norm in the good old days.
<I think I still have an old book that cautioned if you configured your refresh rates and monitor settings incorrectly your monitor could catch on fire.> Are you telling me that one dev for X.org could set someone’s monitor on fire by fucking with four lines of code?
Jesus Christ, thanks for that, I didn’t need to sleep tonight.
Monitors don’t work like that anymore. The ones that could catch on fire are pretty much all in the landfills by now.
It was S.u.S.E. Linux 5.3
Great manual.
I was lucky that my NIC, graphics and sound card were supported out of the box.
But everything was still much worse than on Windows.
But I could taste the freedom.
Now all my devices run on Linux (except my Nintendo Switch).
Yeah those manuals were great i still have mine.