Back in Linux Journal issue #2 handy, the title of my “From the Editor” column was “Linux 2000...” The abstract goes on to say:
I see Linux as a progressive movement as well as an operating system. And, with any movement you need to chart your direction. To help with that charting I decided that rather than write the April, 1994 editorial I would just go ahead and write the one for January, 2000.
“Well, here we are and it's time to see how well I did. If you don't happen to have issue #2, you can check out the original article here.”
My first idea, that Linux would turn into a movement to provide affordable, reliable, multi-tasking software, was certainly right on. Looks like my growth prediction was a bit off, though. I estimated Linux would be in 100 million homes—seems a bit high.
I also implied that few would remember an old program loader called MS-DOS. Well, I think that has essentially happened. While it continues to live under Windows, not many people seem to realize it.
Another miss on my part was personal satellite stations in the late 1990s. They exist—you can buy a system that offers a bidirectional connection to the Internet—but the price of that connection is still too high to be practical for personal computing. What has happened is that the cost of wired Internet connectivity, most specifically DSL and cable modem, has offered significant bandwidth to many at a very low cost.
I had some predictions about Linux Journal. Specifically, I projected that 90% of our subscribers would receive their magazines via the Internet rather than on paper. This could have happened, if not for two reasons: most readers want the print version of the magazine, and our advertisers want you to see their ads. We do offer an electronic version of the articles to our subscribers at http://interactive.linuxjournal.com/, but we continue to print over 100,000 magazines each month.
I predicted two events that would make Linux a huge success. The first was the formation of MoAmI Semiconductor, a company that produced a chip where Linux was the first OS to run native. We're still waiting on this one. Linux will most likely be the first OS on Intel's Merced chip, with VA the first vendor to ship systems. And, well, keep watching Transmeta.
The second event was Linux becoming the most popular OS in computer science classes. I don't have the numbers, but it sure seems like that could have happened.
I also predicted that NT would become a niche operating system. Did it? We don't know yet, as its name is now Windows 2000 and, well, it really hasn't shipped yet. I still feel that most corporate customers waiting for Windows 2000 will, once they see it, decide Linux is the right answer for servers.
I also predicted those old Linux developers would still be writing code or books on Linux, rather than being CEOs of multi-billion dollar corporations. Generally, this seems to be true, with some also writing articles for Linux Journal. Why? Because these people are doing what they want.
I'm fairly happy with how things have turned out. I took some rather SWAGs (Scientific Wild Guesses) and wasn't totally out of line. The good news is Linux is still here, you are still here, and we are still here. Hopefully, the only big arguments of the next five years will be about programming in Perl or Python and whether to use GNOME or KDE on your Linux system.
Practical Task Scheduling Deployment
July 20, 2016 12:00 pm CDT
One of the best things about the UNIX environment (aside from being stable and efficient) is the vast array of software tools available to help you do your job. Traditionally, a UNIX tool does only one thing, but does that one thing very well. For example, grep is very easy to use and can search vast amounts of data quickly. The find tool can find a particular file or files based on all kinds of criteria. It's pretty easy to string these tools together to build even more powerful tools, such as a tool that finds all of the .log files in the /home directory and searches each one for a particular entry. This erector-set mentality allows UNIX system administrators to seem to always have the right tool for the job.
Cron traditionally has been considered another such a tool for job scheduling, but is it enough? This webinar considers that very question. The first part builds on a previous Geek Guide, Beyond Cron, and briefly describes how to know when it might be time to consider upgrading your job scheduling infrastructure. The second part presents an actual planning and implementation framework.
Join Linux Journal's Mike Diehl and Pat Cameron of Help Systems.
Free to Linux Journal readers.Register Now!
- Stunnel Security for Oracle
- Murat Yener and Onur Dundar's Expert Android Studio (Wrox)
- SourceClear Open
- SUSE LLC's SUSE Manager
- My +1 Sword of Productivity
- Managing Linux Using Puppet
- Google's SwiftShader Released
- Non-Linux FOSS: Caffeine!
- Parsing an RSS News Feed with a Bash Script
- SuperTuxKart 0.9.2 Released
With all the industry talk about the benefits of Linux on Power and all the performance advantages offered by its open architecture, you may be considering a move in that direction. If you are thinking about analytics, big data and cloud computing, you would be right to evaluate Power. The idea of using commodity x86 hardware and replacing it every three years is an outdated cost model. It doesn’t consider the total cost of ownership, and it doesn’t consider the advantage of real processing power, high-availability and multithreading like a demon.
This ebook takes a look at some of the practical applications of the Linux on Power platform and ways you might bring all the performance power of this open architecture to bear for your organization. There are no smoke and mirrors here—just hard, cold, empirical evidence provided by independent sources. I also consider some innovative ways Linux on Power will be used in the future.Get the Guide