Mano a Mano with Microsoft: Update
A couple of weeks ago, I wrote about an impending meeting with Microsoft to discuss some of its actions during the standardisation process of OOXML at the ISO. I asked Linux Journal readers for some help in preparing for this, and you responded with a generosity entirely in keeping with the spirit of free software. The many helpful comments to that post give some indication of the scale of the response, but that overlooks the extraordinary emails I received from others, packed with useful information, which clearly represented many hours' work. To everyone, I'd like to express my thanks. The bad news is that the meeting is not going to take place after all.
That's not because either I or Microsoft chickened out, but a consequence of the fact that the CIO of Newham, Richard Steel, for whose benefit the meeting was being held, has rather suddenly and unexpectedly announced that he is retiring. In the wake of this news, I contacted Microsoft, and we agreed that there was little point holding the meeting as originally planned.
However, I was very conscious of the huge amount of effort that people had put into this matter, and so I wondered how that might be salvaged. Since the issue of how Microsoft behaved during the ISO process is still a very live one, it seemed to me a worthwhile exercise trying to pull together a summary of certain aspects of what happened last year. Since the whole point of doing this was to hear Microsoft's response, I asked the company whether they would be interested in writing a reply to my post, and they have agreed to do so. Quite what this will produce, I don't know, but I think it's worth trying. At the very least, there will be some more documentation of what happened, and maybe some new points from Microsoft too.
To clarify: what I aim to concentrate on is not the technical side – what was and still is wrong with OOXML as an ISO standard – so much as the procedural issues at the national level. Sifting through the wonderful resources provided by Linux Journal readers, I aim to pull out what seem to me the clearest cases of Microsoft bending the system to produce the result that it wanted.
I will omit the many vague accusations that were flying around at the time, simply because it wouldn't be fruitful to include them. Microsoft could rightly say that they are merely unsubstantiated rumours. Even though this will mean leaving out some of the more egregious flouting of rules, I think it will make the case stronger by virtue of concentrating on those incidents that are most substantiated and least easy to dismiss.
I'm currently putting that document together; I will then post it here and send the link to Microsoft for comments, which can be in whatever form they wish. I then aim to publish the comments, unedited, alongside my original post in some way, depending on what format Microsoft chooses. It will be interesting to see what, if anything, all this will produce.
Practical Task Scheduling Deployment
July 20, 2016 12:00 pm CDT
One of the best things about the UNIX environment (aside from being stable and efficient) is the vast array of software tools available to help you do your job. Traditionally, a UNIX tool does only one thing, but does that one thing very well. For example, grep is very easy to use and can search vast amounts of data quickly. The find tool can find a particular file or files based on all kinds of criteria. It's pretty easy to string these tools together to build even more powerful tools, such as a tool that finds all of the .log files in the /home directory and searches each one for a particular entry. This erector-set mentality allows UNIX system administrators to seem to always have the right tool for the job.
Cron traditionally has been considered another such a tool for job scheduling, but is it enough? This webinar considers that very question. The first part builds on a previous Geek Guide, Beyond Cron, and briefly describes how to know when it might be time to consider upgrading your job scheduling infrastructure. The second part presents an actual planning and implementation framework.
Join Linux Journal's Mike Diehl and Pat Cameron of Help Systems.
Free to Linux Journal readers.Register Now!
- Paranoid Penguin - Building a Secure Squid Web Proxy, Part IV
- SUSE LLC's SUSE Manager
- Google's SwiftShader Released
- Murat Yener and Onur Dundar's Expert Android Studio (Wrox)
- Managing Linux Using Puppet
- My +1 Sword of Productivity
- Non-Linux FOSS: Caffeine!
- SuperTuxKart 0.9.2 Released
- Parsing an RSS News Feed with a Bash Script
- Doing for User Space What We Did for Kernel Space
With all the industry talk about the benefits of Linux on Power and all the performance advantages offered by its open architecture, you may be considering a move in that direction. If you are thinking about analytics, big data and cloud computing, you would be right to evaluate Power. The idea of using commodity x86 hardware and replacing it every three years is an outdated cost model. It doesn’t consider the total cost of ownership, and it doesn’t consider the advantage of real processing power, high-availability and multithreading like a demon.
This ebook takes a look at some of the practical applications of the Linux on Power platform and ways you might bring all the performance power of this open architecture to bear for your organization. There are no smoke and mirrors here—just hard, cold, empirical evidence provided by independent sources. I also consider some innovative ways Linux on Power will be used in the future.Get the Guide