Work the Shell - More Fun with Days and Dates
I received a very interesting note from a reader—a note that referred to a very interesting problem:
Many UNIX commands (for example, last) and log files show brain-dead date strings, such as “Thu Feb 24”. Does anybody out there have a script that will convert that to a year, given a five-year interval and defaulting to the present?
Given a day of the week, a month and a day, is it possible to calculate quickly the most recent year in the past when that particular date occurred on that day of the week? Of course it is!
Various formulas exist for calculating this sort of thing, but I realized pretty quickly that the handy cal utility can do the work for us. If you haven't experimented with it, you'll be surprised at what it can do. Here are two quick, relevant examples:
$ cal March 2011 Su Mo Tu We Th Fr Sa 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 $ cal mar 2007 March 2007 Su Mo Tu We Th Fr Sa 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31
Any light bulb starting to glow for you? If you know the month and day, you simply can go backward looking at that month's day-of-week layout until finally you find a match.
In a rudimentary fashion, the basic idea can be illustrated with a loop, like this:
repeat cal $month $year | grep $day if day-of-week matches echo date $month $day most recently occurred in $year else year=$(( $year - 1 )) end repeat
Of course, the problem is a bit more complicated (as they always are), partially because of the complexity of calculating what day a specific date occurs in the cal output. There's another complication too, however; the requested date actually might have occurred in the current year, so it's not as simple as starting with the year 2010 and going backward.
The first task is to figure out how to get the information from the user. We'll have only three input parameters and do relatively little testing for misspelled day names and so on:
if [ $# -ne 3 ] ; then echo "Usage: $(basename $0) weekday month day" echo " (example: $(basename $0) wed aug 3 )" exit 1 fi
That's straightforward and pretty typical, offering a nice usage tip if you forget how to use the script. As is typical of scripts, we return a nonzero result upon error too.
We can't work with completely arbitrary data, however, so when we grab the first few parameters, we'll transliterate them into lowercase and chop off all but the first three letters:
weekday=$(echo $1 | tr '[[:upper:]]' '[[:lower:]]'; | cut -c1-3) month=$(echo $2 | tr '[[:upper:]]' '[[:lower:]]'; | cut -c1-3) day=$3
Given “Monday, February 8”, it'd be converted automatically to “mon” and “feb” for subsequent testing.
We also need the current date fields for testing, and to do this, I'll demonstrate a very neat trick of date that makes this incredibly efficient:
eval $(date "+thismonth=%m; thisday=%d; thisyear=%Y")
The eval function interprets its argument as if it were a direct script command. More interesting, date can output arbitrary formats (as documented in strftime if you want to read the man page) by using the + prefix, with %m the month number, %d the day of the month and %Y the year. The result is that date creates the string:
thismonth=03; thisday=01; thisyear=2011
which then is interpreted by the shell to create and instantiate the three named variables. Neat, eh?
It turns out that users can specify a month by name or number on the command line. If it's already a number, it'll survive the transforms intact. If it's a name though, we also need the number, so we can figure out whether the date specified could be earlier this year. There are several ways to do this, including a case statement, but that's a lot of work. Instead, I'll lean on sed as I quite frequently do:
monthnum=$(echo $month | sed 's/jan/1/;s/feb/2/;s/mar/3/;s/apr/4/;s/may/5/;s/jun/ ↪6/;s/jul/7/;s/aug/8/;s/sep/9/;s/oct/10/;s/ ↪nov/11/;s/dec/12/')
Here's where a misspelled month name is a problem, but that's a challenge beyond the scope of this script. For now, however, we'll just roll with it.
Dave Taylor has been hacking shell scripts for over thirty years. Really. He's the author of the popular "Wicked Cool Shell Scripts" and can be found on Twitter as @DaveTaylor and more generally at www.DaveTaylorOnline.com.
|Non-Linux FOSS: libnotify, OS X Style||Jun 18, 2013|
|Containers—Not Virtual Machines—Are the Future Cloud||Jun 17, 2013|
|Lock-Free Multi-Producer Multi-Consumer Queue on Ring Buffer||Jun 12, 2013|
|Weechat, Irssi's Little Brother||Jun 11, 2013|
|One Tail Just Isn't Enough||Jun 07, 2013|
|Introduction to MapReduce with Hadoop on Linux||Jun 05, 2013|
- Containers—Not Virtual Machines—Are the Future Cloud
- Non-Linux FOSS: libnotify, OS X Style
- Lock-Free Multi-Producer Multi-Consumer Queue on Ring Buffer
- Linux Systems Administrator
- Introduction to MapReduce with Hadoop on Linux
- RSS Feeds
- Weechat, Irssi's Little Brother
- Tech Tip: Really Simple HTTP Server with Python
- Validate an E-Mail Address with PHP, the Right Way
- Android's Limits
- Reply to comment | Linux Journal
18 min 56 sec ago
- Welcome to 1998
1 hour 7 min ago
- notifier shortcomings
1 hour 31 min ago
3 hours 7 min ago
- Android User
3 hours 9 min ago
- Reply to comment | Linux Journal
5 hours 2 min ago
7 hours 52 min ago
- This is a good post. This
13 hours 5 min ago
- Great, This is really amazing
13 hours 7 min ago
- These posts are really good
13 hours 8 min ago
Free Webinar: Hadoop
How to Build an Optimal Hadoop Cluster to Store and Maintain Unlimited Amounts of Data Using Microservers
Realizing the promise of Apache® Hadoop® requires the effective deployment of compute, memory, storage and networking to achieve optimal results. With its flexibility and multitude of options, it is easy to over or under provision the server infrastructure, resulting in poor performance and high TCO. Join us for an in depth, technical discussion with industry experts from leading Hadoop and server companies who will provide insights into the key considerations for designing and deploying an optimal Hadoop cluster.
Some of key questions to be discussed are:
- What is the “typical” Hadoop cluster and what should be installed on the different machine types?
- Why should you consider the typical workload patterns when making your hardware decisions?
- Are all microservers created equal for Hadoop deployments?
- How do I plan for expansion if I require more compute, memory, storage or networking?