Low-Bandwidth Communication Tools for Science
Dissemination and management of knowledge is essential for scientific enterprise and sustainable development. For several decades, the Abdus Salam International Centre for Theoretical Physics (ICTP) in Trieste, Italy, has paid special attention to the needs of developing countries to foster, through training and research, the progress of science.
The Centre long ago realized the importance of information retrieval systems on the Internet, including the distribution of in-house preprints, yearly activities and public access catalogs.
On a technical level, Linux provides us with a cost-effective alternative for promoting distance electronic collaboration (see Resources). Based on the Linux OS, virtual laboratories and the extensive use of digital communication tools can help reduce scientific isolation, while filling the need to transfer knowledge to developing countries in the Southern Hemisphere in an unprecedented way (see Resources).
Following these principles, we have started building prototype, on-line scientific tools to further enhance electronic collaboration and to support the use of web navigation and database search by e-mail. Below, we describe two tools that Salam ICTP offers the low-bandwidth scientific community. Both packages use state-of-the-art technologies and software developed in-house, and are distributed under the GNU General Public License (GPL).
The ICTP www4mail software allows navigation and search of the entire Internet via e-mail, using any standard web browser and a MIME (Multipurpose Internet Mail Exchange)-aware e-mail program. At first glance, it may appear similar to one of the several web-to-mail software interfaces; but the www4mail program introduces new features not previously available. In short, e-mail messages containing filtered HTML pages are automatically passed to the www4mail server when links to other web sites are selected while browsing.
Written in modular Perl, the program allows retrieval of web pages, searching of arbitrary databases, filling out of web forms (GET and POST conduct web database searches) and following of links (on-line browsing), all by e-mail. It is multi-lingual, easy to manage and supports current Internet standards (MIME, HTML 4.0, etc.).
Developed from scratch on the Linux platform, www4mail has been used successfully on the BSD platform and contains some optional optimisations that are Linux-specific. For example, www4mail can monitor the system load average, directly from the Linux /proc file system and, at high load averages, queue requests for later processing.
Here are some major features of www4mail:
sends replies as e-mail attachments or in the body of an e-mail message, depending on the type of request options sent by the e-mail client through the web browser
supports scripting, once the browser can display it
handles dynamic contents, parsing text HTML and source HTML
preserves the original layout of requested web pages
retrieves information from FTP sites and Usenet news servers
handles meta tags; that is, if a web page is redirected or relocated by the use of a meta statement, www4mail automatically warns about the possible relocation of the information and provides suitable links for the new location at the top of the reply page
handles frames, inserting suitable links to each framed document
supports user authentication for password-protected web/FTP sites
traps error messages and sends them back to the user
provides support for text-only access for compatibility with the alternative “Agora” and “GetWeb” web-mail servers
serves filtered requests to reduce bandwidth
supports the transfer of binary data
allows web pages to be downloaded as PostScript files, to be viewed or printed locally (see for manuals)
4.23.00 - It was in the C-edit directory as of May 1.
www4mail (see logo in Figure 1) was developed mainly to help researchers from developing countries browse the Web using only e-mail and slow Internet links. While the amount of information on the Web has grown exponentially in the last few years, there is still a large community of Internet users who have access to only e-mail, or their Internet providers do not offer full Internet connections (some of them still use UUCP) or who cannof afford to have an expensive account with full Internet capabilities. Many of these users live in rural areas of developing countries, and rely on e-mail to access essential medical and business information as well as for interpersonal communication and world news. Having the ability to query available databases (such as AltaVista, HotBot, etc.) or preprint repositories with one simple e-mail and receive the output in a few minutes (or hours) could help them tremendously with their scientific work.
At present, www4mail can be tested by sending an e-mail message to email@example.com, or to any other place where the gateway is installed (e.g., Bellanet-Canada, www.bellanet.org/email.htm), listing the requested URL(s) in the body of the message.
Over 50 server configuration options are currently available for setting parameters such as maximum quota per user, gateway administrators, maximum size of each request, or to split sizes for large files. (Type help in the body of the e-mail message for further details).
The installation procedure of the server is simple. For example, under Red Hat Linux, create a user account called www4mail (adduser www4mail), log on as user www4mail (su - www4mail), extract the tar archive in the home directory for www4mail (tar zxvf www4mail.tar.gz) and perform a few extra operations (e.g., to enable forwarding). It is necessary to create a link from the executable /home/www4mail/bin/www4mail to the /etc/smrsh subdirectory in order to keep the sendmail MTA (mail transfer agent) happy. To optimize its configuration, some preliminary monitoring is necessary.
www4mail has been very useful for many people from many different countries, often receiving over 12,000 requests per day. You can view weekly statistics at http://web.bellanet.org/www4mail/).
|Designing Electronics with Linux||May 22, 2013|
|Dynamic DNS—an Object Lesson in Problem Solving||May 21, 2013|
|Using Salt Stack and Vagrant for Drupal Development||May 20, 2013|
|Making Linux and Android Get Along (It's Not as Hard as It Sounds)||May 16, 2013|
|Drupal Is a Framework: Why Everyone Needs to Understand This||May 15, 2013|
|Home, My Backup Data Center||May 13, 2013|
- I once had a better way I
3 hours 40 min ago
- Not only you I too assumed
3 hours 58 min ago
- another very interesting
5 hours 51 min ago
- Reply to comment | Linux Journal
7 hours 44 min ago
- Reply to comment | Linux Journal
14 hours 38 min ago
- Reply to comment | Linux Journal
14 hours 54 min ago
- Favorite (and easily brute-forced) pw's
16 hours 46 min ago
- Have you tried Boxen? It's a
22 hours 38 min ago
- seo services in india
1 day 3 hours ago
- For KDE install kio-mtp
1 day 3 hours ago
Enter to Win an Adafruit Pi Cobbler Breakout Kit for Raspberry Pi
It's Raspberry Pi month at Linux Journal. Each week in May, Adafruit will be giving away a Pi-related prize to a lucky, randomly drawn LJ reader. Winners will be announced weekly.
Fill out the fields below to enter to win this week's prize-- a Pi Cobbler Breakout Kit for Raspberry Pi.
Congratulations to our winners so far:
- 5-8-13, Pi Starter Pack: Jack Davis
- 5-15-13, Pi Model B 512MB RAM: Patrick Dunn
- 5-21-13, Prototyping Pi Plate Kit: Philip Kirby
- Next winner announced on 5-27-13!
Free Webinar: Hadoop
How to Build an Optimal Hadoop Cluster to Store and Maintain Unlimited Amounts of Data Using Microservers
Realizing the promise of Apache® Hadoop® requires the effective deployment of compute, memory, storage and networking to achieve optimal results. With its flexibility and multitude of options, it is easy to over or under provision the server infrastructure, resulting in poor performance and high TCO. Join us for an in depth, technical discussion with industry experts from leading Hadoop and server companies who will provide insights into the key considerations for designing and deploying an optimal Hadoop cluster.
Some of key questions to be discussed are:
- What is the “typical” Hadoop cluster and what should be installed on the different machine types?
- Why should you consider the typical workload patterns when making your hardware decisions?
- Are all microservers created equal for Hadoop deployments?
- How do I plan for expansion if I require more compute, memory, storage or networking?