3/29/2011

Initiatives of Autonomic Computing - Autonomic computing today



By reading this paper you can gain the knowledge about Autonomic Computing. I detailed it under the following sub topics.

 1.0 Introduction
 2.0 Why need Autonomic?
 3.0 Major Researches 
           3.1 Research projects in Autonomic computing.
           3.2 University research projects in autonomic computing
4.0 Four basic elements of autonomic computing
           4.1self-configuring
           4.2self-healing
           4.3self-optimizing
           4.4 self-protecting
           AutonomicComputing Vs Current Computing
5.0 Autonomic computing architecture
6.0. Autonomic computing today
           6.1. Initiatives of autonomic computing 
           6.2. Benefits
           6.3. Applications
7.0 What happen if does not have autonomic in the future?
8.0 Autonomic Computing Research Issues and Challenges
9.0 Future direction of Autonomic computing
10.0 Conclusion

6.0. Autonomic Computing today 

It is possible to find some aspects of autonomic computing already in today’s software products .This section describes the current initiatives (section 6.1.), benefits (section 6.2.), applications which are using autonomic concept (section 5.3.).

6.1. Initiatives of Autonomic Computing

In today’s various computer vendors and research institutions are involved in autonomic computing. The technology is not only applicable to servers, but also extends to databases, software applications, and Grid Computing networks.

Perhaps the first elements of autonomous computing were ‘software agents’ that made waves around 1999. A prime example is Computer Associates’ Neugents. According to Computer Associates, Neugents look for patterns in data and can extrapolate from the patterns to predict future events[8].

‘Spiders’ or software agents from search engines are another example. Also called ‘Bots’, these agents scour the Web looking for new websites and then return to the search engine and update its database with the new URLs[8].

Windows XP also incorporates self-management technology. When an application crashes, the user can shut it down systematically, thereby preventing the entire system from freezing or hanging. Further, Windows XP looks out for updates and automatically downloads these when available.

Recent versions of Microsoft Office include a Repair feature. So if key program file (such as Winword.exe) gets corrupted or accidentally deleted, the software can reinstall it. Such features will soon be present in other desktop software.
Figure 6.1- plug and play devices
Plug-and-play is another element of autonomous computing. Plug in a new device to your PC and the system will automatically detect it. The operating system will then fire up its hardware wizard, which guides you through the process of installing the appropriate drivers for the new device.

One company that is actively working towards fully autonomous systems is IBM, which has an initiative named Project eLiza .IBM has incorporated some elements of eLiza in its servers. The company is demonstrating software called Enterprise Workload Manager, which monitors groups of servers, managing the machines and moving work between them without the aid of human administrators [8].

An initiative similar to eLiza is Project Oceano. It will enable a group of Linux servers to share jobs, and reassign jobs when new servers are added or removed from the cluster.

Compaq is also pursuing autonomic computing. It is offering a suite of tools collectively called Proliant Essentials. The tool with autonomic characteristics is Compaq Insight Manager. This software delivers pre-failure alerts for Compaq ProLiant servers, thereby proactively detecting potential server failures before they result in unplanned system downtime. Another tool in the suite is ActiveUpdate, an advanced Web-based application that provides proactive notification and automatic download of software updates for all Compaq systems that range from handhelds to servers.

Another example of an autonomic system is the Adaptive data flow engine, a technology used to scan ‘Deep Web’ databases and collect information. Deep Web refers to information on the Internet that cannot be found using traditional search engines by using Telegraph software. Telegraph does much more than traditional search engines. It fetches data from Web-accessible databases, analyses it, does cross-referencing, collates data and presents it all on one screen.

Research in autonomic computing is also taking place in labs at MIT, University of Texas, University of Michigan and other universities.

No comments: