56444741 case-study-of-os

36
Assignments On: Case study on os

Upload: homeworkping2

Post on 11-Feb-2017

209 views

Category:

Education


0 download

TRANSCRIPT

Assignments On:

Case study on os

CASE STUDY OF OPERATING SYSTEM

1. MS DOS2. MICROSOFT WINDOWS A. XP B.WIN VISTA C.WIN 73. UNIX 4. APPLE MACINTOSH OS A. MAC OS (1984-2001) B. MAC OS X

CASE STUDY OF OPERATING SYSTEM

Case study 1 – MS-DOS

Nowadays, it’s impossible to introduce desktop computers without seeing the name of Microsoft around sooner or later. This company totally managed to take over this world. If they tell your computer to commit suicide, unless you’re some kind of bearded geek not using their software, it will.

The story of Microsoft begins with the emergence of personal computers, and with two clever guys called Bill Gates and Paul Allen who loved to mess around with computers. They were popular at the time in the computing world, because they wrote useful software (things that made writing other software much easier, a spreadsheet, a world processor…) on a low-cost computer called the Altair 8800 (that didn’t even included a keyboard and a monitor). When people at IBM – who ruled a big part of the computer world at the time – started having an interest in micro-computers, building what was going to be a big hit called the IBM PC, they were attracted by this popularity and licensed some of Gates’s software. And when they failed to get the major operating system of those days, CP/M, they went back to Gates and asked him if he could provide them for an OS for it. Gates found one, bought it, and hired the one who wrote it in order to modify it so that it fits IBM’s need. MS-DOS was born.

The first releases of MS-DOS essentially provided the file abstraction (ie the ability to consider that disk space is divided into chunks of data called files, allowing one to ignore the actual structure of the floppy disk he works on), some primitive tools to manipulate internal memory and text files (ie files containing text and numbers), and basic configuration routines, everything being stolen from the dominant OS of those days, CP/M. One used it by typing commands on a keyboard, pressing Return, and reading a resulting dump of text on the screen, an interface that’s called command line. However, it introduced a new commercial idea that was here to last : the idea of selling an operating system bundled with a computer, betting that users won’t bother using another one if there’s already something working there.

The command line interface

Subsequent releases included support for higher-capacity data storage, directories (a kind of box which may contain files and other directories, and hence a hierarchical way to organize files), then for more languages, then for clocks (a chip whose goal is to measure time, useful not only for calendar/watches-like applications but also for real-time applications) and even higher data storage, then for computer

networks. After that, DOS 4, 10 times heavier than the first release of DOS with 110 KB memory usage (which is 1/30 the size of a common MP3 file), introduced “dosshell”, a new optional way to explore files that was a little less linear than the command line, direct ancestor of windows’s file explorer.

The Shell interface

Subsequent releases of MS-DOS, besides adding up support for a higher storage space (again…) through multiple hard drives management and a new file management system, and managing more RAM, introduced three new tools : one to compress files in order to economize disk space (at the cost of slower data access), one to check the hard drive for errors (due to bad machine shutdown as an example) and try to fix them, and a simple antivirus, MSAV.

As one may figure out, at this stage, DOS didn’t really evolved anymore. Those last things really were secondary, questionable features, that were added just to sell new releases of the operating system and make money from it, and maintenance improvements for keeping in touch with new hardware. MS-DOS had reached a mature stage of evolution, and was a bit left behind while Microsoft were now working on their new product, Windows, initially running on top of DOS, which we’re going to describe in a later article.

Now that we have described the features and evolution of MS-DOS, we can discuss them. A few points shine, especially :

Most of the development effort on the hardware abstraction side was spent on file and storage space management : Some releases of DOS took months, years of development, just to add support for newer, higher-capacity diskettes. Improvement on OS support of other kind of hardware was very limited.This may be partly explained by the fact that, at the time, everyone was okay with letting programs deal with the bare hardware. Hardware manufacturers made the thing a lot easier by letting everyone know how to interact with their hardware, and by introducing hardware that was dead easy to manipulate for someone used to assembly language. Standards were the rule rather than the exception, so if you could make your program work on a computer, you were 95% sure that it would work on other hardware without modifying it in any way. There wasn’t that much viral software at the time to harm the machine given direct access to it, especially due to the fact that people didn’t downloaded and run random software from the internet, for the internet didn’t exist for most people. On the other hand,

performance was a critical issues, and there’s no program faster on a specific hardware that one written specifically for it, using it directly.Part of that wasn’t true for storage hardware. First, there wasn’t any kind of dominating storage medium, there was a jungle of incompatible technologies (multiple kinds of tapes, various flavors of diskettes, the first hard drive disks…), and they shared the common characteristic of being awful to manipulate. Then, performance wasn’t such an issue for file storage : if you store something, it’s not for using it right away, and you don’t spend your time reading and storing files on a diskette in your programs when you’ve got a main memory that’s a lotfaster. Last, file storage and manipulation was almost the sole thing that an average unskilled user, trying to use software rather than write it, was forced to get into, in order to find and run the software he used, or to copy the text files he wrote to disk, so the process had to be as simple as possible. The Shell, whose only purpose was to better visualize the hierarchical structure of directories and find files quicker, perfectly illustrates that.

Little to no planning, features come as needed : Let’s see… As DOS aged up, 2 different file systems – ways to manage the file abstraction – were used, one after another, called FAT12 and FAT16. FAT16 was introduced to address the maximum disk capacity limit of FAT12. They are extremely close to each other for older programs maximum compatibility reasons, so that differentiating FAT12 and FAT16 is extremely difficult, but at the same time they are structurally incompatible. This is a typical example of a hack, a modification in a program that a developer introduces when he understands that he messed up, but don’t want to make a new design doc and other silly rigorous conception, just want to fix the sole thing that doesn’t work. Perhaps the most common source of bugs in software is when people forget about the hack (or don’t even know about it) and push it to its limits.At the time of DOS, they couldn’t plan things, since they didn’t have a clue of how the computer business would move on, so they introduced gradual changes, leading to hack accumulation, and hence bug multiplications. Let’s anticipate the following articles by saying that this is one of the reasons that led to the abandon of DOS later : managing, correcting, and more generally modifying it had become too complicated, since no one could ever understand how it exactly worked. The operating system needed a complete rewrite.

From the history of DOS, we may extract the following keys of its success and subsequent fall :

Always focus on what’s important first… : Well, reasons for that are pretty obvious

…but don’t neglect the OS’s theoretical foundation and design : Or, sooner or latter, hacks will occur.

Make the doc simple, complete, and easily accessible : And you save people a lot of trouble, preventing effort duplication on the way. Hardware manufacturers knew that, at the time.

Don’t neglect the average user : After all, there’s a lot more users than there’s developers, so you know which one makes lots of sales.

Cleanness is the friend of reliability : It always seems fine at the time, but it is the source of most of computer lack of reliability nowadays.

Compatibility may be the enemy of cleanness if the software is poorly designed to start with : This will be even more obvious after we study the story of Windows.

The search of maximal performance may be an enemy too : There are times when only hack will make a program faster. Though if I were you, I’d better choose reliability and simplicity of design over performance, as long as said performance is sufficient.

With those last three points, one figures out that OS design is often made of dilemmas. This is not the last time I’ll be saying it. Thanks for reading !

Case study 2 – Microsoft WindowsMicrosoft were close to getting rid of MS-DOS and introducing a new product, Windows. We’re now going to describe why they did that, how they did it, and what kind of issues they’ve been facing then and are still facing now. Please note that we’ll focus on Windows releases for the consumer market only.

DEATH OF THE CLI

First, what is windows initially, and why would Microsoft wish to introduce such a product ? Well, from an average user’s point of view, DOS sucked badly by one point : it was using a command-line interface (CLI), which is, I recall you, when one types in a command, hits Return, after what the computer starts some computation and displays the results.

Command-line is a preferred interface for most power users and developers because :

It’s fast : Text manipulation is a relatively easy task from a computing point of view, and the command line syntax is made so that the computer may analyze commands very quickly. On the user side, typing a command is a speedy task because it doesn’t require you to analyze what the computer tells you. The whole command line syntax is engraved in your brain, you just have to remember it, which is pretty quick if you use it everyday. And even if you’ve got thousands of programs, typing in command-line isn’t slowed down the tiniest bit.

It has very low hardware requirements : Modern graphical interfaces have important resource consumption. They have to track the motion of the mouse, constantly re-draw windows as needed, putting flashy, real-time-rendered, animated eye-candy everywhere, keep in memory big uncompressed images corresponding to what’s displayed on the screen, along with some temporary data used for quicker and smoother drawing… By comparison, command line only requires a few KB of memory, few CPU power, and no advanced graphic card.

It’s a perfect tool for repeated tasks : Suppose that there is a single thing that you want to do on twelve computers. You know what commands you have to type in order to obtain this result. Then you can just put those commands in a text file (called a batch file or a shell script depending on the computing

church you belong to), one after the other, and give the text file to your operating system, which is going to automatically read and execute all the commands in it. Think of it : twelve computers to manipulate, a single batch file to carry around and run…

It is streamable : Command-line input and output may be redirected from and to various I/O devices and disk files. The former allows one very low-powered, dumb terminal computer to only read input from and send input to a central server, which was practical when computers were big and expensive because it allowed people to work on tiny, inexpensive machines, that would cost far less to replace if they died because of a misplaced cup of coffee. Though not needed anymore, this way of doing things still is loved by the paranoiac and dictatorial system administrator. Streaming to a file allows to keep a record of the huge dump of text spurred out by some command-line programs in order to keep a track of it and carefully analyze it later. Newer command-line interfaces also allows one to use the output from a program as input for another program directly. Just think of the possibilities : if you want to delete all files containing “foo” in their name, you just use “DIR *foo*” to list such files and directly send the output of it to the “REM” program which deletes files.

It has some features of modern programming languages : Latter interfaces also introduced loops (do a task multiple times) and conditional structures (if certain things happen, do this, otherwise do that), making command-line a (relatively) simple way to make little programs that work on almost any computer provided they use the same operating system.

It’s easily extensible : Any program may, under certain simple conditions, become a new command in a command-line interface, and managing thousands of programs in command-line isn’t nearly as complicated as it can be with other interface paradigms

However, command-line is just awful for normal people because :

Reading the manual is mandatory : You can’t learn how the system works progressively, knowing only a few basic concepts. You have to immediately stick in your memory several basic commands (CD, REM, DIR, TYPE, HELP…) and what they’re used for, because without that you just can’t do anything. Such a necessity is very frustrating for new users, who generally prefer to jump in and learn later as much as possible. It’s also a major source of errors while one gets used to the system, which is even more frustrating.

Error prevention does not exist : Suppose that you’ve been starting to type a big command. But the second word is wrong, because you were sleepy or don’t know anything by heart yet. In a command line system, the computer does not analyses what you told him to do before you press Return. So it’s only after you’ve sent the whole command that the computer says “I’m sorry, Dave. I’m afraid I can’t do that” (or rather “Bad command or filename” in DOS’s poetic words), and then you have to check the entire long command for errors.

Error detection and correction is poor : The basic structure of most command line systems is made so that several frequent and simple errors can’t be properly detected and reported to the user by the system. Hence the user must do error correction all by himself. Simple example : several command

accept an infinite number of parameters, so that you can’t know by the number of parameters given by the user if a mandatory parameter is missing. All the operating system can say is that there is an incorrect parameter somewhere. Even worse, there are (several) cases where the operating system can’t detect the error at all and does nonsense, making the user think that it’s buggy.

There is a huge lack of hierarchy : Even though the average user is only going to use five commands or so, the operating system treats ALL commands as being equal. This means that if you want to type “DIR”, the command for displaying the content of directories, but only remember that it starts with a D, have lost your manual, and want the system to display all commands starting with a D, you may get something like…DATDATPCKDCHECKDDDEFLATEDESTORDIRDRWATSONDUCKDWMDZBRWZ

…or, usually, something much longer, and have to read through the entire list, looking for your command.

So while command-line satisfies perfectly old-school and professional people, it should now be obvious that it’s not okay for average users, that together form, say, 95% of Microsoft’s targeted audience. Clearly, something was wrong in this core concept of MS-DOS, and had to be fixed.

A solution to this problem came from research made by the team of Doug Engelbart (Stanford Research institute), and later by people working in a huge laboratory called Xerox PARC. It’s called the GUI (Graphical User Interface), and more specifically the WIMP (Windows, Icons, Menus, Pointer) paradigm. The user interacts with programs, including the operating system, through the use of a pointing-and-clicking device (generally a mouse, but touchscreen are gradually getting popular nowadays) that allows an on-screen cursor, the pointer, to be moved on the screen. Programs and certain functionalities are launched by clicking on icons, tiny pictures that should more or less indicate what they’re used for, but usually helpfully associated with a tiny text explaining it. Those programs work in well-defined areas of the screen called windows (that the user can usually conveniently move or resize). The icon model isn’t that great when a large set of actions must be provided, so one uses the menu system : a list of action that is shown temporary for interaction purposes an then disappears (just right-click on the present text and you’ll get a perfect example of menu).

WINDOWS XP

Just one year after Windows Me, Windows XP was out. It merged together professional-oriented features and superior core from Windows NT on one side and consumer-oriented features from Windows Me and 9x on the other side, in one product that sold like hotcakes, including :

True application and kernel data isolation : Coming from the NT branch, this neat feature makes sure that applications can’t mess with each other’s data and that they can’t gain direct access to the hardware. This makes Windows XP the first consumer-oriented OS from Microsoft with serious security capabilities.

Multiple users management : Also coming from the NT branch. One may now create several users for the OS. Those users might have more or less limited capabilities as needed, and they have their own private folder, which other users (except administrators) cannot read, in which any user-specific data (wallpaper used, games saves) are stored..

Relative hardware independence : NT-related too. All the hardware specific-code is limited to a subset of the kernel and can easily be changed. This allowed Windows XP to be easily ported to 64-bit processors as they came out, for the sake of improved performance.

Far increased reliability : Windows XP was by far one of the most reliable OSs ever conceived by Microsoft. Complete system crashes (aka Blue Screens of Death or BSODs) that were quite common at the time of Windows 98 and even more with Windows Me became rather the exception than the rule, especially after the introduction of several Service Packs, because applications bugs weren’t capable of doing much harm anymore.

DOS emulation and general good compatibility : Though not perfect on that side, Windows XP made great efforts to ensure application compatibility with Windows Me, Windows 9x, and DOS. Even some Windows 1 applications can still run flawlessly on it.

Mature multimedia support : With Windows XP, multimedia features finally reached a perfectly mature state. As an example, there was out-of-the-box excellent support for scanners and cameras. Movie maker became somewhat less laughable. All the power of Windows 9x DirectX multimedia infrastructure was flawlessly introduced to Windows XP. Folders could display little thumbnails of images in order to help managing photos. One could burn data CDs easily since CD burners became a common thing at the time. And so on…

More networking : The successive releases of Windows XP implemented all newer networking devices and protocols one by one : DSL, Wi-Fi, WEP then WPA encryption protocols…

Various UI tweaks and stupid default theme : The new start menu now displays frequently used applications and various user folders, the file explorer displays “common tasks” in order to interact quickly with files, big icons makes it easy to target important elements, and other delicate attentions, sadly including even more annoying popup windows filled with garbage than ever… Windows XP also introduced ability for the user to change the appearance of windows and controls using “themes”, a very popular feature. Especially looking at the childish and hard on the eyes theme bundled with Windows XP.

Bloated task manager : Windows 95 and later featured the Ctl+Alt+Suppr keystroke, used in order to close hanged programs. It displayed a list of opened programs, along with a button to close one, with extreme simplicity and efficiency. However, starting with Windows XP, things won’t go this way anymore. The Task Manager, as it’s called, will allow one to know about CPU usage, memory usage, networking, users logged in, and so forth. When an application is hanged, taking 100% CPU, this application will take minutes to load. Hence the “improvements” make it quite inefficient at closing buggy application, and one has to wonder what else it had to be used for in Microsoft’s engineers mind.

Definitive failure at making a simple design : Though NT included more interesting feature in its internals than I may describe here (in fact, Microsoft included almost any existing solution for each OS design problem in it), this came with a cost : Windows NT internals are horribly complicated, and have a high footprint on performance and memory usage. Developing for it is a hard task, because there are, say, ten alternative approaches to each problem, with no clue which one is the fastest for a specific purpose..

Bloated kernel : Somewhat related to the preceding issue. Windows’s NT core includes a lot of abstractions, alternative ways to manage the same task, but also running as privileged applications were hardware drivers. This makes a lot of machine code running as maximum-privilege software, reducing the benefit of kernel data isolation by the hardware (since several unauthorized operations can occur using bugs in this huge mass of computer code), and more generally that the system is prone to failure due to buggy software with high security privileges. This led to poor security in Windows XP, albeit still better than that of preceding consumer-oriented releases of Windows.

Slower than ever : Windows XP was quite slow at doing everyday tasks like file management or launching applications, close to Windows 98 in that respect despite running on hardware that was four times faster. Microsoft’s plans to reduce boot time to 30s turned out to take the form of showing the desktop a long time before it’s really usable, with the operating system silently continuing to load in the background.

Product activation : On the early days of computing, one would just buy software, put the diskette or cd-rom in the drive, install it, and use it. However, a new trend dawned by the time of Windows XP : clearly stating that the user wasn’t proprietary anymore of the software he was using. Alongside with several piracy concern, this led to an “activation” process, where the owner of a Windows XP licence had to call microsoft or go on some website, with failure to do so leading to software not functioning after 30 days. This process had to be repeated anytime an important piece of hardware was changed on the PC. This

process allowed Microsoft to make sure that a product key of Windows XP wasn’t used multiple times.(As an aside, it didn’t stop piracy at all, but it made several users angry, like any copy protection measure of this type.)

Bozo the clown-oriented theming and new start menu

Windows XP was extremely well perceived, noticeably because of its great reliability and well-implemented multimedia features. Its slowness and somewhat poor security led to some criticism, but globally there was far more supporters than haters. This led Microsoft to give a focus on improving Windows XP’s security, rather than release a new operating system right away, perhaps one of the wisest design decision they ever made.

WINDOWS VISTA

The Veterans Health Information Systems and Technology Architecture (VistA) is an enterprise-wide information system built around an Electronic Health Record (EHR), used throughout the United States Department of Veterans Affairs (VA) medical system, known as the Veterans Health Administration (VHA). It's a collection of about 100 integrated software modules.

By 2003, the VHA was the largest single medical system in the United States, providing care to over 4 million veterans, employing 180,000 medical personnel and operating 163 hospitals, over 800 clinics, and 135 nursing homes. About a quarter of the nation's population is potentially eligible for VA benefits and services because they are veterans, family members, or survivors of veterans.

By providing electronic health records capability, VistA is thereby one of the most widely used EHRs in the world. Nearly half of all US hospitals that have a full implementation of an EHR are VA hospitals using VistA.

FeaturesThe Department of Veterans Affairs (VA) has had automated data processing systems, including extensive clinical and administrative capabilities, within its medical facilities since before 1985.[7] Initially called the Decentralized Hospital Computer Program (DHCP) information system, DHCP was enshrined as a recipient of the Computerworld Smithsonian Award for best use of Information Technology in Medicine in 1995.

VistA supports both ambulatory and inpatient care, and includes several significant enhancements to the original DHCP system. The most significant is a graphical user interface for clinicians known as the Computerized Patient Record System (CPRS), which was released in 1997. In addition, VistA includes computerized order entry, bar code medication administration, electronic prescribing and clinical

guidelines.

CPRS provides a client–server interface that allows health care providers to review and update a patient's electronic medical record. This includes the ability to place orders, including those for medications, special procedures, X-rays, nursing interventions, diets, and laboratory tests. CPRS provides flexibility in a wide variety of settings so that a consistent, event-driven, Windows-style interface is presented to a broad spectrum of health care workers.

Windows Vista Advantages:

1. Improved Graphical User Interface, the Aero:

There is an entire new GUI for Windows Vista. The appearance of window, desktop, start button, start menu, taskbar, everything got a new look. Now, the window in Windows Vista looks similar to that of in Mac OSX. Minimize, maximize and the close buttons wore an entire new strange look. The windows can be made to appear in translucent, 3D modes, if the new Aero is applied. The start button has been modified to give a new look. The word “start” has been removed. Only the new Windows logo gets appear in start button. Start menu also got a new look. The desktop wallpapers got a new look too. The desktop contains a side bar, which shows a clock, and some of the widgets. Widgets are small programs, which is equivalent to standard short cuts.

2. Manage your knotty kids, using parental controls:

One can control kids by using parental controls in Windows Vista. Parents can now deny access to certain programs especially, the games. Through separate child account, parents can now keep track of what all activities done by the child. It also enables parents to control the activities by blocking certain activities such as running a game program, blocking some sites that, the child might have visited etc. For this to work, parent must setup separate account for the child with different password.

3. Play multimedia contents using latest media player :

The latest version of Windows Media Player bundled with Windows Vista. The Windows Media Player

11 enhances video, audio quality when playing sound, music, videos. It has got new look and has got a series of rich features. These features include URGE services from MTV digital music services, Simplified Library, Library layouts. The newest URGE services provide access to the huge collection of online music, song collections. The simplified library improves customization options. Whenever, one plugs some of the latest media player compatible devices, the media player detects the device and installs the drivers associated with the device automatically.

4. Improved search, Windows Explorer :

In Windows Vista, you will get the search result as and when you type some letters of a particular word. The explorer shows a list of all files in the folder by default. When one types in, the few letters of words in the search box, the files will get filtered out according to the letters. Suppose, if one searches for filenames of “govt. services”, the search will filter out and displays only files with filenames with the name “gov”, if you type “gov” in the search text box. It will further filter the filenames, when you type full form. Also, the look of Windows Explorer got changed rapidly. More sorting options are also provided. Different kinds of listing is also got introduced.

5. Manage your photos using photo gallery :

Windows Vista has got an inbuilt tool called as Photo Gallery to manage digital photos. Using it, one can modify the photos. It has got ability to show previews of movie files, afeature, which was not there with Windows XP. Using this tool, one can transfer or share the photos from computer to devices such as mobile phones, PDA’s, etc. The photo gallery also has the ability to work high precision, high quality images, which are produced by latest high quality cameras.

Vista Disadvantages:

1. Requirements high, too hot deal:

Windows Vista, the next generation operating system requires superb computer to install. It consumes a lot more resources than its predecessor windows XP. For vista Aero to work, you need to invest something more than what you would have expected to invest before. The graphics card requirement is quite high. The graphics card must be of DirectX 10 supported. The required minimum graphics memory on the graphics card is 128 MB. For better performance, graphics memory must be of 256 MB. The

minimum hard disk space required is 20 GB for installation of Windows Vista. Processor speed is also expected to be high for Aero to work smoothly without causing any hiccups. The memory requirement for Windows Vista system is high. The system will work smoothly when one has got RAM of about 2 GB installed. Totally, I would like to say that, Windows vista eats up lot of resources, than its predecessor, the Windows XP.

2. Regarding price, sucking your hard earned money:

The price of Windows Vista Ultimate edition seems to be too high. Ordinary user cannot thus, have a look into all the features of Vista, which is only available in the Ultimate Edition. Microsoft seems to be not so interested in country wise markets. The prices are set according to the US market. However these prices are on higher side for developing asian countries. Microsoft should set the price of Vista as per the market. It is the time for Microsoft to think over this aspect, having introduced stricter validation process.

3. Replace the oldies, another way of extracting:

This is a usual problem, which always exists whenever new operating system gets released, and you have an old system or have some old components in the system. If one decides to install windows vista in the old system, then they should check the compatibility of system components, checking whether Vista supports them or not. Some manufacturers will still provide Vista support for the oldies by providing the latest drivers for the same. If manufacturers are not providing the latest drivers, you need to purchase a new device as a replacement for the existing device. Before, you purchase a new device you need to check out for Vista compatible or Vista ready logo, which is put up on the device.

4.Window appearance. Did you like the different look?

In Windows Vista, the window appearance underwent lot of unwanted changes. The windows in Windows Vista, the window appears similar to that of MAC OSX. The Minimize, Maximize, and Close buttons, wore a different look. The three buttons got reduced in their sizes, making them unclear to aged people, people with eye sight problems. The minimize, maximize, and close buttons have reduced in sizes. The icon in the other end has disappeared, which makes the window slightly dull in appearance.

In overall, the new operating system, Windows Vista is said to be an ideal replacement to its predecessor, the Windows XP.

WINDOWS 7

This led Microsoft to be more cautious with the next release of Windows, Windows 7. They didn’t gave up on any idea or technology from Vista, even keeping the highly laughable Flip 3D, but didn’t introduced lots of destabilizing new feature this time, and rather focused on improving performance as much as technically possible and fine-tuning any new concept from Vista. It included :

Major debugging and optimisation effort : UAC showed less popups. Disk footprint was reduced to around 6 GB and memory footprint to around 512 MB. CPU usage was dropped. Animations were a little bit more quiet. The system boots and shuts down faster than Vista, though a bit slower than Windows XP (but the desktop is available faster). In one word, they made software that used Vista technology while being able to run smoothly on a netbook, and that was somewhat less annoying from an end user perspective.

Some applications disappeared : Perhaps the first time Microsoft do this. Most Windows Live-related content (Messenger/Calendar/Mail) was moved out of Windows, along with the movie making functions. Those are now only available for download on Windows Live’s website.

Some advanced options disappeared : In Windows Vista, it was possible to turn off the transparency effects of the windows, in order to see their title better or to push laptop battery life a bit further. In Windows 7, I didn’t get to find that settings window anymore and suspect that it’s gone. Windows 7 also introduced a new way to manage peripherals that’s a lot more user-friendly than the old one, but that feels somewhat incomplete facing its older counterpart. The advanced search windows, along with its interesting options like specifying where to search and the like, is now gone, making using search bars mandatory.

The Sidebar disappeared : No gray translucent thing on the desktop anymore. However, its gadgets are still around. They just show up directly on the desktop now.

Focus on touch screens : Windows 7 finishes the work in Vista to introduce proper support for this. By proper support, I mean that one may now easily right-click, scroll in windows, zoom in and out, get an on-screen keyboard, or use some windows features through gestures. In no way were Windows applications display modified to make them touch-friendly, meaning that menus, toolbars, and the like are pretty much unusable because of their small size. Only some special touch-enhanced applications are available. Still Microsoft has time : computers with touch screen still aren’t common or useful enough to make improvements like this the top priority.

Libraries : Microsoft discovers symbolic linking technology. Libraries are a special kind of folder that only contains links to the files in it. Their link nature is hidden to normal applications, making it possible to interact with those links exactly the same way one interacts with the real file. Such technology allows one to organize its file better, by putting them in more than one place without needing extra disk space. Its worth nothing to point out that this feature has been available in about all other significant OSs for ages, without the need to create a special kind of folder in order to use it.

Major Taskbar revamp : Made bigger in order to be easier operated on touch screens, the Taskbar now steals several ideas and design principles from Apple OSX’s own taskbar-like application called the Dock

(Apple having themselves stolen the taskbar idea from Microsoft, this looks somewhat fair) and introduces some new concepts. First, it’s now application-oriented, meaning that instead of showing one button for each opened window it shows one button for each opened application. Then there’s only few distinction between buttons to launch favorite programs and opened windows, especially since program icons aren’t accompanied by their title anymore : there’s only icons. That’s somewhat bad, because it’s sometimes not that easy to figure out what an icon is about, introducing the need to move the mouse on the icon in order to know more.When one moves the mouse on top of the icon corresponding to an opened program, a popup appears, allowing one to choose one of the opened windows (this choice being helped by hiding all other windows with a useless glassy effect, see screenshot below). Another improvement, minor at the moment but having some potential, is the introduction of Jump lists : by right-clicking on one program’s icon, one gets access to a list of commonly used options and recently opened files, provided said program is compatible with this feature.Last but not least. the Notification area has been redone. Windows XP introduced the ability to auto-hide inactive icons, which was a feature somewhere between inefficient and useless, but Windows 7 finally introduces something grand : the ability to prevent them from showing popup balloons endlessly. Basically, if a program keeps annoying you with stupid popups such as “connexion lost, awaiting connexion…” or “is your antivirus up to date ?”, you may now tell it to shut up either definitely, or the time before you’re ready to listen to them. That’s what any long-time Windows user will call really great !

The new Taskbar

Various Start menu tweaks : The shut down button now really shuts down the computer, and is labelled as such to advertise this important improvement (though its function may be changed, too). Jump lists of applications are available here too. The search box may now search control panel items.

Various windows management tweaks : Some windows management can now be done using mouse gestures, by grabbing the windows to one side of the screen. It’s now easier to have two maximized windows side by side : dragging one to the left makes it maximized and taking the left half of the screen, dragging the other to the right make it do the same but on the right side of the screen.

Microsoft Windows 7

Okay, so now is the time for a global review of Microsoft Windows and for a word of conclusion.

CONCLUSION

So what can be globally said about Microsoft Windows and remembered as a lesson from the past ?

Microsoft are sometimes extremely good at keeping compatibility… : This is one of the best explanations to Microsoft’s monopoly nowadays, since it’s one of those things that allow Windows’s application catalog to be so huge. Though, sadly, it’s often the applications that matter (like games or working applications costing hundreds of dollars) that won’t work, lots of Windows 1 applications still run on windows vista 32bit, even though it uses a totally different core. In the same way, core interface concepts like the Taskbar, windows maximize/iconize/close buttons, or file explorer, have gone a long way without significant modification, and still no one would complain about them before Vista and 7 would tweak it.

…but can’t prevent themselves from introducing new features and re-designing the API… : This is not always a bad thing, since it brought several innovations in the computing world. But overdoing this is dangerous. The Windows NT structure is insanely complicated, because for almost each problem encountered in operating system design it’s made so that it would support many different solutions provided in the past. And that’s just the core. Almost each new releases of Windows, introduces quite a revolution in the development kit, meaning that developers have to get used to a new way of doing things in order to benefit from the new features and to get compatibility. And, as you’d guess, only the latest version is totally supported by Microsoft engineers, though the older GDI, MFC, and the like still remain around. In one word, compatibility + feature-adding frenzy = a mess. From an average user’s point of view, Windows used to be somewhat more stable and less… surprising, but starting with Vista, it looks like those times are over.

Features aren’t always well-chosen nor well-designed : Who the hell needs Wordpad for anything ? Whatever you may think of, this piece of software sucks at it… Vista’s explorer revamp was globally unwanted and not very well-perceived, whereas one had to wait for Windows 7 to see the cleanup and increase in performance that just everybody was waiting for after the release of Windows XP.

Early releases introduced a good equilibrium between use of icons and text : This has somewhat been lost starting with Windows Vista, sadly.

Windows introduced the problem of managing big menus or lists : Just look at the Programs menu length on most computers…

Windows introduced monopoly abuse, too : And how ! If I was to choose just one example, it would be the “integration” of IE, Media Player, and C#/.NET…

Windows gets globally too much in the way of the user since W95 : Popups filled with garbage, endless “are you sure you want to do this ?” questioning, lack of system access for the user since windows Me… And let’s not talk about all those Windows Update popup just waiting for you to click “OK” instead of just doing it themselves…

Case study 3 - UNIX

Introduction:Unix (officially trademarked as UNIX, sometimes also written as UNIX with small caps) is a multitasking, multi-user computer operating system originally developed in 1969 by a group of AT&T employees at Bell Labs, including Ken Thompson, Dennis Ritchie, Brian Kernighan, Douglas McIlroy, and Joe Ossanna. The Unix operating system was first developed in assembly language, but by 1973 had been almost entirely recoded in C, greatly facilitating its further development and porting to other hardware. Today's Unix systems are split into various branches, developed over time by AT&T as well as various commercial vendors and non-profit organizations.The Open Group, an industry standards consortium, owns the “UNIX” trademark. Only systems fully compliant with and certified according to the Single UNIX Specification are qualified to use the trademark; others might be called "Unix system-like" or "Unix-like" (though the Open Group disapproves of this term[1]). However, the term "Unix" is often used informally to denote any operating system that closely resembles the trademarked system.During the late 1970s and early 1980s, the influence of Unix in academic circles led to large-scale adoption of Unix (particularly of the BSD variant, originating from the University of California, Berkeley) by commercial startups, the most notable of which are Solaris, HP-UX and AIX. Today, in addition to certified Unix systems such as those already mentioned, Unix-like operating systems such as Linux and BSD descendants (FreeBSD, NetBSD, and OpenBSD) are commonly encountered. The term "traditional Unix" may be used to describe a Unix or an operating system that has the characteristics of either Version 7 Unix or UNIX System V.

Features Of Unix• multi-usermore than one user can use the machine at a timesupported via terminals (serial or network connection) • multi-tasking

more than one program can be run at a time • hierarchical directory structureto support the organisation and maintenance of files • portabilityonly the kernel ( <10%) written in assemblera wide range of support tools (debuggers, compilers) • The tools for program development

UNIX Operating SystemConsists of • kernelschedules tasksmanages data/file access and storageenforces security mechanismsperforms all hardware access • shellpresents each user with a promptinterprets commands types by a userexecutes user commandssupports a custom environment for each user • utilitiesfile management (rm, cat, ls, rmdir, mkdir)user management (passwd, chmod, chgrp)process management (kill, ps)printing (lpr) 

Unix ShellA Unix shell is a command-line interpreter or shell that provides a traditional user interface for the Unix operating

system and forUnix-like systems. Users direct the operation of the computer by entering command input as text for a command line interpreterto execute or by creating text scripts of one or more such commands.

The most influential Unix shells have been the Bourne shell and the C shell. The Bourne shell, sh, was written by Stephen Bourne at AT&T as the original Unix command line interpreter; it introduced the basic features common to all the Unix shells, including piping, here documents, command substitution, variables, control

structures for condition-testing and looping and filename wildcarding. The language, including the use of a reversed keyword to mark the end of a block, was influenced byALGOL 68.[1]

The C shell, csh, was written by Bill Joy while a graduate student at University of California, Berkeley. The language, including the control structures and the expression grammar, was modeled on C. The C shell also introduced a large number of features for interactive work, including the history and editing mechanisms, aliases, directory stacks, tilde notation, cdpath, job control andpath hashing.

Both shells have been used as coding base and model for many derivative and work-alike shells with extended feature sets.

Concept

The most generic sense of the term shell means any program that users employ to type commands. In the Unix operating system users may select which shell to use for interactive sessions. When the user logs in to the system the shell program is automatically executed. Many types of shells have been developed for this purpose. The program is called a "shell" because it hides the details of the underlying operating system behind the shell's interface. The shell manages the technical details of the operating system kernel interface, which is the lowest-level, or 'inner-most' component of an operating system.

Similarly, graphical user interfaces for Unix, such as GNOME, KDE, and Xfce can be called visual shells or graphical shells.

By itself, the term shell is usually associated with the command line. In Unix, users who want to use a different syntax for typing commands can specify a different program as their shell, though in practice this usually requires administrator rights.

The Unix shell was unusual when it was first created. Since it is both an interactive command language as well as a scripting programming language it is used by Unix as the facility to control (see shell script) the execution of the system.

Many shells created for other operating systems offer rough equivalents to Unix shell functionality.

On systems using a windowing system, some users may never use the shell directly. On Unix systems, the shell is still the implementation language of system startup scripts, including the program that starts the windowing system, the programs that facilitate access to the Internet, and many other essential functions.

Many users of a Unix system still find a modern command line shell more convenient for many tasks than any GUI application.

Due to the recent movement in favor of free and open source software, most Unix shells have at least one version that is distributed under an open source or free software license.

Bourne shell

The Bourne shell was one of the major shells used in early versions of the Unix operating system and became a de facto standard. It was written by Stephen Bourne at Bell Labs and was first distributed with Version 7 Unix, circa 1977. Every Unix-like system has at least one shell compatible with the Bourne shell. The Bourne shell program name is sh and it is typically located in the Unix file system hierarchy at /bin/sh. On many systems, however, /bin/sh may be a symbolic link or hard link to a compatible, but more feature-rich shell than the Bourne shell. The POSIX standard specifies its standard shell as a strict subset of the Korn shell. From a user's perspective the Bourne shell was immediately recognized when active by its characteristic default command line prompt character, the dollar sign ($).

C shell

The C shell was developed by Bill Joy for the Berkeley Software Distribution, a line of Unix operating systems derived from Unix and developed at the University of California, Berkeley. It was originally derived from the 6th Edition Unix shell (Thompson shell). Its syntax is modeled after the C programming language. It is used primarily for interactive terminal use, but less frequently for scripting and operating system control. C shell has many interactive commands.

Advantages Of Unix• Full multitasking with protected memory. Multiple users can run multiple programs each at the same time without interfering with each other or crashing the system. • Very efficient virtual memory, so many programs can run with a modest amount of physical memory. • Access controls and security. All users must be authenticated by a valid account and password to use the system at all. All files are owned by particular accounts. The owner can decide whether others have read or write access to his files. • A rich set of small commands and utilities that do specific tasks well -- not cluttered up with lots of special options. Unix is a well-stocked toolbox, not a giant do-it-all Swiss Army Knife. • A lean kernel that does the basics for you but doesn't get in the way when you try to do the unusual. • Available on a wide variety of machines - the most truly portable operating system. • Optimized for program development, and thus for the unusual circumstances that are the rule in research. 

Disadvantages Of Unix• The traditional command line shell interface is user hostile -- designed for the programmer, not the casual user. • Commands often have cryptic names and give very little response to tell the user what they are doing. Much use of special keyboard characters - little typos have unexpected results. • To use Unix well, you need to understand some of the main design features. Its power comes from knowing how to make commands and programs interact with each other, not just from treating each as a fixed black box.

Case study 4 – APPLE MAC OS

Mac OS is a series of graphical user interface-based operating systems developed by Apple Inc. (formerly Apple Computer, Inc.) for their Macintosh line of computer systems. The Macintosh user experience is credited with popularizing the graphical user interface. The original form of what Apple would later name the "Mac OS" was the integral and unnamed system software first introduced in 1984 with the original Macintosh, usually referred to simply as the System software.

Apple deliberately downplayed the existence of the operating system in the early years of the Macintosh[citation needed] to help make the machine appear more user-friendly and to distance it from other operating systems such as MS-DOS, which was more arcane and technically challenging. Much of this early system software was held in ROM, with updates typically provided free of charge by Apple dealers on floppy disk. As increasing disk storage capacity and performance gradually eliminated the need for storing much of the advanced GUI operating system in the ROM, Apple explored clones while positioning major operating system upgrades as separate revenue-generating products, first with System 7.1 and System 7.5, then with Mac OS 7.6 in 1997.

Early versions of the Mac OS were compatible only with Motorola 68000-based Macintoshes. As Apple introduced computers with PowerPC hardware, the OS was ported to support this architecture as well. Mac OS 8.1 was the last version that could run on a 68000-class processor (the 68040). Mac OS X, which has superseded the "Classic" Mac OS, is compatible with both PowerPC and Intel processors through to version 10.5 ("Leopard"). Version 10.6 ("Snow Leopard") supports only Intel processors.

Mac OS can be divided into two families:

The Mac OS Classic family, which was based on Apple's own code

The Mac OS X operating system, developed from Mac OS Classic family, and NeXTSTEP, which was UNIX-based.

Classic" Mac OS (1984–2001)Main article: History of Mac OS

Original 1984 Macintosh desktop

The "classic" Mac OS is characterized by its total lack of a command line; it is a completely graphical operating system. Noted for its ease of use and its cooperative multitasking, it was criticized for its very limited memory management, lack of protected memory, and susceptibility to conflicts among operating system "extensions" that provide additional functionality (such as networking) or support for a particular

device. Some extensions may not work properly together, or work only when loaded in a particular order. Troubleshooting Mac OS extensions could be a time-consuming process of trial and error.

The Macintosh originally used the Macintosh File System (MFS), a flat file system with only one level of folders. This was quickly replaced in 1985 by the Hierarchical File System (HFS), which had a true directory tree. Both file systems are otherwise compatible.

Most file systems used with DOS, Unix, or other operating systems treat a file as simply a sequence of bytes, requiring an application to know which bytes represent what type of information. By contrast, MFS and HFS give files two different "forks". The data fork contains the same sort of information as other file systems, such as the text of a document or the bitmaps of an image file. The  resource

fork contains other structured data such as menu definitions, graphics, sounds, or code segments. A file might consist only of resources with an empty data fork, or only a data fork with no resource fork. A text file could contain its text in the data fork and styling information in the resource fork, so that an application, which doesn’t recognize the styling information, can still read the raw text. On the other hand, these forks provide a challenge to interoperability with other operating systems; copying a file from a Mac to a non-Mac system strips it of its resource fork, necessitating such encoding schemes as BinHex and MacBinary.

PowerPC versions of Mac OS X up to and including Mac OS X v10.4 Tiger (support for Classic was dropped by Apple with Leopard's release and it is no longer included) include a compatibility layer for running older Mac applications, the Classic Environment. This runs a full copy of the older Mac OS, version 9.1 or later, in a Mac OS X process. PowerPC-based Macs shipped with Mac OS 9.2 as well as Mac OS X. Mac OS 9.2 had to be installed by the user — it was not installed by default on hardware revisions released after the release of Mac OS X 10.4 Tiger. Most well-written "classic" applications function properly under this environment, but compatibility is only assured if the software was written to be unaware of the actual hardware, and to interact solely with the operating system. The Classic Environment is not available on Intel-based Macintosh systems due to the incompatibility of  Mac OS

9 with the x86 hardware.

Users of the classic Mac OS generally upgraded to Mac OS X, but many criticized it as being more difficult and less user-friendly than the original Mac OS, for the lack of certain features that had not been re-implemented in the new OS, or for being slower on the same hardware (especially older hardware), or other, sometimes serious incompatibilities with the older OS. Because drivers (for printers, scanners, tablets, etc.) written for the older Mac OS are not compatible with Mac OS X, and due to the lack of Mac OS X support for older Apple machines, a significant number of Macintosh users have still continued using the older classic Mac OS.

In June 2005, Steve Jobs announced at the Worldwide Developers Conference keynote that Apple computers would be transitioning from PowerPC to Intel processors and thus dropping compatibility on new machines for Mac OS Classic. At the same conference, Jobs announced Developer Transition Kits that included beta versions of Apple software including Mac OS X that developers could use to test their applications as they ported them to run on Intel-powered Macs. In January 2006, Apple released the

first Macintosh computers with Intel processors, an iMac and the MacBook Pro, and in February 2006, Apple released a Mac mini with an Intel Core Solo and Duo processor. On May 16, 2006, Apple released the MacBook, before completing the Intel transition on August 7 with the Mac Pro. To ease the transition for early buyers of the new machines, Intel-based Macs include an emulation technology called  Rosetta, which allows them to run Mac OS X software that was compiled for PowerPC-based Macintoshes. Rosetta runs transparently, creating a user experience identical to running the software on a PowerPC machine, though execution is typically slower than with native code.

Mac OS X Mac OS X is the newest of Apple Inc.'s Mac OS line of operating systems. Although it is officially designated as simply "version 10" of the Mac OS, it has a history largely independent of the earlier Mac OS releases.

The operating system is the successor to Mac OS 9 and the "classic" Mac OS. It is a Unix operating system, based on the NeXTSTEP operating system and the Mach kernel which Apple acquired after purchasing NeXT Computer, with its CEO Steve Jobs returning to Apple at this time. Mac OS X also makes use of the BSD code base. There have been six significant releases of the client version, the most recent being Mac OS X 10.6, referred to as Snow Leopard. On Apple's October 20th 2010 "Back to the Mac" event, Mac OS X 10.7 Lion was previewed, showing improvements and additions including a Mac App Store.

As well as the client versions, Mac OS X has also had six significant releases as a server version, called Mac OS X Server. The first of these, Mac OS X Server 1.0, was released in beta in 1999. The server versions are architecturally identical to the client versions, with the differentiation found in their inclusion of tools for server management, including tools for managing Mac OS X-basedworkgroups, mail servers, and web servers, amongst other tools. It is currently the default operating system for the Xserve server hardware, and as an optional feature on the Mac Mini, as well as being installable on most other Macs. Unlike the client version, Mac OS X Server can be run in a virtual machine using emulation software such as Parallels Desktop.

Mac OS X is also the basis for iOS, (previously iPhone OS) used on Apple's iPhone, iPod Touch, and iPad, as well as being the basis for the operating system used in the Apple TV.

Advantages of Apple Computers

Security. Apple computers are much more secure than Windows PCs. Viruses, adware and malware designed for a Windows-based processor simply will not run on a Mac. The main reason for this is simply that there are much fewer Macs in use. People who make malicious software want to have it spread to as many people as possible and designing a virus for a Mac simply will not accomplish this.

Reliability. The people who make Mac software are the same people who make the Mac hardware. While Windows programmers have to take into account nearly infinite variations in hardware, Mac OS is designed to be used on a very limited amount of computers: those built by Apple. The software is designed specifically to run on the hardware. This means that the operating system is much more stable.

Macs are convenient. Do you own an iPod? An iPhone? Do you listen to music with iTunes? Do you appreciate how simple, elegant and easy to use these products are? Apple applies all of the things you like about these products to their computers. Yes, there will be a time of adjustment to using Mac OS, but once you are used to it, you will love it.

Advanced technology. The current generation of Macbook Pros feature LED monitors, multi-touch mouse trackpad and a CNC machined aluminum case. Desktop Mac Pros feature up to 8 core processors (yes, eight) and up to 4 TB of storage (TB=Terabyte). Thats four thousand (4000) GB.

A good overview of the Macbook Pro is available here.

A Mac can do anything a Windows PC can. With Apple computers now using Intel processors, they are able to run Windows. With programs like Bootcamp or software emulation, you are able to install and run Windows on your Mac and switch between the two OSs easily. Macs come with a Mail utility with the same functionality as Microsoft Outlook. The iWork line of software comes with Pages, Keynote and Numbers which can function as replacements for Word, Powerpoint and Excel. Each of these programs can export files in a format that can be used in Windows software. Or, if you don’t want to try iWork, Microsoft offers a Mac-compatible version of Office.

Disadvantages of Mac Computers.

They’re only more secure because fewer people use them. If everybody used Macs, there would

be a lot more viruses and malware for them. If Apple computers became more popular, they would

become less secure.

Cannot be upgraded/customized. There are upgrade options when you buy a Mac, but unlike a

Windows PC, you cannot mix and match components. The easiest way to upgrade a Mac is to buy a new

one.

Price. Macs are very expensive. Even the cheapest laptop costs $999.99. Windows machines cost as

little as $500.

Playing games requires Windows. There are very few games available for the Mac. If you are a

gamer, a Mac is probably not the best choice. You could run Bootcamp, but if you are going to spend

most of your time on a Mac running Windows, you have to ask yourself if it is worth it. Another option (and

the one I have taken) is to play games on a video game console, such as a PS3 or Wii, and simply use

your computer for computing.

It requires adjustment. It’s a Windows-based world. Most people are used to using Windows. Changing to

a Mac requires that you get used to a number of differences. Some people are simply not willing to make

that change.

Research Paper help

https://www.homeworkping.com/