design alternative for parallel systems
DESCRIPTION
Bachelor Thesis on Parallel computing.TRANSCRIPT
Design Alternative for Parallel System[root@aissms ~ ]# mount /dev/Parallex /mnt/presentation
Presented by:
Amit Kumar B32*****7
Ankit Singh B32*****8
Sushant Bhadkamkar B32*****2
GUIDE: Mr. Anil.J. Kadam
Department of Computer Engineering,
AISSMS College of Engineering,
Pune - 1
[root@aissms]# cat /mnt/presentation/AUTHORS
Overview[root@aissms ~]# tree /mnt/presentation
Introduction What is Parallel computing?
- Introduction to parallel computing,
- Who uses parallel computing.? - Why parallel computing?
Hardware & software resources.
Technical design overview
Implementation briefing
Phase I results
Applications
Advantages
Conclusion
References
Introduction- What is Parallel computing?
Parallel computing is the simultaneous execution of the
same task (split up and specially adapted) on multiple Processors in
order to obtain results faster.
In the simplest sense, parallel computing is the simultaneous use of
multiple compute resources to solve a computational problem.
- To be run using multiple CPUs
- A problem is broken into discrete parts that can be solved
concurrently
[root@aissms ~]# grep /mnt/parallex Introduction
Introduction[root@aissms ~]# grep /mnt/parallex Introduction
-Amdahl’s Law
If the sequential component of an algorithm accounts for 1/s of the program's execution time, then the maximum possible speedup that can be achieved on a parallel computer is ‘s’
[root@aissms ~]# awk USAGE /mnt/parallex
- Who uses parallel computing?
Introduction
[root@aissms ~]# sed /mnt/parallex PARALLEL
- Why parallel computing?
The primary reasons for using parallel computing:
-Save time - wall clock time -Solve larger problems -Provide concurrency (do multiple things at the same time) -Taking advantage of non-local resources -Cost savings
Limits to serial computing :
-Transmission speeds -Limits to miniaturization -Economic limitations
Introduction
[root@aissms ~]# cat Hardware | more
Hardware:
x686 Class PCs (installed with intranet connection)SwitchSerial port connectors100 BASE T LAN cable , RJ 45 connectors
Software:
Linux (2.6.x kernel)Intel Compiler suite (Noncommercial)LSB ( Linux Standard Base ) Set of GNU Kits with GNU CC/C++/F77/LD/AS
Hardware and Software Resources
Design Overview[root@aissms ~]# echo $DESIGN_OVERVIEW
Phase I Implementation [root@aissms ~]# echo
-NFS mounted on all nodes. (implementing shared memory)
-Status of nodes
-A test application sent to all host to determine current load on the processor.
-Developed a distribution algorithm to break the task according to load capacity of processor given by test app.
-All task received by server & integrate the result & give output on server terminal.
Phase I Results[root@aissms ~]# echo $CONCLUSION
Execution of application on single machine
Applications[root@aissms ~]#
- High processing requirement tasks
- Molecular dynamics
- Astronomical modeling
- Data mining
- Image rendering
- Clustering is now used for mission-critical applications such as web and FTP
servers
- Google uses an ever-growing cluster composed of tens of thousands of
computers
- Scientific Calculations consisting of complex numerical calculations
Advantages[root@aissms ~]# ls -lh ‘Advantages*’
- Implemented parallelism at every level.
- Parallel systems implemented on available hardware.
- Diskless technology.
-Cost (central storage solution)
-Error recovery
-Initialization
-Optimum utilization of available resources.
Conclusion[root@aissms ~]# echo $CONCLUSION
By Implementing parallelism on all levels and making efficient utilization of
available hardware resources, we attempt to provide cost effective solution
for small & medium scale businesses and research institutes.
And,
We are in process of developing Mini Super computer.
References[root@aissms ~]# find / -name “*Parallex*”
[1] Parallel Computer Architectures : Hardware/Software Approach. Culler, David. Morgran Coffman Publishers. San Fransisco,CA.
[2] High Performance Computing. 2nd Edition, Dowd Kavin and Charles. Sebastopol , CA : ORielly and Associates
[3] Source Book of Parallel Computing: Dongara, Jack. Morgran Coffman Publishers. San Fransisco,CA
[4] High Performance Linux Cluster,Joseph Sloan.Sebastopol,CA:O’ReillyMedia Inc.
[5] Parallel Computing on Heterogeneous Networks by Alexey L. Lastovetsky
[6] Kernel Sources from http://ww.kernel.org
Thank you![root@aissms ~]# killall Parallex
Any Questions ?