SlideShare a Scribd company logo
1 of 21
Create an Office Mix 
Let’s Get Started…
Topic Overview 
•What is Parallel Computing? 
•Why Parallel Computing ? 
•Scope of Parallel Computing Applications
What is Parallel Computing?
What is Parallel Computing? 
• Traditionally, software has been written for serial 
computation: 
– To be run on a single computer having a single 
Central Processing Unit (CPU); 
– A problem is broken into a discrete series of 
instructions. 
– Instructions are executed one after another. 
– Only one instruction may execute at any moment 
in time.
• In the simplest sense, parallel computing is the 
simultaneous use of multiple compute resources to solve a 
computational problem. 
– To be run using multiple CPUs 
– A problem is broken into discrete parts that can be 
solved concurrently 
– Each part is further broken down to a series of 
instructions 
• Instructions from each part execute simultaneously on 
different CPUs
Why parallel computing ?
Limitations of Serial Computing….. 
•Limits to serial computing - both physical and practical reasons pose 
significant constraints to simply building ever faster serial computers. 
•Transmission speeds - the speed of a serial computer is directly dependent 
upon how fast data can move through hardware. Absolute limits are the 
speed of light (30 cm/nanosecond) and the transmission limit of copper wire 
(9 cm/nanosecond). Increasing speeds necessitate increasing proximity of 
processing elements. 
•Limits to miniaturization - processor technology is allowing an increasing 
number of transistors to be placed on a chip. However, even with molecular or 
atomic-level components, a limit will be reached on how small components 
can be. 
•Economic limitations - it is increasingly expensive to make a single processor 
faster. Using a larger number of moderately fast commodity processors to 
achieve the same (or better) performance is less expensive.
Why Parallel 
Computing? 
The primary reasons for using 
parallel computing: 
•Save time - wall clock time 
•Solve larger problems 
•Provide concurrency (do multiple 
things at the same time)
Other reasons might include: 
– Taking advantage of non-local resources - using 
available compute resources on a wide area network, or 
even the Internet when local compute resources are 
scarce. 
– Cost savings - using multiple "cheap" computing 
resources instead of paying for time on a 
supercomputer. 
– Overcoming memory constraints - single computers 
have very finite memory resources. For large problems, 
using the memories of multiple computers may 
overcome this obstacle.
Moore’s Law Moore's law states [1965]: 
``The complexity for minimum 
component costs has increased at a rate of 
roughly a factor of two per year. Certainly 
over the short term this rate can be expected 
to continue, if not to increase. Over the 
longer term, the rate of increase is a bit more 
uncertain, although there is no reason to 
believe it will not remain nearly constant for 
at least 10 years. That means by 1975, the 
number of components per integrated circuit 
for minimum cost will be 65,000.''
Moore attributed this doubling rate to exponential behavior 
of die sizes, finer minimum dimensions, and ``circuit and 
device cleverness''. 
In 1975, he revised this law as follows: 
``There is no room left to squeeze anything out by 
being clever. Going forward from here we have to depend on the 
two size factors - bigger dies and finer dimensions.'' 
He revised his rate of circuit complexity doubling to 
18 months and projected from 1975 onwards at this reduced 
rate.
The Computational Power Argument 
•If one is to buy into Moore's law, the question still 
remains - how does one translate transistors into useful 
OPS (operations per second)? 
•The logical recourse is to rely on parallelism, both 
implicit and explicit. 
•Most serial (or seemingly serial) processors rely 
extensively on implicit parallelism. 
•We focus in this class, for the most part, on explicit 
parallelism.
The Memory/Disk Speed Argument 
• While clock rates of high-end processors have increased at 
roughly 40% per year over the past decade, DRAM access times 
have only improved at the rate of roughly 10% per year over this 
interval. 
• This mismatch in speeds causes significant performance 
bottlenecks. 
• Parallel platforms provide increased bandwidth to the memory 
system. 
• Parallel platforms also provide higher aggregate caches. 
• Principles of locality of data reference and bulk access, which 
guide parallel algorithm design also apply to memory 
optimization. 
• Some of the fastest growing applications of parallel computing 
utilize not their raw computational speed, rather their ability to 
pump data to memory and disk faster.
The Data Communication Argument 
• As the network evolves, the vision of the Internet as 
one large computing platform has emerged. 
• This view is exploited by applications such as 
SETI@home and Folding@home. 
• In many other applications (typically databases and 
data mining) the volume of data is such that they 
cannot be moved. 
• Any analyses on this data must be performed over 
the network using parallel techniques.
Scope of Parallel 
Computing Applications
Parallelism finds applications in very diverse 
application domains for different motivating reasons. 
These range from improved application performance to 
cost considerations.
• Design of airfoils (optimizing lift, drag, 
stability), internal combustion engines 
(optimizing charge distribution, burn), 
high-speed circuits (layouts for delays and 
capacitive and inductive effects), and 
structures (optimizing structural integrity, 
design parameters, cost, etc.). 
• Design and simulation of micro- and 
nano-scale systems. 
• Process optimization, operations research. 
Applications in 
Engineering 
and Design
• Functional and structural characterization of 
genes and proteins. 
• Advances in computational physics and chemistry 
have explored new materials, understanding of 
chemical pathways, and more efficient processes. 
• Applications in astrophysics have explored the 
evolution of galaxies, thermonuclear processes, 
and the analysis of extremely large datasets from 
telescopes. 
• Weather modeling, mineral prospecting, flood 
prediction, etc., are other important applications. 
• Bioinformatics and astrophysics also present 
some of the most challenging problems with 
respect to analyzing extremely large datasets. 
Scientific 
Applications
•Some of the largest parallel computers 
power the wall street! 
•Data mining and analysis for optimizing 
business and marketing decisions. 
•Large scale servers (mail and web servers) 
are often implemented using parallel 
platforms. 
•Applications such as information retrieval 
and search are typically powered by large 
clusters. 
Commercial 
Applications
• Network intrusion detection, cryptography, 
multiparty computations are some of the 
core users of parallel computing 
techniques. 
• Embedded systems increasingly rely on 
distributed control algorithms. 
• A modern automobile consists of tens of 
processors communicating to perform 
complex tasks for optimizing handling and 
performance. 
• Conventional structured peer-to-peer 
networks impose overlay networks and 
utilize algorithms directly from parallel 
computing. 
Applications 
in Computer 
Systems
Thank You!

More Related Content

What's hot

Introduction to Parallel and Distributed Computing
Introduction to Parallel and Distributed ComputingIntroduction to Parallel and Distributed Computing
Introduction to Parallel and Distributed ComputingSayed Chhattan Shah
 
System interconnect architecture
System interconnect architectureSystem interconnect architecture
System interconnect architectureGagan Kumar
 
Shared-Memory Multiprocessors
Shared-Memory MultiprocessorsShared-Memory Multiprocessors
Shared-Memory MultiprocessorsSalvatore La Bua
 
Multithreading computer architecture
 Multithreading computer architecture  Multithreading computer architecture
Multithreading computer architecture Haris456
 
Lecture 6.1
Lecture  6.1Lecture  6.1
Lecture 6.1Mr SMAK
 
Parallel computing and its applications
Parallel computing and its applicationsParallel computing and its applications
Parallel computing and its applicationsBurhan Ahmed
 
Parallel Programming
Parallel ProgrammingParallel Programming
Parallel ProgrammingUday Sharma
 
Limitations of memory system performance
Limitations of memory system performanceLimitations of memory system performance
Limitations of memory system performanceSyed Zaid Irshad
 
Operating system 31 multiple processor scheduling
Operating system 31 multiple processor schedulingOperating system 31 multiple processor scheduling
Operating system 31 multiple processor schedulingVaibhav Khanna
 
system interconnect architectures in ACA
system interconnect architectures in ACAsystem interconnect architectures in ACA
system interconnect architectures in ACAPankaj Kumar Jain
 
program partitioning and scheduling IN Advanced Computer Architecture
program partitioning and scheduling  IN Advanced Computer Architectureprogram partitioning and scheduling  IN Advanced Computer Architecture
program partitioning and scheduling IN Advanced Computer ArchitecturePankaj Kumar Jain
 
Pipelining and vector processing
Pipelining and vector processingPipelining and vector processing
Pipelining and vector processingKamal Acharya
 
Computer architecture multi processor
Computer architecture multi processorComputer architecture multi processor
Computer architecture multi processorMazin Alwaaly
 

What's hot (20)

Introduction to Parallel and Distributed Computing
Introduction to Parallel and Distributed ComputingIntroduction to Parallel and Distributed Computing
Introduction to Parallel and Distributed Computing
 
System interconnect architecture
System interconnect architectureSystem interconnect architecture
System interconnect architecture
 
Shared-Memory Multiprocessors
Shared-Memory MultiprocessorsShared-Memory Multiprocessors
Shared-Memory Multiprocessors
 
Cluster Computing
Cluster ComputingCluster Computing
Cluster Computing
 
Multithreading computer architecture
 Multithreading computer architecture  Multithreading computer architecture
Multithreading computer architecture
 
Lecture 6.1
Lecture  6.1Lecture  6.1
Lecture 6.1
 
Parallel computing and its applications
Parallel computing and its applicationsParallel computing and its applications
Parallel computing and its applications
 
Memory virtualization
Memory virtualizationMemory virtualization
Memory virtualization
 
Parallel Programming
Parallel ProgrammingParallel Programming
Parallel Programming
 
Limitations of memory system performance
Limitations of memory system performanceLimitations of memory system performance
Limitations of memory system performance
 
Memory management
Memory managementMemory management
Memory management
 
Mobile computing (Wireless) Medium Access Control (MAC)
Mobile computing (Wireless) Medium Access Control (MAC)Mobile computing (Wireless) Medium Access Control (MAC)
Mobile computing (Wireless) Medium Access Control (MAC)
 
Storage Structure in OS
Storage Structure in OSStorage Structure in OS
Storage Structure in OS
 
Operating system 31 multiple processor scheduling
Operating system 31 multiple processor schedulingOperating system 31 multiple processor scheduling
Operating system 31 multiple processor scheduling
 
system interconnect architectures in ACA
system interconnect architectures in ACAsystem interconnect architectures in ACA
system interconnect architectures in ACA
 
Mainframe systems
Mainframe systemsMainframe systems
Mainframe systems
 
program partitioning and scheduling IN Advanced Computer Architecture
program partitioning and scheduling  IN Advanced Computer Architectureprogram partitioning and scheduling  IN Advanced Computer Architecture
program partitioning and scheduling IN Advanced Computer Architecture
 
Advanced computer architecture
Advanced computer architectureAdvanced computer architecture
Advanced computer architecture
 
Pipelining and vector processing
Pipelining and vector processingPipelining and vector processing
Pipelining and vector processing
 
Computer architecture multi processor
Computer architecture multi processorComputer architecture multi processor
Computer architecture multi processor
 

Similar to Introduction to Parallel Computing

3 - Grid Computing.pptx
3 - Grid Computing.pptx3 - Grid Computing.pptx
3 - Grid Computing.pptxRiazSalim1
 
Fundamentals.pptx
Fundamentals.pptxFundamentals.pptx
Fundamentals.pptxdhivyak49
 
Simulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud InfrastructuresSimulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud InfrastructuresCloudLightning
 
Chap 2 classification of parralel architecture and introduction to parllel p...
Chap 2  classification of parralel architecture and introduction to parllel p...Chap 2  classification of parralel architecture and introduction to parllel p...
Chap 2 classification of parralel architecture and introduction to parllel p...Malobe Lottin Cyrille Marcel
 
CC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfCC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfHasanAfwaaz1
 
Grid computing Seminar PPT
Grid computing Seminar PPTGrid computing Seminar PPT
Grid computing Seminar PPTUpender Upr
 
Cloud Computing and Virtualization Overview by Amr Ali
Cloud Computing and Virtualization Overview by Amr AliCloud Computing and Virtualization Overview by Amr Ali
Cloud Computing and Virtualization Overview by Amr AliAmr Ali
 
(19-23)CC Unit-1 ppt.pptx
(19-23)CC Unit-1 ppt.pptx(19-23)CC Unit-1 ppt.pptx
(19-23)CC Unit-1 ppt.pptxNithishaYadavv
 
54665962-Nav-Cluster-Computing.pptx
54665962-Nav-Cluster-Computing.pptx54665962-Nav-Cluster-Computing.pptx
54665962-Nav-Cluster-Computing.pptxYashAhire28
 

Similar to Introduction to Parallel Computing (20)

Chap1 slides
Chap1 slidesChap1 slides
Chap1 slides
 
High–Performance Computing
High–Performance ComputingHigh–Performance Computing
High–Performance Computing
 
3 - Grid Computing.pptx
3 - Grid Computing.pptx3 - Grid Computing.pptx
3 - Grid Computing.pptx
 
CC unit 1.pptx
CC unit 1.pptxCC unit 1.pptx
CC unit 1.pptx
 
42 grid computing
42 grid computing42 grid computing
42 grid computing
 
42 grid computing
42 grid computing42 grid computing
42 grid computing
 
Grid computing
Grid computingGrid computing
Grid computing
 
Cluster Computing
Cluster ComputingCluster Computing
Cluster Computing
 
Fundamentals.pptx
Fundamentals.pptxFundamentals.pptx
Fundamentals.pptx
 
Simulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud InfrastructuresSimulation of Heterogeneous Cloud Infrastructures
Simulation of Heterogeneous Cloud Infrastructures
 
Chap2 slides
Chap2 slidesChap2 slides
Chap2 slides
 
CF-Chapitre1.pdf
CF-Chapitre1.pdfCF-Chapitre1.pdf
CF-Chapitre1.pdf
 
Grid Computing
Grid ComputingGrid Computing
Grid Computing
 
Chap 2 classification of parralel architecture and introduction to parllel p...
Chap 2  classification of parralel architecture and introduction to parllel p...Chap 2  classification of parralel architecture and introduction to parllel p...
Chap 2 classification of parralel architecture and introduction to parllel p...
 
CC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdfCC LECTURE NOTES (1).pdf
CC LECTURE NOTES (1).pdf
 
Grid computing Seminar PPT
Grid computing Seminar PPTGrid computing Seminar PPT
Grid computing Seminar PPT
 
Aca module 1
Aca module 1Aca module 1
Aca module 1
 
Cloud Computing and Virtualization Overview by Amr Ali
Cloud Computing and Virtualization Overview by Amr AliCloud Computing and Virtualization Overview by Amr Ali
Cloud Computing and Virtualization Overview by Amr Ali
 
(19-23)CC Unit-1 ppt.pptx
(19-23)CC Unit-1 ppt.pptx(19-23)CC Unit-1 ppt.pptx
(19-23)CC Unit-1 ppt.pptx
 
54665962-Nav-Cluster-Computing.pptx
54665962-Nav-Cluster-Computing.pptx54665962-Nav-Cluster-Computing.pptx
54665962-Nav-Cluster-Computing.pptx
 

More from Roshan Karunarathna

More from Roshan Karunarathna (8)

UI/UX presentation by Roshan Karunarathna
UI/UX presentation by Roshan KarunarathnaUI/UX presentation by Roshan Karunarathna
UI/UX presentation by Roshan Karunarathna
 
LIVE CHAT ETIQUETTE
LIVE CHAT ETIQUETTELIVE CHAT ETIQUETTE
LIVE CHAT ETIQUETTE
 
Iplanet
IplanetIplanet
Iplanet
 
onlinemarketing
onlinemarketingonlinemarketing
onlinemarketing
 
Pay Pal Introduction.........!
Pay Pal Introduction.........!Pay Pal Introduction.........!
Pay Pal Introduction.........!
 
Chou fasman algorithm for protein structure prediction
Chou fasman algorithm for protein structure predictionChou fasman algorithm for protein structure prediction
Chou fasman algorithm for protein structure prediction
 
Iterative and Incremental Development (RAD)
Iterative  and Incremental Development (RAD)Iterative  and Incremental Development (RAD)
Iterative and Incremental Development (RAD)
 
Introduction to backwards learning algorithm
Introduction to backwards learning algorithmIntroduction to backwards learning algorithm
Introduction to backwards learning algorithm
 

Recently uploaded

Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupFlorian Wilhelm
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Mark Simos
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxNavinnSomaal
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Manik S Magar
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfAlex Barbosa Coqueiro
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024The Digital Insurer
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 3652toLead Limited
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfRankYa
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostZilliz
 
The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfSeasiaInfotech2
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyAlfredo García Lavilla
 

Recently uploaded (20)

E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Streamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project SetupStreamlining Python Development: A Guide to a Modern Project Setup
Streamlining Python Development: A Guide to a Modern Project Setup
 
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationBeyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
Beyond Boundaries: Leveraging No-Code Solutions for Industry Innovation
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
Tampa BSides - Chef's Tour of Microsoft Security Adoption Framework (SAF)
 
SAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptxSAP Build Work Zone - Overview L2-L3.pptx
SAP Build Work Zone - Overview L2-L3.pptx
 
Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!Anypoint Exchange: It’s Not Just a Repo!
Anypoint Exchange: It’s Not Just a Repo!
 
Unraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdfUnraveling Multimodality with Large Language Models.pdf
Unraveling Multimodality with Large Language Models.pdf
 
My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024My INSURER PTE LTD - Insurtech Innovation Award 2024
My INSURER PTE LTD - Insurtech Innovation Award 2024
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365Ensuring Technical Readiness For Copilot in Microsoft 365
Ensuring Technical Readiness For Copilot in Microsoft 365
 
Search Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdfSearch Engine Optimization SEO PDF for 2024.pdf
Search Engine Optimization SEO PDF for 2024.pdf
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage CostLeverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
Leverage Zilliz Serverless - Up to 50X Saving for Your Vector Storage Cost
 
The Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdfThe Future of Software Development - Devin AI Innovative Approach.pdf
The Future of Software Development - Devin AI Innovative Approach.pdf
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
Commit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easyCommit 2024 - Secret Management made easy
Commit 2024 - Secret Management made easy
 

Introduction to Parallel Computing

  • 1. Create an Office Mix Let’s Get Started…
  • 2. Topic Overview •What is Parallel Computing? •Why Parallel Computing ? •Scope of Parallel Computing Applications
  • 3. What is Parallel Computing?
  • 4. What is Parallel Computing? • Traditionally, software has been written for serial computation: – To be run on a single computer having a single Central Processing Unit (CPU); – A problem is broken into a discrete series of instructions. – Instructions are executed one after another. – Only one instruction may execute at any moment in time.
  • 5. • In the simplest sense, parallel computing is the simultaneous use of multiple compute resources to solve a computational problem. – To be run using multiple CPUs – A problem is broken into discrete parts that can be solved concurrently – Each part is further broken down to a series of instructions • Instructions from each part execute simultaneously on different CPUs
  • 7. Limitations of Serial Computing….. •Limits to serial computing - both physical and practical reasons pose significant constraints to simply building ever faster serial computers. •Transmission speeds - the speed of a serial computer is directly dependent upon how fast data can move through hardware. Absolute limits are the speed of light (30 cm/nanosecond) and the transmission limit of copper wire (9 cm/nanosecond). Increasing speeds necessitate increasing proximity of processing elements. •Limits to miniaturization - processor technology is allowing an increasing number of transistors to be placed on a chip. However, even with molecular or atomic-level components, a limit will be reached on how small components can be. •Economic limitations - it is increasingly expensive to make a single processor faster. Using a larger number of moderately fast commodity processors to achieve the same (or better) performance is less expensive.
  • 8. Why Parallel Computing? The primary reasons for using parallel computing: •Save time - wall clock time •Solve larger problems •Provide concurrency (do multiple things at the same time)
  • 9. Other reasons might include: – Taking advantage of non-local resources - using available compute resources on a wide area network, or even the Internet when local compute resources are scarce. – Cost savings - using multiple "cheap" computing resources instead of paying for time on a supercomputer. – Overcoming memory constraints - single computers have very finite memory resources. For large problems, using the memories of multiple computers may overcome this obstacle.
  • 10. Moore’s Law Moore's law states [1965]: ``The complexity for minimum component costs has increased at a rate of roughly a factor of two per year. Certainly over the short term this rate can be expected to continue, if not to increase. Over the longer term, the rate of increase is a bit more uncertain, although there is no reason to believe it will not remain nearly constant for at least 10 years. That means by 1975, the number of components per integrated circuit for minimum cost will be 65,000.''
  • 11. Moore attributed this doubling rate to exponential behavior of die sizes, finer minimum dimensions, and ``circuit and device cleverness''. In 1975, he revised this law as follows: ``There is no room left to squeeze anything out by being clever. Going forward from here we have to depend on the two size factors - bigger dies and finer dimensions.'' He revised his rate of circuit complexity doubling to 18 months and projected from 1975 onwards at this reduced rate.
  • 12. The Computational Power Argument •If one is to buy into Moore's law, the question still remains - how does one translate transistors into useful OPS (operations per second)? •The logical recourse is to rely on parallelism, both implicit and explicit. •Most serial (or seemingly serial) processors rely extensively on implicit parallelism. •We focus in this class, for the most part, on explicit parallelism.
  • 13. The Memory/Disk Speed Argument • While clock rates of high-end processors have increased at roughly 40% per year over the past decade, DRAM access times have only improved at the rate of roughly 10% per year over this interval. • This mismatch in speeds causes significant performance bottlenecks. • Parallel platforms provide increased bandwidth to the memory system. • Parallel platforms also provide higher aggregate caches. • Principles of locality of data reference and bulk access, which guide parallel algorithm design also apply to memory optimization. • Some of the fastest growing applications of parallel computing utilize not their raw computational speed, rather their ability to pump data to memory and disk faster.
  • 14. The Data Communication Argument • As the network evolves, the vision of the Internet as one large computing platform has emerged. • This view is exploited by applications such as SETI@home and Folding@home. • In many other applications (typically databases and data mining) the volume of data is such that they cannot be moved. • Any analyses on this data must be performed over the network using parallel techniques.
  • 15. Scope of Parallel Computing Applications
  • 16. Parallelism finds applications in very diverse application domains for different motivating reasons. These range from improved application performance to cost considerations.
  • 17. • Design of airfoils (optimizing lift, drag, stability), internal combustion engines (optimizing charge distribution, burn), high-speed circuits (layouts for delays and capacitive and inductive effects), and structures (optimizing structural integrity, design parameters, cost, etc.). • Design and simulation of micro- and nano-scale systems. • Process optimization, operations research. Applications in Engineering and Design
  • 18. • Functional and structural characterization of genes and proteins. • Advances in computational physics and chemistry have explored new materials, understanding of chemical pathways, and more efficient processes. • Applications in astrophysics have explored the evolution of galaxies, thermonuclear processes, and the analysis of extremely large datasets from telescopes. • Weather modeling, mineral prospecting, flood prediction, etc., are other important applications. • Bioinformatics and astrophysics also present some of the most challenging problems with respect to analyzing extremely large datasets. Scientific Applications
  • 19. •Some of the largest parallel computers power the wall street! •Data mining and analysis for optimizing business and marketing decisions. •Large scale servers (mail and web servers) are often implemented using parallel platforms. •Applications such as information retrieval and search are typically powered by large clusters. Commercial Applications
  • 20. • Network intrusion detection, cryptography, multiparty computations are some of the core users of parallel computing techniques. • Embedded systems increasingly rely on distributed control algorithms. • A modern automobile consists of tens of processors communicating to perform complex tasks for optimizing handling and performance. • Conventional structured peer-to-peer networks impose overlay networks and utilize algorithms directly from parallel computing. Applications in Computer Systems