SlideShare a Scribd company logo
1 of 47
PARALLEL PROCESSING
Prepared By
Ms. Sheethal Aji
Mani
Assistant Professor
Kristu Jayanti
College
Parallel processing is an efficient form of information processing which
emphasizes the exploitation of concurrent events in the computing
process.
Parallel Processing demands concurrent execution of many programs
in the computer.
Concurrency implies parallelism, simultaneity, and pipelining.
INTRODUCTION TO
PARALLELISM IN
UNIPROCESSOR SYSTEM
From an operating point of view, computer systems have improved
chronologically in four phases:
• batch processing
• multiprogramming
• time sharing
•multiprocessing
In these four operating modes, the degree of parallelism increase sharply
from phase to phase.
We define parallel processing as
Parallel processing is an efficient form of information processing which emphasizes the exploitation of
concurrent events in the computing process.
Parallel Processing demands concurrent execution of many programs in the computer.
Concurrency implies parallelism, simultaneity, and pipelining.
The highest level of parallel processing is conducted among multiple jobs or programs through
multiprogramming, time sharing, and multiprocessing.
Parallel processing can be challenged in four programmatic levels:
1.Job or program level – highest level of parallel processing conducted among multiple jobs or programs
2. Task or procedure level – next higher level parallel processing conducted among procedures or tasks
within the same program
3. Interinstruction level- Third level to exploit concurrency among multiple instructions
4. Intrainstruction level-Finally concurrent operations within each instruction
BASIC UNIPROCESSOR
ARCHITECTURE
•A typical uniprocessor computer consists of the three major components the
main memory, the CPU (Central Processing Unit) and the I/O(Input-Output)
subsystem.
• There are two architectures of commercially available uniprocessor
computers to show the relation between three subsystems
SYSTEM ARCHITECTURE OF
THE SUPERMINI VAX-11/780
UNIPROCESSOR SYSTEM
There are 16, 32-bit general purpose register one of which is a Program Counter (PC).
There is also a special CPU status register containing about the current state of the
processor being executed
The CPU contains an ALU with an optional Floating-point accelerator, and some local
cache memory with an optional diagnostic memory
Floating-point accelerator (FPA) -A device to improve the overall performance of a computer
by removing the burden of performing floating-point arithmetic from the central processor.
optional diagnostic memory –checks the errors
•The CPU can be intervened by the operator through the console connected to floppy
disk
• The CPU, the main memory( 2^32 words of 32 bit each) and the I/O subsystem are
all connected to a common bus, the synchronous backplane interconnection(SBI)
• Through this bus, all I/O devices can communicate with each other with CPU or
with the memory
•I/O devices can be connected directly to the SBI through the unibus and its controller
or through a mass bus and its controller
SYSTEM ARCHITECTURE OF THE MAINFRAME IBM
SYSTEM 370/MODEL 168 UNIPROCESSOR
COMPUTER
•
•The CPU contains the instruction decoding and execution units as
well as cache
•Main memory is divided into four units referred to as logical storage
units (LSU) that are four ways interleaved
•The storage controller provides multiport Connections between the
CPU and the four LSU’s
• Peripherals are connected to the system via high speed I/O
channels which operate asynchronously with the CPU
Parallelism can be promoted by
Hardware means
Software means
PARALLELLISM IN UNIPROCESSOR
SYSTEM
A number of parallel processing mechanisms have been developed in
uniprocessor computers.
We identify them in the following six categories:
1.Multiplicity of functional units
2.Parallelism and pipelining within the CPU
3.Overlapped CPU and I/O operations
4.Use of a hierarchical memory system
5.Balancing of subsystem bandwidths
6.Multiprogramming and time sharing
1.MULTIPLICITY OF FUNCTIONAL
UNITS
Early computers - 1. one ALU that perform one operation at a time
2. Slow processing
Multiple and specialized functional units. - operate in parallel.
- two parallel execution units (fixed and floating point arithmetic) in CDC-
6600
→ 10 functional units
2.PARALLELISM AND PIPELINING
WITHIN THE CPU
Instead of serial bit adders , parallel adders are used in almost ALUs
Use of High speed multiplier recoding and convergence division
Sharing of hardware resources for functions of multiply and divide
Various phases of instructions executions (instruction fetch , decode,
operand fetch, arithmetic logic execution, store result) are pipelined
For overlapped instruction executions, instruction prefetch and
buffering techniques have been developed.
3.OVERLAPPED CPU AND I/O
OPERATIONS
I/O operations performed simultaneously with CPU computations
- by using separate I/O controllers, channels or I/O processors. Dat
DMA channel can be used for direct transfer of data between main
memory and I/O devices
USE OF A HIERARCHICAL
MEMORY SYSTEM
CPU is 1000 times faster than
memory access
Heirarchical memory system can
be used to close up the speed gap
Cache memory can be used
to be served as a buffer between
cpu and main memory
BALANCING OF SUBSYSTEM
BANDWIDTHS
CPU is the fastest unit in computer.
The bandwidth of a system is defined as the number of operations performed per unit time.
In case of main memory the memory bandwidth is measured by the number of words that can be accessed per unit
time.
Relationship b/w BWs
Bandwidth Balancing Between CPU and Memory
The speed gap between the CPU and the main memory can be closed up by using fast cache
memory between them.
A block of memory words is moved from the main memory into the cache so that immediate
instructions can be available most of the time from the cache.
Bandwidth Balancing Between Memory and I/O Devices
Input-output channels with different speeds can be used between the slow I/O devices and the
main memory.
The I/O channels perform buffering and multiplexing functions to transfer the data from multiple
disks into the main memory by stealing cycles from the CPU.
PARALLEL COMPUTER
STRUCTURES
A parallel computer is a kind of computer structure where a large complex problem
is broken into multiple small problems which are then executed simultaneously by
several processors. We often consider this as parallel processing.
It is divided into three
1.Pipeline Computer
2.Array Processors
3.Multiprocessor Systems
PIPELINE COMPUTER
Pipeline computers leverage parallel computing by overlapping the execution of
one process with another process.
If we consider executing a small program that has a set of instructions then we
can categorize the execution of the entire program in four steps that must be
repeated continuously till the last instruction in the program gets executed.
The four steps we mentioned above are instruction fetching (IF), instruction
decoding (ID), operand fetch (OF), and instruction execution (IE).
Each instruction of the program is fetched from the main memory, it is decoded
by the processor, and the operands mentioned in the instructions that are
required for the execution of instruction are fetched from the main memory,
and at last, the instruction is executed.
The figure below shows that to execute three instructions completely it takes 12 secs.
And if we implement pipelining then instruction will be executed in an overlapped fashion.
As you can see that the four instructions can be executed in 7 sec.
Pipeline computer synchronizes operations of all stages under a common clock. Hence, we
can say that executing instructions in pipelined fashion is more efficient.
ARRAY PROCESSORS
Array processors leverage parallel computing by implementing multiple arithmetic
logic units i.e. processing elements in a synchronized way.
The processing elements operate in a parallel fashion.
Now consider that if we replicate the ALUs and all ALUs are working in parallel then
we can say that we have achieved spatial parallelism.
Array processors are also capable of processing array elements.
In the figure above, we can see that we have multiple ALUs which are connected in
parallel with the control unit using a data routing network.
Each ALU in the system i.e., processing elements (PE) consists of the processor(P)
and local memory(M).
The pattern in which the processing elements are interconnected depends on the
specific computation to be performed by the control unit.
The scalar instructions are implicitly executed in the control unit whereas the vector
instructions are broadcasted to the parallelly connected processing elements.
The operands are fetched directly from the local memories.
The instruction fetch and instruction decode is done by the control unit and in this
way the vector instructions are executed in a distributed manner.
However, the different array processors may be using different kinds of
interconnection networks for connecting the processing elements.
The array processors are somewhat complex as compared to the pipelined
processors.
MULTIPROCESSOR SYSTEMS
Multiprocessor system supports parallel computing by using a set of interactive processors
that have shared resources.
In a multiprocessor system, there are multiple processors and processors that have access
to a common set of memory modules, peripheral devices, and other input-output devices.
However, the entire system with multiple processors is controlled by a single operating
system.
It is the responsibility of the operating system to provide interaction between the multiple
processors present in the system.
Even though the processors share memories, peripheral devices, and I/O, each processor in
the multiprocessor system has a local memory and even some private devices.
Communication between the processors can be achieved with the help of shared
memories or through the interrupted network.
The interconnection between the shared memories, I/O devices, and multiple
processors in the system can be determined in three different ways:
•Time shared common bus
•Crossbar switch network
•Multiport memories
Using a multiprocessor system improves the throughput, flexibility, availability, and
reliability of the system.
UNIT 5
ARCHITECTURAL CLASSIFICATION SCHEMES
FLYNN’S CLASSIFICATION
TAXONOMY
Flynn's taxonomy distinguishes multi-processor computer architectures
according to how they can be classified along the two independent
dimensions of Instruction Stream and Data Stream.
Each of these dimensions can have only one of two possible states: Single
or Multiple.
The matrix below defines the 4 possible classifications according to Flynn:
SISD- (SINGLE INSTRUCTION
SINGLE DATA)
A serial (non-parallel) computer
Single Instruction: Only one instruction stream is being acted on by
the CPU during any one clock cycle
Single Data: Only one data stream is being used as input during any
one clock cycle
Deterministic execution
This is the oldest type of computer
Examples: older generation mainframes, minicomputers, workstations
and single processor/core PCs.
SINGLE INSTRUCTION,
MULTIPLE DATA (SIMD):
A type of parallel computer
Single Instruction: All processing units execute the same instruction
at any given clock cycle
Multiple Data: Each processing unit can operate on a different data
element
Best suited for specialized problems characterized by a high degree
of regularity, such as graphics/image processing.
Synchronous (lockstep) and deterministic execution
Two varieties: Processor Arrays and Vector Pipelines
Examples:
Processor Arrays: Thinking Machines CM-2, MasPar MP-1 & MP-2,
ILLIAC IV
Vector Pipelines: IBM 9000, Cray X-MP, Y-MP & C90, Fujitsu VP, NEC
SX-2, Hitachi S820, ETA10
Most modern computers, particularly those with graphics processor
units (GPUs) employ SIMD instructions and execution units.
MULTIPLE INSTRUCTION,
MULTIPLE DATA (MIMD):
A type of parallel computer
Multiple Instruction: Every processor may be executing a different
instruction stream
Multiple Data: Every processor may be working with a different data
stream
Currently, the most common type of parallel computer - most
modern supercomputers fall into this category.
Examples: most current supercomputers, networked parallel
computer clusters and "grids", multi-processor SMP computers,
multi-core PCs.
Note: many MIMD architectures also include SIMD execution sub-
components
MULTIPLE INSTRUCTION,
SINGLE DATA (MISD):
A type of parallel computer
Multiple Instruction: Each processing unit operates on the data
independently via separate instruction streams.
Single Data: A single data stream is fed into multiple processing
units.
Few (if any) actual examples of this class of parallel computer have
ever existed.
Some conceivable uses might be:
•multiple frequency filters operating on a single signal stream
•multiple cryptography algorithms attempting to crack a single coded
message.
THANKYOU
SERIAL VS PARALLEL PROCESSING
Degree of parellelism is another criteria for computer architecture
classification
The max number of binary digits(bits) that can be processed within
unit time by a computer system is called maximum parallelism
degree P
A bit slice is a string of bits one from each of the words at the same
vertical position.
under above classification
Word Serial and Bit Serial (WSBS)
Word Parallel and Bit Serial (WPBS)
Word Serial and Bit Parallel(WSBP)
Word Parallel and Bit Parallel (WPBP)
WSBS has been called bit parallel processing because one bit is
processed at a time.
WPBS has been called bit slice processing because m-bit slice is
processes at a time.
WSBP is found in most existing computers and has been called as
Word Slice processing because one word of n bit processed at a time.
WPBP is known as fully parallel processing in which an array on n x m
bits is processes at one time
PARALLEL PROCESSING TECHNIQUES UNIPROCESSOR SYSTEMS

More Related Content

What's hot

Cisc vs risc
Cisc vs riscCisc vs risc
Cisc vs riscKumar
 
Ch9: Memory Management
Ch9: Memory ManagementCh9: Memory Management
Ch9: Memory ManagementAhmar Hashmi
 
Multi core-architecture
Multi core-architectureMulti core-architecture
Multi core-architecturePiyush Mittal
 
Ch4 memory management
Ch4 memory managementCh4 memory management
Ch4 memory managementBullz Musetsho
 
Operating System BCA 301
Operating System BCA 301Operating System BCA 301
Operating System BCA 301cpjcollege
 
Storage Management
Storage ManagementStorage Management
Storage ManagementSURBHI SAROHA
 
Operating system concepts (notes)
Operating system concepts (notes)Operating system concepts (notes)
Operating system concepts (notes)Sohaib Danish
 
Parallel Programming
Parallel ProgrammingParallel Programming
Parallel ProgrammingUday Sharma
 
Unit 3-pipelining & vector processing
Unit 3-pipelining & vector processingUnit 3-pipelining & vector processing
Unit 3-pipelining & vector processingvishal choudhary
 
Parallel computing persentation
Parallel computing persentationParallel computing persentation
Parallel computing persentationVIKAS SINGH BHADOURIA
 
Troubleshooting methods of computer peripherals
Troubleshooting methods of computer peripheralsTroubleshooting methods of computer peripherals
Troubleshooting methods of computer peripheralsDevyani Chaudhari
 
01 introduction to microcomputers
01 introduction to microcomputers01 introduction to microcomputers
01 introduction to microcomputersayeshasafdar8
 
Computer architecture
Computer architectureComputer architecture
Computer architectureZuhaib Zaroon
 
Direct access memory
Direct access memoryDirect access memory
Direct access memorymaliksiddique1
 
Distributed Operating System
Distributed Operating SystemDistributed Operating System
Distributed Operating SystemSanthiNivas
 
Operating System.pdf
Operating System.pdfOperating System.pdf
Operating System.pdfSyed Zaid Irshad
 

What's hot (20)

Computer architecture
Computer architecture Computer architecture
Computer architecture
 
Cisc vs risc
Cisc vs riscCisc vs risc
Cisc vs risc
 
Ch9: Memory Management
Ch9: Memory ManagementCh9: Memory Management
Ch9: Memory Management
 
Multi core-architecture
Multi core-architectureMulti core-architecture
Multi core-architecture
 
Ch4 memory management
Ch4 memory managementCh4 memory management
Ch4 memory management
 
Operating System BCA 301
Operating System BCA 301Operating System BCA 301
Operating System BCA 301
 
Storage Management
Storage ManagementStorage Management
Storage Management
 
Operating system concepts (notes)
Operating system concepts (notes)Operating system concepts (notes)
Operating system concepts (notes)
 
Parallel Programming
Parallel ProgrammingParallel Programming
Parallel Programming
 
Unit 3-pipelining & vector processing
Unit 3-pipelining & vector processingUnit 3-pipelining & vector processing
Unit 3-pipelining & vector processing
 
Parallelism
ParallelismParallelism
Parallelism
 
Parallel computing persentation
Parallel computing persentationParallel computing persentation
Parallel computing persentation
 
MEMORY & I/O SYSTEMS
MEMORY & I/O SYSTEMS                          MEMORY & I/O SYSTEMS
MEMORY & I/O SYSTEMS
 
Troubleshooting methods of computer peripherals
Troubleshooting methods of computer peripheralsTroubleshooting methods of computer peripherals
Troubleshooting methods of computer peripherals
 
01 introduction to microcomputers
01 introduction to microcomputers01 introduction to microcomputers
01 introduction to microcomputers
 
Lecture 2 process
Lecture 2   processLecture 2   process
Lecture 2 process
 
Computer architecture
Computer architectureComputer architecture
Computer architecture
 
Direct access memory
Direct access memoryDirect access memory
Direct access memory
 
Distributed Operating System
Distributed Operating SystemDistributed Operating System
Distributed Operating System
 
Operating System.pdf
Operating System.pdfOperating System.pdf
Operating System.pdf
 

Similar to PARALLEL PROCESSING TECHNIQUES UNIPROCESSOR SYSTEMS

Parallel Processing Presentation2
Parallel Processing Presentation2Parallel Processing Presentation2
Parallel Processing Presentation2daniyalqureshi712
 
Advanced computer architecture
Advanced computer architectureAdvanced computer architecture
Advanced computer architectureAjithaSomasundaram
 
Operating system (BCS303) MODULE 1 NOTES
Operating system (BCS303) MODULE 1 NOTESOperating system (BCS303) MODULE 1 NOTES
Operating system (BCS303) MODULE 1 NOTESKopinathMURUGESAN
 
OS - Ch1
OS - Ch1OS - Ch1
OS - Ch1sphs
 
Chapter 1 - Introduction
Chapter 1 - IntroductionChapter 1 - Introduction
Chapter 1 - IntroductionWayne Jones Jnr
 
Operating systems. replace ch1 with numbers for next chapters
Operating systems. replace ch1 with numbers for next chaptersOperating systems. replace ch1 with numbers for next chapters
Operating systems. replace ch1 with numbers for next chapterssphs
 
operating system over view.ppt operating sysyems
operating system over view.ppt operating sysyemsoperating system over view.ppt operating sysyems
operating system over view.ppt operating sysyemsJyoReddy9
 
What is operating system
What is operating systemWhat is operating system
What is operating systemSuvithak
 
Processor management
Processor managementProcessor management
Processor managementdev3993
 
Unix operating system basics
Unix operating system basicsUnix operating system basics
Unix operating system basicsSankar Suriya
 
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTES
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTESPARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTES
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTESsuthi
 
Lecture_01 Operating System Course Introduction
Lecture_01 Operating System Course IntroductionLecture_01 Operating System Course Introduction
Lecture_01 Operating System Course IntroductionArnoyKhan
 
LM1 - Computer System Overview, system calls
LM1 - Computer System Overview, system callsLM1 - Computer System Overview, system calls
LM1 - Computer System Overview, system callsmanideepakc
 
Operating system
Operating systemOperating system
Operating systemmak120
 
Basics of Operating System
Basics of Operating SystemBasics of Operating System
Basics of Operating SystemSoumit Ghosh
 

Similar to PARALLEL PROCESSING TECHNIQUES UNIPROCESSOR SYSTEMS (20)

Parallel Processing Presentation2
Parallel Processing Presentation2Parallel Processing Presentation2
Parallel Processing Presentation2
 
Advanced computer architecture
Advanced computer architectureAdvanced computer architecture
Advanced computer architecture
 
Ch1
Ch1Ch1
Ch1
 
Operating system (BCS303) MODULE 1 NOTES
Operating system (BCS303) MODULE 1 NOTESOperating system (BCS303) MODULE 1 NOTES
Operating system (BCS303) MODULE 1 NOTES
 
OS - Ch1
OS - Ch1OS - Ch1
OS - Ch1
 
Chapter 1 - Introduction
Chapter 1 - IntroductionChapter 1 - Introduction
Chapter 1 - Introduction
 
Operating systems. replace ch1 with numbers for next chapters
Operating systems. replace ch1 with numbers for next chaptersOperating systems. replace ch1 with numbers for next chapters
Operating systems. replace ch1 with numbers for next chapters
 
operating system over view.ppt operating sysyems
operating system over view.ppt operating sysyemsoperating system over view.ppt operating sysyems
operating system over view.ppt operating sysyems
 
What is operating system
What is operating systemWhat is operating system
What is operating system
 
Processor management
Processor managementProcessor management
Processor management
 
CH01.pdf
CH01.pdfCH01.pdf
CH01.pdf
 
Unix operating system basics
Unix operating system basicsUnix operating system basics
Unix operating system basics
 
Chapter 10
Chapter 10Chapter 10
Chapter 10
 
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTES
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTESPARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTES
PARALLEL ARCHITECTURE AND COMPUTING - SHORT NOTES
 
Lecture_01 Operating System Course Introduction
Lecture_01 Operating System Course IntroductionLecture_01 Operating System Course Introduction
Lecture_01 Operating System Course Introduction
 
100-E
100-E100-E
100-E
 
LM1 - Computer System Overview, system calls
LM1 - Computer System Overview, system callsLM1 - Computer System Overview, system calls
LM1 - Computer System Overview, system calls
 
Operating system
Operating systemOperating system
Operating system
 
Basics of Operating System
Basics of Operating SystemBasics of Operating System
Basics of Operating System
 
Cs8493 unit 1
Cs8493 unit 1Cs8493 unit 1
Cs8493 unit 1
 

Recently uploaded

Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfUmakantAnnand
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTiammrhaywood
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionSafetyChain Software
 

Recently uploaded (20)

Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
Concept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.CompdfConcept of Vouching. B.Com(Hons) /B.Compdf
Concept of Vouching. B.Com(Hons) /B.Compdf
 
Staff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSDStaff of Color (SOC) Retention Efforts DDSD
Staff of Color (SOC) Retention Efforts DDSD
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPTECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
ECONOMIC CONTEXT - LONG FORM TV DRAMA - PPT
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 
Mastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory InspectionMastering the Unannounced Regulatory Inspection
Mastering the Unannounced Regulatory Inspection
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
CĂłdigo Creativo y Arte de Software | Unidad 1
CĂłdigo Creativo y Arte de Software | Unidad 1CĂłdigo Creativo y Arte de Software | Unidad 1
CĂłdigo Creativo y Arte de Software | Unidad 1
 

PARALLEL PROCESSING TECHNIQUES UNIPROCESSOR SYSTEMS

  • 1. PARALLEL PROCESSING Prepared By Ms. Sheethal Aji Mani Assistant Professor Kristu Jayanti College
  • 2. Parallel processing is an efficient form of information processing which emphasizes the exploitation of concurrent events in the computing process. Parallel Processing demands concurrent execution of many programs in the computer. Concurrency implies parallelism, simultaneity, and pipelining.
  • 3. INTRODUCTION TO PARALLELISM IN UNIPROCESSOR SYSTEM From an operating point of view, computer systems have improved chronologically in four phases: • batch processing • multiprogramming • time sharing •multiprocessing In these four operating modes, the degree of parallelism increase sharply from phase to phase.
  • 4. We define parallel processing as Parallel processing is an efficient form of information processing which emphasizes the exploitation of concurrent events in the computing process. Parallel Processing demands concurrent execution of many programs in the computer. Concurrency implies parallelism, simultaneity, and pipelining. The highest level of parallel processing is conducted among multiple jobs or programs through multiprogramming, time sharing, and multiprocessing. Parallel processing can be challenged in four programmatic levels: 1.Job or program level – highest level of parallel processing conducted among multiple jobs or programs 2. Task or procedure level – next higher level parallel processing conducted among procedures or tasks within the same program 3. Interinstruction level- Third level to exploit concurrency among multiple instructions 4. Intrainstruction level-Finally concurrent operations within each instruction
  • 5. BASIC UNIPROCESSOR ARCHITECTURE •A typical uniprocessor computer consists of the three major components the main memory, the CPU (Central Processing Unit) and the I/O(Input-Output) subsystem. • There are two architectures of commercially available uniprocessor computers to show the relation between three subsystems
  • 6. SYSTEM ARCHITECTURE OF THE SUPERMINI VAX-11/780 UNIPROCESSOR SYSTEM
  • 7. There are 16, 32-bit general purpose register one of which is a Program Counter (PC). There is also a special CPU status register containing about the current state of the processor being executed The CPU contains an ALU with an optional Floating-point accelerator, and some local cache memory with an optional diagnostic memory Floating-point accelerator (FPA) -A device to improve the overall performance of a computer by removing the burden of performing floating-point arithmetic from the central processor. optional diagnostic memory –checks the errors
  • 8. •The CPU can be intervened by the operator through the console connected to floppy disk • The CPU, the main memory( 2^32 words of 32 bit each) and the I/O subsystem are all connected to a common bus, the synchronous backplane interconnection(SBI) • Through this bus, all I/O devices can communicate with each other with CPU or with the memory •I/O devices can be connected directly to the SBI through the unibus and its controller or through a mass bus and its controller
  • 9. SYSTEM ARCHITECTURE OF THE MAINFRAME IBM SYSTEM 370/MODEL 168 UNIPROCESSOR COMPUTER
  • 10. • •The CPU contains the instruction decoding and execution units as well as cache •Main memory is divided into four units referred to as logical storage units (LSU) that are four ways interleaved •The storage controller provides multiport Connections between the CPU and the four LSU’s • Peripherals are connected to the system via high speed I/O channels which operate asynchronously with the CPU
  • 11. Parallelism can be promoted by Hardware means Software means
  • 12. PARALLELLISM IN UNIPROCESSOR SYSTEM A number of parallel processing mechanisms have been developed in uniprocessor computers. We identify them in the following six categories: 1.Multiplicity of functional units 2.Parallelism and pipelining within the CPU 3.Overlapped CPU and I/O operations 4.Use of a hierarchical memory system 5.Balancing of subsystem bandwidths 6.Multiprogramming and time sharing
  • 13. 1.MULTIPLICITY OF FUNCTIONAL UNITS Early computers - 1. one ALU that perform one operation at a time 2. Slow processing Multiple and specialized functional units. - operate in parallel. - two parallel execution units (fixed and floating point arithmetic) in CDC- 6600 → 10 functional units
  • 14.
  • 15. 2.PARALLELISM AND PIPELINING WITHIN THE CPU Instead of serial bit adders , parallel adders are used in almost ALUs Use of High speed multiplier recoding and convergence division Sharing of hardware resources for functions of multiply and divide Various phases of instructions executions (instruction fetch , decode, operand fetch, arithmetic logic execution, store result) are pipelined For overlapped instruction executions, instruction prefetch and buffering techniques have been developed.
  • 16. 3.OVERLAPPED CPU AND I/O OPERATIONS I/O operations performed simultaneously with CPU computations - by using separate I/O controllers, channels or I/O processors. Dat DMA channel can be used for direct transfer of data between main memory and I/O devices
  • 17. USE OF A HIERARCHICAL MEMORY SYSTEM CPU is 1000 times faster than memory access Heirarchical memory system can be used to close up the speed gap Cache memory can be used to be served as a buffer between cpu and main memory
  • 18. BALANCING OF SUBSYSTEM BANDWIDTHS CPU is the fastest unit in computer. The bandwidth of a system is defined as the number of operations performed per unit time. In case of main memory the memory bandwidth is measured by the number of words that can be accessed per unit time. Relationship b/w BWs
  • 19. Bandwidth Balancing Between CPU and Memory The speed gap between the CPU and the main memory can be closed up by using fast cache memory between them. A block of memory words is moved from the main memory into the cache so that immediate instructions can be available most of the time from the cache. Bandwidth Balancing Between Memory and I/O Devices Input-output channels with different speeds can be used between the slow I/O devices and the main memory. The I/O channels perform buffering and multiplexing functions to transfer the data from multiple disks into the main memory by stealing cycles from the CPU.
  • 20. PARALLEL COMPUTER STRUCTURES A parallel computer is a kind of computer structure where a large complex problem is broken into multiple small problems which are then executed simultaneously by several processors. We often consider this as parallel processing. It is divided into three 1.Pipeline Computer 2.Array Processors 3.Multiprocessor Systems
  • 21. PIPELINE COMPUTER Pipeline computers leverage parallel computing by overlapping the execution of one process with another process. If we consider executing a small program that has a set of instructions then we can categorize the execution of the entire program in four steps that must be repeated continuously till the last instruction in the program gets executed. The four steps we mentioned above are instruction fetching (IF), instruction decoding (ID), operand fetch (OF), and instruction execution (IE). Each instruction of the program is fetched from the main memory, it is decoded by the processor, and the operands mentioned in the instructions that are required for the execution of instruction are fetched from the main memory, and at last, the instruction is executed.
  • 22. The figure below shows that to execute three instructions completely it takes 12 secs.
  • 23. And if we implement pipelining then instruction will be executed in an overlapped fashion. As you can see that the four instructions can be executed in 7 sec. Pipeline computer synchronizes operations of all stages under a common clock. Hence, we can say that executing instructions in pipelined fashion is more efficient.
  • 24. ARRAY PROCESSORS Array processors leverage parallel computing by implementing multiple arithmetic logic units i.e. processing elements in a synchronized way. The processing elements operate in a parallel fashion. Now consider that if we replicate the ALUs and all ALUs are working in parallel then we can say that we have achieved spatial parallelism. Array processors are also capable of processing array elements.
  • 25. In the figure above, we can see that we have multiple ALUs which are connected in parallel with the control unit using a data routing network. Each ALU in the system i.e., processing elements (PE) consists of the processor(P) and local memory(M). The pattern in which the processing elements are interconnected depends on the specific computation to be performed by the control unit.
  • 26. The scalar instructions are implicitly executed in the control unit whereas the vector instructions are broadcasted to the parallelly connected processing elements. The operands are fetched directly from the local memories. The instruction fetch and instruction decode is done by the control unit and in this way the vector instructions are executed in a distributed manner.
  • 27.
  • 28. However, the different array processors may be using different kinds of interconnection networks for connecting the processing elements. The array processors are somewhat complex as compared to the pipelined processors.
  • 29. MULTIPROCESSOR SYSTEMS Multiprocessor system supports parallel computing by using a set of interactive processors that have shared resources. In a multiprocessor system, there are multiple processors and processors that have access to a common set of memory modules, peripheral devices, and other input-output devices. However, the entire system with multiple processors is controlled by a single operating system. It is the responsibility of the operating system to provide interaction between the multiple processors present in the system. Even though the processors share memories, peripheral devices, and I/O, each processor in the multiprocessor system has a local memory and even some private devices.
  • 30. Communication between the processors can be achieved with the help of shared memories or through the interrupted network. The interconnection between the shared memories, I/O devices, and multiple processors in the system can be determined in three different ways: •Time shared common bus •Crossbar switch network •Multiport memories Using a multiprocessor system improves the throughput, flexibility, availability, and reliability of the system.
  • 31.
  • 33. FLYNN’S CLASSIFICATION TAXONOMY Flynn's taxonomy distinguishes multi-processor computer architectures according to how they can be classified along the two independent dimensions of Instruction Stream and Data Stream. Each of these dimensions can have only one of two possible states: Single or Multiple. The matrix below defines the 4 possible classifications according to Flynn:
  • 34. SISD- (SINGLE INSTRUCTION SINGLE DATA) A serial (non-parallel) computer Single Instruction: Only one instruction stream is being acted on by the CPU during any one clock cycle Single Data: Only one data stream is being used as input during any one clock cycle Deterministic execution This is the oldest type of computer Examples: older generation mainframes, minicomputers, workstations and single processor/core PCs.
  • 35.
  • 36. SINGLE INSTRUCTION, MULTIPLE DATA (SIMD): A type of parallel computer Single Instruction: All processing units execute the same instruction at any given clock cycle Multiple Data: Each processing unit can operate on a different data element Best suited for specialized problems characterized by a high degree of regularity, such as graphics/image processing.
  • 37. Synchronous (lockstep) and deterministic execution Two varieties: Processor Arrays and Vector Pipelines Examples: Processor Arrays: Thinking Machines CM-2, MasPar MP-1 & MP-2, ILLIAC IV Vector Pipelines: IBM 9000, Cray X-MP, Y-MP & C90, Fujitsu VP, NEC SX-2, Hitachi S820, ETA10 Most modern computers, particularly those with graphics processor units (GPUs) employ SIMD instructions and execution units.
  • 38.
  • 39. MULTIPLE INSTRUCTION, MULTIPLE DATA (MIMD): A type of parallel computer Multiple Instruction: Every processor may be executing a different instruction stream Multiple Data: Every processor may be working with a different data stream Currently, the most common type of parallel computer - most modern supercomputers fall into this category. Examples: most current supercomputers, networked parallel computer clusters and "grids", multi-processor SMP computers, multi-core PCs. Note: many MIMD architectures also include SIMD execution sub- components
  • 40.
  • 41. MULTIPLE INSTRUCTION, SINGLE DATA (MISD): A type of parallel computer Multiple Instruction: Each processing unit operates on the data independently via separate instruction streams. Single Data: A single data stream is fed into multiple processing units. Few (if any) actual examples of this class of parallel computer have ever existed. Some conceivable uses might be: •multiple frequency filters operating on a single signal stream •multiple cryptography algorithms attempting to crack a single coded message.
  • 42.
  • 44. SERIAL VS PARALLEL PROCESSING Degree of parellelism is another criteria for computer architecture classification The max number of binary digits(bits) that can be processed within unit time by a computer system is called maximum parallelism degree P A bit slice is a string of bits one from each of the words at the same vertical position.
  • 45. under above classification Word Serial and Bit Serial (WSBS) Word Parallel and Bit Serial (WPBS) Word Serial and Bit Parallel(WSBP) Word Parallel and Bit Parallel (WPBP)
  • 46. WSBS has been called bit parallel processing because one bit is processed at a time. WPBS has been called bit slice processing because m-bit slice is processes at a time. WSBP is found in most existing computers and has been called as Word Slice processing because one word of n bit processed at a time. WPBP is known as fully parallel processing in which an array on n x m bits is processes at one time