SlideShare a Scribd company logo
Introduction to NUMA (Non-Uniform Memory Access)
vote
This is a primer on the NUMA hardware architecture...
In a typical SMP (Symmetric MultiProcessor
architecture), all memory access are posted to the
same shared memory bus. This works fine for a
relatively small number of CPUs, but the problem with
the shared bus appears when you have dozens, even
hundreds, of CPUs competing for access to the shared
memory bus. This leads to a major performance
bottleneck due to the extremely high contention rate
between the multiple CPU's onto the single memory
bus.
The NUMA architecture was designed to surpass these
scalability limits of the SMP architecture. NUMA
computers offer the scalability of MPP(Massively
Parallel Processing), in that processors can be added
and removed at will without loss of efficiency, and the
programming ease of SMP where.
Understanding Non-uniform Memory Access
Updated: 5 December 2005
Microsoft SQL Server 2005 is non-uniform memory access (NUMA) aware, and performs
well on NUMA hardware without special configuration. As clock speed and the number of
processors increase, it becomes increasingly difficult to reduce the memory latency
required to use this additional processing power. To circumvent this, hardware vendors
provide large L3 caches, but this is only a limited solution. NUMA architecture provides a
scalable solution to this problem. SQL Server 2005 has been designed to take advantage of
NUMA-based computers without requiring any application changes.
NUMA Concepts
The trend in hardware has been towards more than one system bus, each serving a small
set of processors. Each group of processors has its own memory and possibly its own I/O
channels. However, each CPU can access memory associated with the other groups in a
coherent way. Each group is called a NUMA node. The number of CPUs within a NUMA node
depends on the hardware vendor. It is faster to access local memory than the memory
associated with other NUMA nodes. This is the reason for the name, non-uniform memory
access architecture.
On NUMA hardware, some regions of memory are on physically different buses from other
regions. Because NUMA uses local and foreign memory, it will take longer to access some
regions of memory than others. Local memory and foreign memory are typically used in
reference to a currently running thread. Local memory is the memory that is on the same
node as the CPU currently running the thread. Any memory that does not belong to the
node on which the thread is currently running is foreign. Foreign memory is also known as
remote memory. The ratio of the cost to access foreign memory over that for local memory
is called the NUMA ratio. If the NUMA ratio is 1, it is symmetric multiprocessing (SMP). The
greater the ratio, the more it costs to access the memory of other nodes. Windows
applications that are not NUMA aware (including SQL Server 2000 SP3 and earlier)
sometimes perform poorly on NUMA hardware.
The main benefit of NUMA is scalability. The NUMA architecture was designed to surpass
the scalability limits of the SMP architecture. With SMP, all memory access is posted to the
same shared memory bus. This works fine for a relatively small number of CPUs, but not
when you have dozens, even hundreds, of CPUs competing for access to the shared
memory bus. NUMA alleviates these bottlenecks by limiting the number of CPUs on any one
memory bus and connecting the various nodes by means of a high speed interconnection.
Hardware-NUMA vs. Soft-NUMA
NUMA can match memory with CPUs through specialized hardware (hardware NUMA) or by
configuring SQL Server memory (soft-NUMA). During startup, SQL Server configures itself
based on underlying operating system and hardware configuration or the soft-NUMA
setting. For both hardware and soft-NUMA, when SQL Server starts in a NUMA
configuration, the SQL Server log records a multimode configuration message for each
node, along with the CPU mask.
Hardware NUMA
Computers with hardware NUMA have more than one system bus, each serving a small set
of processors. Each group of processors has its own memory and possibly its own I/O
channels, but each CPU can access memory associated with other groups in a coherent
way. Each group is called a NUMA node. The number of CPUs within a NUMA node depends
on the hardware vendor. Your hardware manufacturer can tell you if your computer
supports hardware NUMA.
If you have hardware NUMA, it may be configured to use interleaved memory instead of
NUMA. In that case, Windows and therefore SQL Server will not recognize it as NUMA. Run
the following query to find the number of memory nodes available to SQL Server:
SELECT DISTINCT memory_node_id FROM sys.dm_os_memory_clerks
If SQL Server returns only a single memory node (node 0), either you do not have
hardware NUMA, or the hardware is configured as interleaved (non-NUMA). If you think
your hardware NUMA is configured incorrectly, contact your hardware vendor to enable
NUMA. SQL Server ignores NUMA configuration when hardware NUMA has four or less CPUs
and at least one node has only one CPU.
Soft-NUMA
SQL Server 2005 allows you to group CPUs into nodes referred to as soft-NUMA. You
usually configure soft-NUMA when you have many CPUs and do not have hardware NUMA,
but you can also use soft-NUMA to subdivide hardware NUMA nodes into smaller groups.
Only the SQL Server scheduler and SQL Server Network Interface (SNI) are soft-NUMA
aware. Memory nodes are created based on hardware NUMA and therefore not impacted by
soft-NUMA. So, for example, if you have an SMP computer with eight CPUs and you create
four soft-NUMA nodes with two CPUs each, you will only have one memory node serving all
four NUMA nodes. Soft-NUMA does not provide memory to CPU affinity.
The benefits of soft-NUMA include reducing I/O and lazy writer bottlenecks on computers
with many CPUs and no hardware NUMA. There is a single I/O thread and a single lazy
writer thread for each NUMA node. Depending on the usage of the database, these single
threads may be a significant performance bottleneck. Configuring four soft-NUMA nodes
provides four I/O threads and four lazy writer threads, which could increase performance.
You cannot create a soft-NUMA that includes CPUs from different hardware NUMA nodes.
For example, if your hardware has eight CPUs (0..7) and you have two hardware NUMA
nodes (0-3 and 4-7), you can create soft-NUMA by combining CPU(0,1) and CPU(2,3). You
cannot create soft-NUMA using CPU (1, 5), but you can use CPU affinity to affinitize an
instance of SQL Server to CPUs from different NUMA nodes. So in the previous example, if
SQL Server uses CPUs 0-3, you will have one I/O thread and one lazy writer thread. If, in
the previous example SQL Server uses CPUs 1, 2, 5, and 6, you will access two NUMA
nodes and have two I/O threads and two lazy writer threads.
Non-Uniform Memory Access
From Wikipedia, the free encyclopedia
Jump to: navigation, search
Non-Uniform Memory Access or Non-Uniform Memory Architecture (NUMA) is a
computer memory design used in multiprocessors, where the memory access time
depends on the memory location relative to a processor. Under NUMA, a processor can
access its own local memory faster than non-local memory, that is, memory local to
another processor or memory shared between processors.
NUMA architectures logically follow in scaling from symmetric multiprocessing (SMP)
architectures. Their commercial development came in work by Burroughs, Convex
Computer (later HP), SGI, Sequent and Data General during the 1990s. Techniques
developed by these companies later featured in a variety of Unix-like operating systems,
as well as to some degree in Windows NT and in later versions of Microsoft Windows.

More Related Content

What's hot

Lecture 6
Lecture  6Lecture  6
Lecture 6Mr SMAK
 
Lecture 6
Lecture  6Lecture  6
Lecture 6Mr SMAK
 
Parallel architecture
Parallel architectureParallel architecture
Parallel architectureMr SMAK
 
Multiple processor (ppt 2010)
Multiple processor (ppt 2010)Multiple processor (ppt 2010)
Multiple processor (ppt 2010)
Arth Ramada
 
Computer architecture
Computer architecture Computer architecture
Computer architecture Ashish Kumar
 
Parallel computing
Parallel computingParallel computing
Parallel computing
Vinay Gupta
 
Shared-Memory Multiprocessors
Shared-Memory MultiprocessorsShared-Memory Multiprocessors
Shared-Memory Multiprocessors
Salvatore La Bua
 
Multiprocessor architecture
Multiprocessor architectureMultiprocessor architecture
Multiprocessor architecture
Arpan Baishya
 
message passing vs shared memory
message passing vs shared memorymessage passing vs shared memory
message passing vs shared memory
Hamza Zahid
 
Lecture 6.1
Lecture  6.1Lecture  6.1
Lecture 6.1Mr SMAK
 
Parallel Processing Presentation2
Parallel Processing Presentation2Parallel Processing Presentation2
Parallel Processing Presentation2daniyalqureshi712
 
Multi Processors And Multi Computers
 Multi Processors And Multi Computers Multi Processors And Multi Computers
Multi Processors And Multi Computers
Nemwos
 
Parallel computing persentation
Parallel computing persentationParallel computing persentation
Parallel computing persentation
VIKAS SINGH BHADOURIA
 
Introduction to parallel_computing
Introduction to parallel_computingIntroduction to parallel_computing
Introduction to parallel_computing
Mehul Patel
 
Parallelism
ParallelismParallelism
Parallelism
Md Raseduzzaman
 
Parallel processing
Parallel processingParallel processing
Parallel processing
Shivalik college of engineering
 

What's hot (18)

Lecture 6
Lecture  6Lecture  6
Lecture 6
 
Lecture 6
Lecture  6Lecture  6
Lecture 6
 
Mimd
MimdMimd
Mimd
 
Parallel architecture
Parallel architectureParallel architecture
Parallel architecture
 
Multiple processor (ppt 2010)
Multiple processor (ppt 2010)Multiple processor (ppt 2010)
Multiple processor (ppt 2010)
 
Computer architecture
Computer architecture Computer architecture
Computer architecture
 
Parallel computing
Parallel computingParallel computing
Parallel computing
 
Shared-Memory Multiprocessors
Shared-Memory MultiprocessorsShared-Memory Multiprocessors
Shared-Memory Multiprocessors
 
Multiprocessor architecture
Multiprocessor architectureMultiprocessor architecture
Multiprocessor architecture
 
message passing vs shared memory
message passing vs shared memorymessage passing vs shared memory
message passing vs shared memory
 
Lecture 6.1
Lecture  6.1Lecture  6.1
Lecture 6.1
 
Parallel Processing Presentation2
Parallel Processing Presentation2Parallel Processing Presentation2
Parallel Processing Presentation2
 
Multi Processors And Multi Computers
 Multi Processors And Multi Computers Multi Processors And Multi Computers
Multi Processors And Multi Computers
 
Parallel processing extra
Parallel processing extraParallel processing extra
Parallel processing extra
 
Parallel computing persentation
Parallel computing persentationParallel computing persentation
Parallel computing persentation
 
Introduction to parallel_computing
Introduction to parallel_computingIntroduction to parallel_computing
Introduction to parallel_computing
 
Parallelism
ParallelismParallelism
Parallelism
 
Parallel processing
Parallel processingParallel processing
Parallel processing
 

Similar to Numa (non uniform memory access)

Overview on NUMA
Overview on NUMAOverview on NUMA
Overview on NUMA
Abed Maatalla
 
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
langonej
 
Distributed operating system amoeba case study
Distributed operating system  amoeba case studyDistributed operating system  amoeba case study
Distributed operating system amoeba case study
RamuAryan
 
New microsoft office word document
New microsoft office word documentNew microsoft office word document
New microsoft office word documentsandya veduri
 
Symmetric multiprocessing and Microkernel
Symmetric multiprocessing and MicrokernelSymmetric multiprocessing and Microkernel
Symmetric multiprocessing and Microkernel
Manoraj Pannerselum
 
Lecture 6
Lecture  6Lecture  6
Lecture 6Mr SMAK
 
Memory interleaving and superscalar processor
Memory interleaving and superscalar processorMemory interleaving and superscalar processor
Memory interleaving and superscalar processor
sshwetasrivastava
 
Ram and types of ram.Cache
Ram and types of ram.CacheRam and types of ram.Cache
Ram and types of ram.Cache
hamza mukhtiar
 
Parallelism in sql server
Parallelism in sql serverParallelism in sql server
Parallelism in sql server
Enrique Catala Bañuls
 
22CS201 COA
22CS201 COA22CS201 COA
22CS201 COA
Kathirvel Ayyaswamy
 
Shared memory Parallelism (NOTES)
Shared memory Parallelism (NOTES)Shared memory Parallelism (NOTES)
Shared memory Parallelism (NOTES)
Subhajit Sahu
 
Computer architecture multi processor
Computer architecture multi processorComputer architecture multi processor
Computer architecture multi processor
Mazin Alwaaly
 
Distributed Shared Memory
Distributed Shared MemoryDistributed Shared Memory
Distributed Shared Memory
Prakhar Rastogi
 
How To Gain Performance On Sql Server By Making Num Ayour Friend
How To Gain Performance On Sql Server By Making Num Ayour FriendHow To Gain Performance On Sql Server By Making Num Ayour Friend
How To Gain Performance On Sql Server By Making Num Ayour FriendThomas Grohser
 
6.distributed shared memory
6.distributed shared memory6.distributed shared memory
6.distributed shared memory
Gd Goenka University
 
Multicore and shared multi processor
Multicore and shared multi processorMulticore and shared multi processor
Multicore and shared multi processor
Sou Jana
 
Multicore and shared multi processor
Multicore and shared multi processorMulticore and shared multi processor
Multicore and shared multi processor
Janani S
 
Presentation1.pptx
Presentation1.pptxPresentation1.pptx
Presentation1.pptx
TheresaSKMansaray
 
CA UNIT IV.pptx
CA UNIT IV.pptxCA UNIT IV.pptx
CA UNIT IV.pptx
ssuser9dbd7e
 

Similar to Numa (non uniform memory access) (20)

Overview on NUMA
Overview on NUMAOverview on NUMA
Overview on NUMA
 
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
Federal VMUG - March - Virtual machine sizing considerations in a numa enviro...
 
Distributed operating system amoeba case study
Distributed operating system  amoeba case studyDistributed operating system  amoeba case study
Distributed operating system amoeba case study
 
New microsoft office word document
New microsoft office word documentNew microsoft office word document
New microsoft office word document
 
Symmetric multiprocessing and Microkernel
Symmetric multiprocessing and MicrokernelSymmetric multiprocessing and Microkernel
Symmetric multiprocessing and Microkernel
 
Lecture 6
Lecture  6Lecture  6
Lecture 6
 
Memory interleaving and superscalar processor
Memory interleaving and superscalar processorMemory interleaving and superscalar processor
Memory interleaving and superscalar processor
 
Ram and types of ram.Cache
Ram and types of ram.CacheRam and types of ram.Cache
Ram and types of ram.Cache
 
Parallelism in sql server
Parallelism in sql serverParallelism in sql server
Parallelism in sql server
 
22CS201 COA
22CS201 COA22CS201 COA
22CS201 COA
 
Shared memory Parallelism (NOTES)
Shared memory Parallelism (NOTES)Shared memory Parallelism (NOTES)
Shared memory Parallelism (NOTES)
 
Computer architecture multi processor
Computer architecture multi processorComputer architecture multi processor
Computer architecture multi processor
 
Distributed Shared Memory
Distributed Shared MemoryDistributed Shared Memory
Distributed Shared Memory
 
How To Gain Performance On Sql Server By Making Num Ayour Friend
How To Gain Performance On Sql Server By Making Num Ayour FriendHow To Gain Performance On Sql Server By Making Num Ayour Friend
How To Gain Performance On Sql Server By Making Num Ayour Friend
 
UNAS-20140123-1800
UNAS-20140123-1800UNAS-20140123-1800
UNAS-20140123-1800
 
6.distributed shared memory
6.distributed shared memory6.distributed shared memory
6.distributed shared memory
 
Multicore and shared multi processor
Multicore and shared multi processorMulticore and shared multi processor
Multicore and shared multi processor
 
Multicore and shared multi processor
Multicore and shared multi processorMulticore and shared multi processor
Multicore and shared multi processor
 
Presentation1.pptx
Presentation1.pptxPresentation1.pptx
Presentation1.pptx
 
CA UNIT IV.pptx
CA UNIT IV.pptxCA UNIT IV.pptx
CA UNIT IV.pptx
 

Recently uploaded

Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Soumen Santra
 
English lab ppt no titlespecENG PPTt.pdf
English lab ppt no titlespecENG PPTt.pdfEnglish lab ppt no titlespecENG PPTt.pdf
English lab ppt no titlespecENG PPTt.pdf
BrazilAccount1
 
Hierarchical Digital Twin of a Naval Power System
Hierarchical Digital Twin of a Naval Power SystemHierarchical Digital Twin of a Naval Power System
Hierarchical Digital Twin of a Naval Power System
Kerry Sado
 
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
Amil Baba Dawood bangali
 
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
thanhdowork
 
road safety engineering r s e unit 3.pdf
road safety engineering  r s e unit 3.pdfroad safety engineering  r s e unit 3.pdf
road safety engineering r s e unit 3.pdf
VENKATESHvenky89705
 
14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application
SyedAbiiAzazi1
 
ML for identifying fraud using open blockchain data.pptx
ML for identifying fraud using open blockchain data.pptxML for identifying fraud using open blockchain data.pptx
ML for identifying fraud using open blockchain data.pptx
Vijay Dialani, PhD
 
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
MdTanvirMahtab2
 
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
bakpo1
 
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdf
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdfGoverning Equations for Fundamental Aerodynamics_Anderson2010.pdf
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdf
WENKENLI1
 
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdfTutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
aqil azizi
 
weather web application report.pdf
weather web application report.pdfweather web application report.pdf
weather web application report.pdf
Pratik Pawar
 
Immunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary AttacksImmunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary Attacks
gerogepatton
 
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Dr.Costas Sachpazis
 
Unbalanced Three Phase Systems and circuits.pptx
Unbalanced Three Phase Systems and circuits.pptxUnbalanced Three Phase Systems and circuits.pptx
Unbalanced Three Phase Systems and circuits.pptx
ChristineTorrepenida1
 
Gen AI Study Jams _ For the GDSC Leads in India.pdf
Gen AI Study Jams _ For the GDSC Leads in India.pdfGen AI Study Jams _ For the GDSC Leads in India.pdf
Gen AI Study Jams _ For the GDSC Leads in India.pdf
gdsczhcet
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
obonagu
 
Fundamentals of Electric Drives and its applications.pptx
Fundamentals of Electric Drives and its applications.pptxFundamentals of Electric Drives and its applications.pptx
Fundamentals of Electric Drives and its applications.pptx
manasideore6
 
Nuclear Power Economics and Structuring 2024
Nuclear Power Economics and Structuring 2024Nuclear Power Economics and Structuring 2024
Nuclear Power Economics and Structuring 2024
Massimo Talia
 

Recently uploaded (20)

Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
 
English lab ppt no titlespecENG PPTt.pdf
English lab ppt no titlespecENG PPTt.pdfEnglish lab ppt no titlespecENG PPTt.pdf
English lab ppt no titlespecENG PPTt.pdf
 
Hierarchical Digital Twin of a Naval Power System
Hierarchical Digital Twin of a Naval Power SystemHierarchical Digital Twin of a Naval Power System
Hierarchical Digital Twin of a Naval Power System
 
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
 
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
 
road safety engineering r s e unit 3.pdf
road safety engineering  r s e unit 3.pdfroad safety engineering  r s e unit 3.pdf
road safety engineering r s e unit 3.pdf
 
14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application
 
ML for identifying fraud using open blockchain data.pptx
ML for identifying fraud using open blockchain data.pptxML for identifying fraud using open blockchain data.pptx
ML for identifying fraud using open blockchain data.pptx
 
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)
 
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
 
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdf
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdfGoverning Equations for Fundamental Aerodynamics_Anderson2010.pdf
Governing Equations for Fundamental Aerodynamics_Anderson2010.pdf
 
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdfTutorial for 16S rRNA Gene Analysis with QIIME2.pdf
Tutorial for 16S rRNA Gene Analysis with QIIME2.pdf
 
weather web application report.pdf
weather web application report.pdfweather web application report.pdf
weather web application report.pdf
 
Immunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary AttacksImmunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary Attacks
 
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
 
Unbalanced Three Phase Systems and circuits.pptx
Unbalanced Three Phase Systems and circuits.pptxUnbalanced Three Phase Systems and circuits.pptx
Unbalanced Three Phase Systems and circuits.pptx
 
Gen AI Study Jams _ For the GDSC Leads in India.pdf
Gen AI Study Jams _ For the GDSC Leads in India.pdfGen AI Study Jams _ For the GDSC Leads in India.pdf
Gen AI Study Jams _ For the GDSC Leads in India.pdf
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
 
Fundamentals of Electric Drives and its applications.pptx
Fundamentals of Electric Drives and its applications.pptxFundamentals of Electric Drives and its applications.pptx
Fundamentals of Electric Drives and its applications.pptx
 
Nuclear Power Economics and Structuring 2024
Nuclear Power Economics and Structuring 2024Nuclear Power Economics and Structuring 2024
Nuclear Power Economics and Structuring 2024
 

Numa (non uniform memory access)

  • 1. Introduction to NUMA (Non-Uniform Memory Access) vote This is a primer on the NUMA hardware architecture... In a typical SMP (Symmetric MultiProcessor architecture), all memory access are posted to the same shared memory bus. This works fine for a relatively small number of CPUs, but the problem with the shared bus appears when you have dozens, even hundreds, of CPUs competing for access to the shared memory bus. This leads to a major performance bottleneck due to the extremely high contention rate between the multiple CPU's onto the single memory bus. The NUMA architecture was designed to surpass these scalability limits of the SMP architecture. NUMA computers offer the scalability of MPP(Massively Parallel Processing), in that processors can be added and removed at will without loss of efficiency, and the programming ease of SMP where. Understanding Non-uniform Memory Access Updated: 5 December 2005 Microsoft SQL Server 2005 is non-uniform memory access (NUMA) aware, and performs well on NUMA hardware without special configuration. As clock speed and the number of processors increase, it becomes increasingly difficult to reduce the memory latency required to use this additional processing power. To circumvent this, hardware vendors provide large L3 caches, but this is only a limited solution. NUMA architecture provides a scalable solution to this problem. SQL Server 2005 has been designed to take advantage of NUMA-based computers without requiring any application changes. NUMA Concepts The trend in hardware has been towards more than one system bus, each serving a small set of processors. Each group of processors has its own memory and possibly its own I/O channels. However, each CPU can access memory associated with the other groups in a coherent way. Each group is called a NUMA node. The number of CPUs within a NUMA node depends on the hardware vendor. It is faster to access local memory than the memory associated with other NUMA nodes. This is the reason for the name, non-uniform memory access architecture. On NUMA hardware, some regions of memory are on physically different buses from other regions. Because NUMA uses local and foreign memory, it will take longer to access some regions of memory than others. Local memory and foreign memory are typically used in reference to a currently running thread. Local memory is the memory that is on the same
  • 2. node as the CPU currently running the thread. Any memory that does not belong to the node on which the thread is currently running is foreign. Foreign memory is also known as remote memory. The ratio of the cost to access foreign memory over that for local memory is called the NUMA ratio. If the NUMA ratio is 1, it is symmetric multiprocessing (SMP). The greater the ratio, the more it costs to access the memory of other nodes. Windows applications that are not NUMA aware (including SQL Server 2000 SP3 and earlier) sometimes perform poorly on NUMA hardware. The main benefit of NUMA is scalability. The NUMA architecture was designed to surpass the scalability limits of the SMP architecture. With SMP, all memory access is posted to the same shared memory bus. This works fine for a relatively small number of CPUs, but not when you have dozens, even hundreds, of CPUs competing for access to the shared memory bus. NUMA alleviates these bottlenecks by limiting the number of CPUs on any one memory bus and connecting the various nodes by means of a high speed interconnection. Hardware-NUMA vs. Soft-NUMA NUMA can match memory with CPUs through specialized hardware (hardware NUMA) or by configuring SQL Server memory (soft-NUMA). During startup, SQL Server configures itself based on underlying operating system and hardware configuration or the soft-NUMA setting. For both hardware and soft-NUMA, when SQL Server starts in a NUMA configuration, the SQL Server log records a multimode configuration message for each node, along with the CPU mask. Hardware NUMA Computers with hardware NUMA have more than one system bus, each serving a small set of processors. Each group of processors has its own memory and possibly its own I/O channels, but each CPU can access memory associated with other groups in a coherent way. Each group is called a NUMA node. The number of CPUs within a NUMA node depends on the hardware vendor. Your hardware manufacturer can tell you if your computer supports hardware NUMA. If you have hardware NUMA, it may be configured to use interleaved memory instead of NUMA. In that case, Windows and therefore SQL Server will not recognize it as NUMA. Run the following query to find the number of memory nodes available to SQL Server:
  • 3. SELECT DISTINCT memory_node_id FROM sys.dm_os_memory_clerks If SQL Server returns only a single memory node (node 0), either you do not have hardware NUMA, or the hardware is configured as interleaved (non-NUMA). If you think your hardware NUMA is configured incorrectly, contact your hardware vendor to enable NUMA. SQL Server ignores NUMA configuration when hardware NUMA has four or less CPUs and at least one node has only one CPU. Soft-NUMA SQL Server 2005 allows you to group CPUs into nodes referred to as soft-NUMA. You usually configure soft-NUMA when you have many CPUs and do not have hardware NUMA, but you can also use soft-NUMA to subdivide hardware NUMA nodes into smaller groups. Only the SQL Server scheduler and SQL Server Network Interface (SNI) are soft-NUMA aware. Memory nodes are created based on hardware NUMA and therefore not impacted by soft-NUMA. So, for example, if you have an SMP computer with eight CPUs and you create four soft-NUMA nodes with two CPUs each, you will only have one memory node serving all four NUMA nodes. Soft-NUMA does not provide memory to CPU affinity. The benefits of soft-NUMA include reducing I/O and lazy writer bottlenecks on computers with many CPUs and no hardware NUMA. There is a single I/O thread and a single lazy writer thread for each NUMA node. Depending on the usage of the database, these single threads may be a significant performance bottleneck. Configuring four soft-NUMA nodes provides four I/O threads and four lazy writer threads, which could increase performance. You cannot create a soft-NUMA that includes CPUs from different hardware NUMA nodes. For example, if your hardware has eight CPUs (0..7) and you have two hardware NUMA nodes (0-3 and 4-7), you can create soft-NUMA by combining CPU(0,1) and CPU(2,3). You cannot create soft-NUMA using CPU (1, 5), but you can use CPU affinity to affinitize an instance of SQL Server to CPUs from different NUMA nodes. So in the previous example, if SQL Server uses CPUs 0-3, you will have one I/O thread and one lazy writer thread. If, in the previous example SQL Server uses CPUs 1, 2, 5, and 6, you will access two NUMA nodes and have two I/O threads and two lazy writer threads. Non-Uniform Memory Access From Wikipedia, the free encyclopedia Jump to: navigation, search Non-Uniform Memory Access or Non-Uniform Memory Architecture (NUMA) is a computer memory design used in multiprocessors, where the memory access time depends on the memory location relative to a processor. Under NUMA, a processor can access its own local memory faster than non-local memory, that is, memory local to another processor or memory shared between processors.
  • 4. NUMA architectures logically follow in scaling from symmetric multiprocessing (SMP) architectures. Their commercial development came in work by Burroughs, Convex Computer (later HP), SGI, Sequent and Data General during the 1990s. Techniques developed by these companies later featured in a variety of Unix-like operating systems, as well as to some degree in Windows NT and in later versions of Microsoft Windows.