Network Virtualization in Windows Server 2012

  • 1,866 views
Uploaded on

 

More in: Technology
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Be the first to comment
No Downloads

Views

Total Views
1,866
On Slideshare
0
From Embeds
0
Number of Embeds
1

Actions

Shares
Downloads
0
Comments
0
Likes
1

Embeds 0

No embeds

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
    No notes for slide
  • We are to group into scenario:-Ensure network solution that deliver continues app availability. Ensure support on multi tenant such as support conflict domain / ip address Leverage on existing and new hardware to get the best performance networkingWe talk about the rich partner in the market that support networking
  • 1st ScenarioGoals isNo interruption to your servicesEnsure consistent bandwidth available to your services. Automation to remediate issue without user interventionFinally we want to make sure that we support different vendor in networking.
  • Next we are going to talk was DHCP Server. In the past, we always configure 80:20 rules. Mean 80% ip to 1 server and another 20% ip in another server.Let say if primary server down, you may not have enough ip to serve to your client. Therefore in Win2012. We provide HA DHCP Server by provide more than 1 DHCP ServerYou can setupA/A or A/PYou can have 1 DHCP in Primary site and 1 DHCP in Secondary Site.In case server, server down, the secondary DHCP will take over.If you setup A/A, both server will load balance itself to distribute IP. And cool is you do no need to setup clustering.
  • Move on QOS. Let say you use virtualization. 1 NIC share with 4 VM. Some VM coz problem.With Win2012, you can configure Max & Min Bandwidth.Min Bandwidth always ensure your services will always get adequate bandwidthThen Max Bandwidth setting, ensure that we configure the cap (limit) to avoid interruption to other VM.
  • In this scenario, we are looking into high performance networkingGoal:-No buffer. Slight buffer is not acceptable in the SLADeliver high speed data transfer
  • Example:- When we transfer data from VM to other VM. It will go from VM NIC to Hyper-V Virtual Switch to Physical NIC.With SR-IOV we open virtual function. The VM can talk direct to physical NIC bypass your Hyper-V Switch.This give you high performance as same as running a physical server.RequirementsTo take advantage of the new Hyper‑V features for host scale and scale-up workload support, you need the following:One or more Windows Server 2012 installations with the Hyper‑V role installed. Hyper‑V requires a server that provides processor support for hardware virtualization.The number of virtual processors that may be configured in a virtual machine depends on the number of processors on the physical machine. You must have at least as many logical processors in the virtualization host as the number of virtual processors required in the virtual machine. For example, to configure a virtual machine with the maximum of 32 virtual processors, you must be running Hyper‑V in Windows Server 2012 on a virtualization host that has 32 or more logical processors.SR-IOV networking requires the following:A host system that supports SR-IOV (such as Intel VT-d2), including chipset support for interrupt and DMA remapping and proper firmware support to enable and describe the platform’s SR-IOV capabilities to the operating system.An SR-IOV–capable network adapter and driver in both the management operating system (which runs the Hyper‑V role) and each virtual machine where a virtual function is assigned.
  • SMB Direct. Without attach FC attached storage, we can also achieve high performance storage.Example:- FS Client & FS ServerIt will go through different layer and Data break into smaller packet and transfer between S & C. All this consume a lot of CPU utilization.To solve this problem, with RDMA compliant network adapter, we can bypass network layer & give full performance capacity. It will bypass CPU & give it to do other thing.RDMA compatibility is currently limited to InfiniBand, iWARP, and RoCE (RDMA over Converged Ethernet)RDMA wont work in Win 2012 NIC Teaming. But MPIO and SMB Multi Channel are preferred.
  • Next scenarioOur goal isTo build script for automation Able to identify the resources use for chargebackMinimize bandwidth transfer in multi site environment.Able to centrally control and monitor DNS & DHCP in your organizationHow do we enable this?
  • Now let look into IP Address management. Previously we setup DHCP & DNS and we cannot centralized manage the IP. Especially when you have a multisite, it is difficult for you to monitor available IP.We want to address this by install IPAM which going to discover your entire DHCP & DNS in your multisite environment.You can also monitorAvailable IPUsed IPStatic IPConfigure scope for your DHCP
  • Next feature is Branch Cache. It allow to you to have centralize data in primary site and customer at remote site will be able to access the data.The data is cache in a dedicated server or client computer. This allow you to save more bandwidth in your environment with less data transferNo configuration in client machine. You just need to configure GPO. Branch cache is now integrate with data dedup, you will have saving of data transfer as well.
  • Final scenario:- Look at richer ecosystemGoal is:Work with various partnerWe built a foundation for partner to leverage on.With 2 goal, we want to have variety of choice for customer to choose.
  • Let move to Hyper-V Network Virtualization.Everyone know about Hyper-V which help you to do Server Virtualization. Yesterday @Singapore launching about Storage Virtualization.Today, we are going to have Network Virtualization.Here we got 2 different set of VM: Blue VM & Orange VM hosted in the single network (both is using the same IP Address and overlap in the same physical network)Here, we are the 1st to come out with network virtualization whereby each VM will have their own VM network without causing conflict . We can do this byIsolation traffic without VLAN – Keep customers on the same server but on different IP subnetsCross Subnet Migration – Move a VM to a new subnet without changing the IP address in the VMHow we do it:-NVGREIP Address rewrite
  • The 1st is we built Hyper-V Virtual Switch. Then we open our system and built the switch extensible to allow partner such as Cisco, InFlow, NEC ,5Nine, etc to build plug in and additional add-on.With this we add new functionality such as monitoring, forwarding, and filtering to the virtual switch.
  • We have talked about various features, next is who is the partner who provide and support these features.RDMABranch CacheHyper-Extensible SwitchNV

Transcript

  • 1. LAI YOONG SENG, MVP VIRTUAL MACHINEREDYNAMICS LAIYS@REDYNAMICS.COM
  • 2. Session Objective and Takeaways Session Objectives: Key Takeaways• Understand different scenario • Networking in Windows Server on the usage of Networking in 2012 is engineered for the Windows Server 2012 networking habits of highly• Demo successful clouds
  • 3. Customer Scenario Goals Dynamic memory Network Larger virtual Open, extensible increase assigned virtualization for machines to support switch to help to virtual machines multitenancy and IP increased workloads support security and to increase capacity portability management needs with no downtime
  • 4. NETWORKING SOLUTION THAT DELIVER CONTINUOUS APPLICATION AVAILABILITY AutomaticallyConsistent remediate issuesbandwidth for with no human DHCP Quality ofservices intervention failover Service
  • 5. IP Address Continuity With DHCPFailoverAutomaticallyreplicates and Inbox support forload balances failover withoutDHCP scopes clusterActive/Active orActive /Passive 5
  • 6. Network QOS Guarantee predictable Help enforce network customer SLAs performance & and maximum fair share pricing caps Support bandwidth floors & Bandwidth caps
  • 7. Network QOS DEMO 7
  • 8. HIGH PERFORMANCE NETWORKINGEnsure that High Speed data Single Rootnetworking is no transfer on I/Olonger the industry standard SMB Direct Virtualizationbottleneck hardware (SR-IOV) (RDMA)
  • 9. SR-IOV Support in Hyper-V Virtual machine Network stack Software NIC Virtual function (VF) Increase network Reduce network throughput latency Hyper-V Extensible Switch Reduces host CPU overhead for processing network traffic SR-IOV network adapter VF VF VF 9
  • 10. Improved Network Performance through SMBDirect (RDMA) Without RDMA With RDMAFile Client File Server App Higher performance through Buffer offloading of network I/O processing SMB SMB Buffer Buffer High OS OS Buffer Buffer Remote storage at throughput the speed of with low Driver Buffer Driver Buffer direct storage latency iWARP Adapter Adapter Buffer rNIC NIC rNIC NIC Buffer InfiniBand 10
  • 11. OPERATIONAL EFFICIENCY Manage IPSimplify datatransfer in address space & consolidate IP Addressmultisite Branch external tools Managementenvironments (IPAM) Cache
  • 12. IP Address Management (IPAM)Inbox feature forintegrated Integrate withmanagement of Microsoft DNSIP Address & DHCPCustom IPAddress space Audit Serverdisplay configuration 12
  • 13. Branch CacheCache content Improvefrom file & webservers on a performanceWAN in local & bandwidthbranch office savingScalability – Hostmultiple cacheservers in singlelocation
  • 14. MULTITENANT & RICHER ECOSYSTEM Provide anSimply infrastructurenetwork that we can Hyper-V Networkisolation build on VirtualizationProvide a wide range of choices to Hyper-Vmeet customer needs Extensible Hardware Switch Partner
  • 15. Hyper-V Network Virtualization Blue sees Orange sees SQL Server Web SQL Server WebIsolate traffic Cross Subnet 10.1.1.1 10.1.1.2 10.1.1.1 10.1.1.2without VLAN 10.1.1.1 192.168.1.10 10.1.1.1 192.168.1.11 Migration 10.1.1.2 192.168.2.12 10.1.1.2 192.168.2.13Server Virtualization Network Virtualization What’s really happening• Run multiple virtual 192.168.n.n • Run multiple virtual servers on a physical networks on a physical server PROVIDER ADDRESS SPACE (PA) network• Each virtual machine 192.168.1.10 192.168.1.11 192.168.2.12 192.168.2.13 • Each virtual network acts as though it is acts as though it is running as a physical running as physical 10.1.1.1 192.168.1.10 10.1.1.1 192.168.1.11 10.1.1.1 192.168.1.11 10.1.1.1 192.168.1.10 server 10.1.1.2 192.168.2.12 10.1.1.2 192.168.2.13 10.1.1.2 192.168.2.13 10.1.1.2 192.168.2.12 fabric 10.1.1.1 10.1.1.1 10.1.1.2 10.1.1.2 SQL Server SQL Server Web Web CUSTOMER ADDRESS SPACE
  • 16. Network Virtualization inHyper-V DEMO 16
  • 17. Extending the Hyper-V ExtensibleSwitchManageability• Windows PowerShell and scripting support Extensibility features Extension monitoring• Unified tracing and enhanced diagnostics Extension uniqueness Extensions that learn virtual machine life cycleBenefits Extensions that can veto state changes• Open platform to fuel plug-ins Multiple extensions on same switch• Free core services Integration with built-in features• Windows reliability/quality Ability to capture extensions• Unified management• Easier support• Live migration support
  • 18. 5Nine Extensible Switch DEMO 18
  • 19. Network Virtualization PartnersRDMA HYPER-V EXTENSIBLE SWITCH NETWORK VIRTUALIZATION Mellanox Intel Chelsio 5NINE inMon NEC TECHNOLOGY PARTNERS ARISTA DELLBRANCH CACHE Net Cisco HP Optics SILICON PARTNERS EMC IVO nAppliance Networks EMULEX Broadcom Intel Riverbed NeApp F5 GATEWAY PARTNERS IVO Networks nAppliance 19
  • 20. Let’s Recap • Resource Metering • D-VMQ • Registered-IO • DNSSEC • IPsec Task Offload • RDP WAN optimization • NIC Teaming • RSS • SNI • Hyper-V Replica • RSC • WebSocket • PVLAN Protocol • SMB Multichannel • LSO • Cross-premises • Direct Access • Loopback Fast Path connectivity • IP portability 20
  • 21. Virtual Lai’s Blog Q&A www.ms4u.info DOWNLOAD Windows MVUG User Group Server 2012 https://www.facebook.com/g http://technet.microsoft.com/en- roups/mvuggroup/ us/evalcenter/hh670538.aspx21