• Share
  • Email
  • Embed
  • Like
  • Save
  • Private Content
Virtual Design Master Challenge 1 - Joe
 

Virtual Design Master Challenge 1 - Joe

on

  • 728 views

The first Technology driven reality competition showcasing the incredible virtualization community members and their talents. Virtually Everywhere · virtualdesignmaster.com

The first Technology driven reality competition showcasing the incredible virtualization community members and their talents. Virtually Everywhere · virtualdesignmaster.com

Statistics

Views

Total Views
728
Views on SlideShare
428
Embed Views
300

Actions

Likes
0
Downloads
15
Comments
0

3 Embeds 300

http://www.virtualdesignmaster.com 297
https://twitter.com 2
http://cloud.feedly.com 1

Accessibility

Categories

Upload Details

Uploaded via as Microsoft Word

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

    Virtual Design Master Challenge 1 - Joe Virtual Design Master Challenge 1 - Joe Document Transcript

    • After the Outbreak: Rebuilding the World Approved:________________________________________ Date:_____________ Original Date: 08/04/13 Revision Date: 08/08/13 Written By: Joe Graziano Sr. Infrastructure Engineer W.R.O – World Rebuild Organization
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 1– Last printed 8/16/2013 5:35 PM Contents 1. INTRODUCTION ........................................................................................................................................................ 2 2. PROBLEM/OPPORTUNITY...................................................................................................................................... 2 3. PROJECT OBJECTIVES ........................................................................................................................................... 2 4. SCOPE........................................................................................................................................................................... 3 5. PREREQUISITES / ASSUMPTIONS ........................................................................................................................ 3 6. SYSTEM ARCHITECTURE AND ARCHITECTURE DESIGN ........................................................................... 4 7. SYSTEM DETAILED DESIGN.................................................................................................................................. 5 8. DIAGRAMS.................................................................................................................................................................. 8 9. ACRONYMS................................................................................................................................................................. 8 10. PEOPLE/RESOURCES............................................................................................................................................... 8
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 2– Last printed 8/16/2013 5:35 PM 1. Introduction 1.1. This document details the task of building up our world’s datacenters to reintroduce a working infrastructure to a land decimated by virus, undead and chaos. 3 datacenters have been located and are waiting for hardware, users and life to again echo through their halls. We will discuss creation of a Windows 2008 active directory domain and an Exchange 2013 email system. A SharePoint farm will be built to enable online document and data storage, collaboration among users and teams as well as the creation of an intranet and web presence for the organization. There will also be a VDI environment for local users as well as remote users with the help of a Cisco VPN client. 2. Problem/Opportunity 2.1. The world is in disarray after a virus outbreak that turned many into zombies. A wealthy philanthropist has recruited us to build a new infrastructure that will enable us to begin putting the world back together. The virus was stopped, but not before much of the world was overrun by zombies. The valiant survivors have taken care of the undead, and for now, we’ve won. A famous billionaire, Mr. Phil N. Thropist, has stepped in to help rebuild. His team of engineers has the Internet backup and running but many datacenters were destroyed. This is our opportunity as a massive warehouse was found containing almost limitless hardware, from April 2008, but it will be just what is needed to build out 3 datacenters. With hardware and software in hand we will tackle each datacenter. Server, network, telecommunication hardware will be racked. VMware clusters will be created. San’s with terabytes of disk will be provisioned and client VDI environments will be built to interface with critical software solutions like Exchange and SharePoint for quick and timely communication with other survivors around the world. This is our time to shine and bring life out of the chaos that has been our existence since the deadly virus nearly wiped out civilization as we know it. 3. Project Objectives 3.1. Use hardware found in the secured warehouse to create buildup 3 datacenters 3.2. Establish email connectivity so that teams can communicate/collaborate quickly and effectively 3.3. Create a document/data storage solution for teams to share and store all critical documents and data gathered 3.4. Focus on standardizing the datacenter design to be repeatable as new buildings/locations are found
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 3– Last printed 8/16/2013 5:35 PM 4. Scope 4.1. Design and implement 3 datacenters in the buildings salvaged form the disaster caused by the virus outbreak 4.2. Create a WAN infrastructure between the 3 locations and remote access so that teams can work to collaborate and rebuild our civilization 4.3. Establish Email system as well as data storageand collaboration for team to use as rebuild efforts begin 4.4. Utilize found hardware/software from a massive warehouse to build the datacenters 4.5. Document and normalize design to turn up new datacenters as they are found 5. Prerequisites / Assumptions 5.1. With the help of Mr. Thropist’s engineering team the internet is once again functional and we have access to all of the data and resources that were currently online at the time of the disaster. 5.2. While fixing ‘The Internet’ the engineering team also established WAN connectivity between datacenters at or near 100mbps. This will be critical in making sure that all 3 datacenters can communicate with each other and the remote users who will be scouring the area, country, world, in the hopes of finding more survivors and taking on the daunting task of rebuilding the world. 5.3. The datacenters that have been found were cleaned up and tested for power, HVAC both in and outside the datacenter room where the server and networking equipment will be housed. Network cabling has been run in the datacenter and server and networking racks are available as well.
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 4– Last printed 8/16/2013 5:35 PM 6. System Architecture and Architecture Design 6.1. The datacenters that we are building are to become the backbone of our new world’s infrastructure. Time is paramount so for now function will be a much bigger concern than form. But it is vitally important that all of the hardwareand software specifications laid out in this document be scalable to the needs of newly found buildings. These designs are going to be setup to be quickly repeatable as well. This way we can multiply or divide the data to accommodate much larger and much smaller spaces. 6.2. The Server and Storage infrastructure will consist of Dell Poweredge R905 servers with maximum ram, multiple core and socket processors to handle a large amount of Virtual Servers and VDI clients. The data will be saved on a Pillar Axiom 500 SAN that features multiple levels of redundancy. 6.3. The software solutions that will be implemented include an Exchange 2013 environment with redundant CAS/HA and Mailbox servers across the datacenters with a DAG configuration allowing the mailbox databases to be replicated in all datacenters for fault tolerance. A SharePoint environment will be created to allow for file storage and collaboration for functional teams and will be managed by a SQL 2012 back end. 6.4. A networking infrastructure will be created to handle traffic internal and across the WAN to the 3 datacenters. This will be accomplished with Cisco Pix Firewalls, ASR 1000 WAN routers, Nexus 7000 core switches, Catalyst IDF switches and a Cisco wireless system in each datacenter using Cisco 2100 series wireless controllers and Cisco Aironet 1250 access points to provide the WiFi signal needed. 6.5. A telecommunications infrastructure will also be created using Avaya S8500 media servers in each datacenter and Avaya 4610 VOIP phones for office and remote communication.
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 5– Last printed 8/16/2013 5:35 PM 7. System Detailed Design 7.1. Hardware Detailed Design 7.1.1. 3 Cisco ASR 1000 WAN routers will be used, one in each datacenter and will provide the connectivity for the WAN link between datacenters and internet traffic that remote users will be traveling on to get through to the client VPN tunnel. Every environment that will need WAN connectivity needs a good edge router and for it’s time the ASR 1000 series was one of the fastest and most productive edge routers you could find. 7.1.2. The Nexus 7000 switch was Cisco’s flagship switch in the early part of 2008. 3 of these switches will sit at the core of the network in each datacenter with dual supervisors for system redundancy and up to 8 more 48 port 1GB Ethernet blades. The switch will be the default gateway for each datacenter and will have VLANS for the Server, Client, VMotion, and Wireless networks. 7.1.3. With the office buildings not being super organized we are expecting seating to be a premium and everyone having access to network drops can’t be guaranteed. For this reason we will add a Cisco 2100 Wireless network controller in each datacenter and will place 6 Cisco Aironet 1250 access points on each floor throughout the buildings to accommodate all the users we’re expecting. Wireless will also allow a much larger group of users to be on site as well as provide more mobility for users with laptops and other Wi-Fi devices. 7.1.4. Using the Cisco PIX 515e Firewall the datacenters will be able to ensure that, even in a post-apocalyptic world access is secure. The firewall will also enable the engineers to create remote access VPN tunnels for clients who need to access the network from remote locations such as internet ready sites, like cyber cafes’ or from their laptops using the Cisco Any Connect VPN client. 7.1.5. Telephony will be accomplished with the Avaya S8500 media gateways that then connect to Avaya SCC controllers carrying the (XX cards) that are then connected to the PSTN. In the offices and remote we can use Avaya 4610SW IP phones as the Avaya can make use of SIP trunking and when a user establishes a VPN tunnel from their home DSL the phone will also be able to communicate. In addition to the S8500 the telecom rack will need an Avaya SCC Control Carrier cabinet that will house a TN464 DS1 card, a TN2602 Med Pro card, a TN2312 IPSI card and a TN799 CLAN card. These are components of the solution and are needed to make the connection to the PSTN as well as support the internal phone infrastructure.
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 6– Last printed 8/16/2013 5:35 PM 7.1.6. For the SAN in our datacenters we found the Axiom 500 from Pillar Data systems. Pillar was once an independent organization and today is part of Oracle. The Axiom 500 will provide connectivity using either fiber channel or ISCSI. Up to 64 bricks can be interconnected, easily allowing us access to the 300TB and maximum we have for this phase of the design. We will also setup SAN replication between the datacenters using the NSS (Network Storage Server) from Falconstor. This virtual appliance runs in each datacenter and allows the Pillar to replicate to the other Axioms and will be a key component in the BC/DR phase of our project. 7.1.7. The ESXi 5.1 hosts will run on Dell PowerEdge R905 Server. These servers boast 256GB Ram, 4x AMD Opteron 8000 processors with 6 cores per socket, expansion slots for ample network adapters and fiber cards for SAN connectivity. The Datacenter diagram will list the number of servers in each datacenter to accommodate the Server VM’s as well as the VDI environment. 7.1.8. We will place 4Catalyst 2960S10/100/1000 switches on each floor of the buildings which will utilize 1GB Ethernet uplinks to the Nexus 7000 in the datacenter. The network diagrams for each datacenter show the number of switches required for each floor of the building to support the maximum number of onsite users. 7.1.9. Dell M1530 laptopswill be deployed for users who have the need to leave the datacenters and bravely venture outside. In another life these were graphics design and gaming power house laptops they will be more than capable of connecting back to the datacenter using the Cisco Any Connect VPN client over the DSL networks available in users homes and remote locations. 7.1.10. The desktops found were Dell Optiplex 755 systems that will be configured with a local copy of Windows 7 professional as well as the Horizon View client which will give users’ access to a Virtual Desktop for central access and storage of data. The workstations also come equipped with a Broadcom wireless network adapter in the event that crowding occurs in the datacenter they are working in. 7.2. Software Detailed Design 7.2.1. Exchange 2013 will be used to create a fault tolerant email solution by building 5 VM servers at each datacenter. 2 servers will hold the CAS and HT role, one server will reside in the DMZ and have the edge transport role and the other 2 VM servers will have the Mailbox server role for the Exchange databases. The mailbox servers will also be part of a DAG and configured so that the 2 backup copies of all databases will be held offsite in the other two datacenters. This will prevent a single point of failure for any Exchange database. OWA will also be configured to allow remote teams and individuals to access their mail even if they find themselves at an internet available location but without their laptops.
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 7– Last printed 8/16/2013 5:35 PM 7.2.2. Windows 2008 R2 will be installed as the Server operating system for all the VM servers in the datacenters. There will be at least 2 domain controllers in each datacenter running DNS, DHCP, NTP, and all will host copies of the global catalog. 7.2.3. MS Office 2010 will be installed on the remote access laptops, desktops and VDI images so that users can access Outlook, create documents, spreadsheets and presentations when/if needed. 7.2.4. A SharePoint 2013 farm will be created to accomplish the design goal of having a central repository for documents and data created and needed by the support teams. An intranet page will also be built with specific access controls for teams to keep their data separate in an attempt to minimize overlap and confusion. The SharePoint farm will be supported on the back end by a SQL 2012 database cluster. 7.2.5. We will create a SQL 2012 database cluster for the SharePoint environment. This cluster will house the database(s) needed by SharePoint. Using a multisite failover cluster design we will enable failover of the SQL 2012 database(s) between the datacenters. 7.2.6. VMware ESXi 5.1 will be installed on all of the Dell PowerEdge R905 servers. DC1 will have 100 servers. DC2 will have 20 servers and DC3 will have 10 servers. Each datacenter will have a Server Cluster and a VDI cluster. This will help to keep resources separate so the server and client environment don’t interfere with each other’s performance. Local subnets have been identified for all of the datacenters and a routing will be configured and enabled for the datacenters. HA/DRS will be enabled but for now only for each local datacenter as the WAN link isn’t setup to handle a long distance VMotion scenario. 7.2.7. VMware Horizon View will be installed and configured so that we can install and configure Windows 7 workstation images and rapidly deploy up to 3000 virtual desktops to users as needed across the 3 datacenters. 7.2.8. A VMware VCenter Server 5.1 will be installed in each Datacenter and the 3 servers will be configured in linked mode so that each datacenter will be aware of its partners. While this is not a must for this phase of the project it will make phase 2 go a great deal smoother when BC/DR is needed. 7.2.9. The Cisco Any Connect VPN Client will be installed on all Laptops and any desktops that will leave the datacenters so that a remove client VPN can be established back to the datacenter network. This will enable teams outside the office the access they need to continue to work and collaborate.
    • Rebuilding the World D:ConFacconversion25327728rebuildingtheworld-130816193543-phpapp02.docx - 8– Last printed 8/16/2013 5:35 PM 8. Diagrams 8.1. Network Diagrams RebuildingTheWorld.vsdx 8.2. IP Spreadsheet RebuildingTheWorld-IPAddressScheme.xls 8.3. Hardware Manifest RebuildingTheWorld-HardwareManifest.xlsx 9. Acronyms 9.1. CAS – Client access Server (Exchange 2013) 9.2. HT – Hub Transport (Exchange 2013) 9.3. DAG – Database Availability Group (Exchange 2013) 9.4. BC – Business Continuity 9.5. DR – Disaster Recovery 9.6. DNS – Domain Name System 9.7. HA – High Availability 9.8. DRS – Distributed Resource Scheduler 9.9. PSTN – Public Switched Telephone Network 9.10. NSS – Network Storage Server (Falconstor) 10.People/Resources 10.1. Information Systems Project Lead – Joe Graziano 10.2. Technology Team – Datacenter 1 10.3. Technology Team – Datacenter 2 10.4. Technology Team – Datacenter 3 10.5. Technology Team – Lan/Wan 10.6. Financial/Owner – Mr. Phil N. Thropist