NBITS is a best hadoop training institute providing customer project-based Training and Placements in Big Data Hadoop. NBITS provides Hadoop Training in Hyderabad by Real time experts faculty with 10+ yrs Experience.
Video and slides synchronized, mp3 and slide download available at URL http://bit.ly/1awkL99.
Details on Pinterest's architeture, its systems -Pinball, Frontdoor-, and stack - MongoDB, Cassandra, Memcache, Redis, Flume, Kafka, EMR, Qubole, Redshift, Python, Java, Go, Nutcracker, Puppet, etc. Filmed at qconsf.com.
Yash Nelapati is an infrastructure engineer at Pinterest where he focusses on scalability, capacity planning and architecture. Prior to Pinterest he was into web development and rapidly prototyping UI. Marty Weiner joined Pinterest in early 2011 as the 2nd engineer. Previously worked at Azul Systems as a VM engineer focused on building/improving the JIT compilers in HotSpot.
Edge performance with in memory nosql, see how you can add high performance and scalability to your application. And try out some of the possible solutions: memcached, redis and aerospike
NBITS is a best hadoop training institute providing customer project-based Training and Placements in Big Data Hadoop. NBITS provides Hadoop Training in Hyderabad by Real time experts faculty with 10+ yrs Experience.
Video and slides synchronized, mp3 and slide download available at URL http://bit.ly/1awkL99.
Details on Pinterest's architeture, its systems -Pinball, Frontdoor-, and stack - MongoDB, Cassandra, Memcache, Redis, Flume, Kafka, EMR, Qubole, Redshift, Python, Java, Go, Nutcracker, Puppet, etc. Filmed at qconsf.com.
Yash Nelapati is an infrastructure engineer at Pinterest where he focusses on scalability, capacity planning and architecture. Prior to Pinterest he was into web development and rapidly prototyping UI. Marty Weiner joined Pinterest in early 2011 as the 2nd engineer. Previously worked at Azul Systems as a VM engineer focused on building/improving the JIT compilers in HotSpot.
Edge performance with in memory nosql, see how you can add high performance and scalability to your application. And try out some of the possible solutions: memcached, redis and aerospike
Redis Developers Day 2014 - Redis Labs TalksRedis Labs
These are the slides that the Redis Labs team had used to accompany the session that we gave during the first ever Redis Developers Day on October 2nd, 2014, London. It includes some of the ideas we've come up with to tackle operational challenges in the hyper-dense, multi-tenants Redis deployments that our service - Redis Cloud - consists of.
Rich storytelling with Drupal, Paragraphs and Islandora DAMSalxbrdg
Drupal's Paragraphs module, combined with a DAMS (Digital Asset Management System) can deliver powerful, rich stories on the web.
This session will show how, showing the inner workings of the Baseball Hall of Fame (baseballhall.org) website as a case study. This site uses Drupal with the Islandora DAMS to leverage the Baseball Hall of Fame's huge archive of images.
Topics covered:
- Building flexible content types using the Paragraphs module
- Multifaceted display of content using view modes
- DAMS & integrating Islandora assets with Drupal content
First presented at DrupalCamp Brighton in January 2015, by Alex Bridge and Tassos Koutlas.
This is the presentation at Percona Live 2015 on MySQL, MariaDB and Percona Orchestration on bare metal, virtualised environments and clouds (AWS and OpenStack).
How to migrate your existing MongoDB and Cassandra Apps to Azure Cosmos DBMicrosoft Tech Community
Bring your Mongo DB and Cassandra applications to Azure Cosmos DB and benefit turnkey global distribution, guaranteed low latency for cloud scale. Learn how easy it is to migrate your existing NoSQL applications to Azure Cosmos DB by using the MongoDB API and Cassandra API. In this sessions you will find out about the tools used in the process, see the code that leverages Azure Cosmos DB, and learn about techniques for working around the differences between the two products.
CBDW2014 - NoSQL Development With Couchbase and ColdFusion (CFML)Ortus Solutions, Corp
NoSQL document stores are reinventing the way we design our databases and cache layers. Couchbase server is a unique offering with unparalleled performance, automatic replication and failover. In this session, we'll talk about how to get started with Couchbase using the open source CFML SDK as well as native caching via the Railo Couchbase Extension.
This is my presentation on MySQL user camp on 26-06-2015.
It gives basic introduction to Ansible and how it can be benefited for MySQL deployment and configuration.
Discover content tokens, variant rules, data source advanced syntax and scaffolding. This deck was presented during SUGPL meeting at Cognifide office in Poznań on October 6th, 2017.
This presentation discusses Windows Azure Blob Storage, covering from the Windows Azure Storage Overview, Blob Storage Basic Concept, Blob Storage Advanced, and finally the Tip of the day.
Joanne Motta, of Destination British Columbia presented how Destination British Columbia built a passionate community of Australians who love BC.
About SoMeT = A new model for destination marketing. SoMeT is a program that gathers the top destination marketers from around the globe. SoMeT conferences take place three times a year in Europe, Australia, and United States.
Follow SoMeT on Twitter: https://twitter.com/SoMeTourism
Like SoMeT on Facebook: https://www.facebook.com/SoMeTourism
Learn about where the next SoMeT global conference is taking place http://www.sometourism.com/
Redis Developers Day 2014 - Redis Labs TalksRedis Labs
These are the slides that the Redis Labs team had used to accompany the session that we gave during the first ever Redis Developers Day on October 2nd, 2014, London. It includes some of the ideas we've come up with to tackle operational challenges in the hyper-dense, multi-tenants Redis deployments that our service - Redis Cloud - consists of.
Rich storytelling with Drupal, Paragraphs and Islandora DAMSalxbrdg
Drupal's Paragraphs module, combined with a DAMS (Digital Asset Management System) can deliver powerful, rich stories on the web.
This session will show how, showing the inner workings of the Baseball Hall of Fame (baseballhall.org) website as a case study. This site uses Drupal with the Islandora DAMS to leverage the Baseball Hall of Fame's huge archive of images.
Topics covered:
- Building flexible content types using the Paragraphs module
- Multifaceted display of content using view modes
- DAMS & integrating Islandora assets with Drupal content
First presented at DrupalCamp Brighton in January 2015, by Alex Bridge and Tassos Koutlas.
This is the presentation at Percona Live 2015 on MySQL, MariaDB and Percona Orchestration on bare metal, virtualised environments and clouds (AWS and OpenStack).
How to migrate your existing MongoDB and Cassandra Apps to Azure Cosmos DBMicrosoft Tech Community
Bring your Mongo DB and Cassandra applications to Azure Cosmos DB and benefit turnkey global distribution, guaranteed low latency for cloud scale. Learn how easy it is to migrate your existing NoSQL applications to Azure Cosmos DB by using the MongoDB API and Cassandra API. In this sessions you will find out about the tools used in the process, see the code that leverages Azure Cosmos DB, and learn about techniques for working around the differences between the two products.
CBDW2014 - NoSQL Development With Couchbase and ColdFusion (CFML)Ortus Solutions, Corp
NoSQL document stores are reinventing the way we design our databases and cache layers. Couchbase server is a unique offering with unparalleled performance, automatic replication and failover. In this session, we'll talk about how to get started with Couchbase using the open source CFML SDK as well as native caching via the Railo Couchbase Extension.
This is my presentation on MySQL user camp on 26-06-2015.
It gives basic introduction to Ansible and how it can be benefited for MySQL deployment and configuration.
Discover content tokens, variant rules, data source advanced syntax and scaffolding. This deck was presented during SUGPL meeting at Cognifide office in Poznań on October 6th, 2017.
This presentation discusses Windows Azure Blob Storage, covering from the Windows Azure Storage Overview, Blob Storage Basic Concept, Blob Storage Advanced, and finally the Tip of the day.
Joanne Motta, of Destination British Columbia presented how Destination British Columbia built a passionate community of Australians who love BC.
About SoMeT = A new model for destination marketing. SoMeT is a program that gathers the top destination marketers from around the globe. SoMeT conferences take place three times a year in Europe, Australia, and United States.
Follow SoMeT on Twitter: https://twitter.com/SoMeTourism
Like SoMeT on Facebook: https://www.facebook.com/SoMeTourism
Learn about where the next SoMeT global conference is taking place http://www.sometourism.com/
A Business Perspective on Building SharePoint 2013 Solutions on Windows AzureChristian Buckley
Presentation prepared for the San Francisco Azure User Group that walks through the latest features in SharePoint 2013, shares some insight into how Axceler is using Windows Azure to develop hybrid and cloud-based solutions, and outlines opportunities for developers to create Azure-based solutions for SharePoint 2013 and Office 365.
SharePoint 2013 on Azure: Your Dedicated Farm in the CloudJamie McAllister
With the Virtual Machine and Virtual Networking services of Windows Azure, it is now possible to deploy and operate a Microsoft SharePoint 2013 Server farm on Windows Azure. In this session we will discuss the key considerations, architecture and operations required to do this successfully. At the end you be able to build your own SharePoint Farm on the Cloud!
Webinar - Migrating Legacy On Premise Solutions to SharePoint Online and Wind...Eric Shupps
As many organizations move away from on-premise SharePoint deployments in favor of externally hosted solutions, developers will increasingly be called upon to make legacy applications compatible with cloud-based services. In this session we will explore the various options available for creating SharePoint applications in the cloud, discuss the pros and cons of each approach, delve into the technical aspects of cloud development, and provide real-world guidance for creating an on-premise-to-cloud migration strategy.
Move your on prem data to a lake in a Lake in CloudCAMMS
With the boom in data; the volume and its complexity, the trend is to move data to the cloud. Where and How do we do this? Azure gives you the answer. In this session, I will give you an introduction to Azure Data Lake and Azure Data Factory, and why they are good for the type of problem we are talking about. You will learn how large datasets can be stored on the cloud, and how you could transport your data to this store. The session will briefly cover Azure Data Lake as the modern warehouse for data on the cloud,
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...DoKC
The storage topology in vogue seems to cycle every few years. Internal storage is followed by centralized Storage Area Networks only to be superseded by one-size-fits-all Hyperconverged models - until scalability constraints led to distributed storage. Then comes NVMe, offering blistering speeds that all of these storage stacks struggle with. Kubernetes inspires Container Attached Storage aspiring to be the perfect model, so why is disaggregated storage now making an appearance?
This talk considers the motivations behind yet another storage topology and examines a modern, flexible architecture for delivering high-performance storage under Kubernetes.
This talk was given by Nick Connolly for DoK Day Europe @ KubeCon 2022.
Disaggregated Container Attached Storage - Yet Another Topology with What Pur...DoKC
Link: https://youtu.be/YhktX1W0geM
https://go.dok.community/slack
https://dok.community/
From the DoK Day EU 2022 (https://youtu.be/Xi-h4XNd5tE)
The storage topology in vogue seems to cycle every few years. Internal storage is followed by centralized Storage Area Networks only to be superseded by one-size-fits-all Hyperconverged models - until scalability constraints led to distributed storage. Then comes NVMe, offering blistering speeds that all of these storage stacks struggle with. Kubernetes inspires Container Attached Storage aspiring to be the perfect model, so why is disaggregated storage now making an appearance?
This talk considers the motivations behind yet another storage topology and examines a modern, flexible architecture for delivering high-performance storage under Kubernetes.
-----
Nick Connolly is a pioneer of storage virtualisation and the Chief Scientist at DataCore, where his background in real-time computing and multiprocessing led to the creation of a world-class high-performance storage stack on Windows. He holds patents ranging from highly scalable algorithms through to data protection techniques. Recently he has been working with OpenEBS to bring the power and performance of NVMe to Kubernetes.
Boost the Performance of SharePoint Today!Brian Culver
Is your farm struggling to server your organization? How long is it taking between page requests? Where is your bottleneck in your farm? Is your SQL Server tuned properly? Worried about upgrading due to poor performance? We will look at various tools for analyzing and measuring performance of your farm. We will look at simple SharePoint and IIS configuration options to instantly improve performance. I will discuss advanced approaches for analyzing, measuring and implementing optimizations in your farm as well as Performance Improvements in SharePoint 2013.
I3 - Running SharePoint 2016 in Azure the do's and dont's - Jasjit ChopraSPS Paris
In this session we are going to discuss the entire process of planning and successfully running SharePoint 2016 in the cloud. We will cover topics related to Security, Performance, Scalability, High Availability, Backup & Restore and Disaster Recovery. We will also cover the IaC (Infrastructure as Code) use case scenario in this deployment.
This presentation is for those of you who are interested in moving your on-prem SQL Server databases and servers to Azure virtual machines (VM’s) in the cloud so you can take advantage of all the benefits of being in the cloud. This is commonly referred to as a “lift and shift” as part of an Infrastructure-as-a-service (IaaS) solution. I will discuss the various Azure VM sizes and options, migration strategies, storage options, high availability (HA) and disaster recovery (DR) solutions, and best practices.
Storage Requirements and Options for Running Spark on KubernetesDataWorks Summit
In a world of serverless computing users tend to be frugal when it comes to expenditure on compute, storage and other resources. Paying for the same when they aren’t in use becomes a significant factor. Offering Spark as service on cloud presents very unique challenges. Running Spark on Kubernetes presents a lot of challenges especially around storage and persistence. Spark workloads have very unique requirements of Storage for intermediate data, long time persistence, Share file system and requirements become very tight when it same need to be offered as a service for enterprise to mange GDPR and other compliance like ISO 27001 and HIPAA certifications.
This talk covers challenges involved in providing Serverless Spark Clusters share the specific issues one can encounter when running large Kubernetes clusters in production especially covering the scenarios related to persistence.
This talk will help people using Kubernetes or docker runtime in production and help them understand various storage options available and which is more suitable for running Spark workloads on Kubernetes and what more can be done
Azure for IaaS - Global Windows Azure Bootcamp (GWAB)Loryan Strant
A presentation on Windows Azure as a platform for Infrastructure as a Service.
Presented in Melbourne, Australia as part of the Global Windows Azure Bootcamp in April 2013.
4Developers 2018: Przetwarzanie Big Data w oparciu o architekturę Lambda na p...PROIDEA
Według szacunków do 2020 roku wygenerujmy 40 Zetta byte’ów, a do roku 2025 aż 163 Zetta byte’ów różnego rodzaju danych, a ich dokładna analiza ACpozwali na odkrywanie nowych zjawisk, optymalizacje procesów, czy wspomaganie procesów decyzyjnych. Aby efektywnie przetwarzać tak duże zbiory danych potrzebujemy nowych technik analizy danych oraz innowacyjnych rozwiązań technologicznych. Ważną role pełni tutaj chmura Azure, która oferuje szereg usług, przy użyciu których możemy tworzyć rozwiązania na potrzeby przetwarzania Big Data zarówno w trybie batch’owych jak i ‘near real time’. Podczas sesji stworzymy przykładowe rozwiązanie przetwarzania Big Data oparte o architekturę Lambda , z wykorzystaniem usług platformy Azure, takich jak Azure Data Factory, Azure Stream Analytics, Azure HdInsight, Azure Event (IoT) Hub, czy Azure Data Lake.
Percona Live 4/14/15: Leveraging open stack cinder for peak application perfo...Tesora
In this session, speakers Amrith Kumar (Tesora), Steven Walchek (SolidFire), and Chris Merz (SolidFire) discuss Cinder, the OpenStack block storage service, and OpenStack Trove.
Amazon Web Services (AWS) can make hosting scalable, highly-available websites and web applications easier and less expensive for the Enterprise Education customers. Join us for an informative webinar on tools AWS provides to elastically scale your architecture to avoid underutilized resources while reducing complexity with templates, partners, and tools to do much of the heavy lifting of creating and running a website for you.
Listen to the keynote address and hear about the latest developments from Rachana Ananthakrishnan and Ian Foster who review the updates to the Globus Platform and Service, and the relevance of Globus to the scientific community as an automation platform to accelerate scientific discovery.
A Comprehensive Look at Generative AI in Retail App Testing.pdfkalichargn70th171
Traditional software testing methods are being challenged in retail, where customer expectations and technological advancements continually shape the landscape. Enter generative AI—a transformative subset of artificial intelligence technologies poised to revolutionize software testing.
Innovating Inference - Remote Triggering of Large Language Models on HPC Clus...Globus
Large Language Models (LLMs) are currently the center of attention in the tech world, particularly for their potential to advance research. In this presentation, we'll explore a straightforward and effective method for quickly initiating inference runs on supercomputers using the vLLM tool with Globus Compute, specifically on the Polaris system at ALCF. We'll begin by briefly discussing the popularity and applications of LLMs in various fields. Following this, we will introduce the vLLM tool, and explain how it integrates with Globus Compute to efficiently manage LLM operations on Polaris. Attendees will learn the practical aspects of setting up and remotely triggering LLMs from local machines, focusing on ease of use and efficiency. This talk is ideal for researchers and practitioners looking to leverage the power of LLMs in their work, offering a clear guide to harnessing supercomputing resources for quick and effective LLM inference.
Globus Compute wth IRI Workflows - GlobusWorld 2024Globus
As part of the DOE Integrated Research Infrastructure (IRI) program, NERSC at Lawrence Berkeley National Lab and ALCF at Argonne National Lab are working closely with General Atomics on accelerating the computing requirements of the DIII-D experiment. As part of the work the team is investigating ways to speedup the time to solution for many different parts of the DIII-D workflow including how they run jobs on HPC systems. One of these routes is looking at Globus Compute as a way to replace the current method for managing tasks and we describe a brief proof of concept showing how Globus Compute could help to schedule jobs and be a tool to connect compute at different facilities.
How to Position Your Globus Data Portal for Success Ten Good PracticesGlobus
Science gateways allow science and engineering communities to access shared data, software, computing services, and instruments. Science gateways have gained a lot of traction in the last twenty years, as evidenced by projects such as the Science Gateways Community Institute (SGCI) and the Center of Excellence on Science Gateways (SGX3) in the US, The Australian Research Data Commons (ARDC) and its platforms in Australia, and the projects around Virtual Research Environments in Europe. A few mature frameworks have evolved with their different strengths and foci and have been taken up by a larger community such as the Globus Data Portal, Hubzero, Tapis, and Galaxy. However, even when gateways are built on successful frameworks, they continue to face the challenges of ongoing maintenance costs and how to meet the ever-expanding needs of the community they serve with enhanced features. It is not uncommon that gateways with compelling use cases are nonetheless unable to get past the prototype phase and become a full production service, or if they do, they don't survive more than a couple of years. While there is no guaranteed pathway to success, it seems likely that for any gateway there is a need for a strong community and/or solid funding streams to create and sustain its success. With over twenty years of examples to draw from, this presentation goes into detail for ten factors common to successful and enduring gateways that effectively serve as best practices for any new or developing gateway.
Check out the webinar slides to learn more about how XfilesPro transforms Salesforce document management by leveraging its world-class applications. For more details, please connect with sales@xfilespro.com
If you want to watch the on-demand webinar, please click here: https://www.xfilespro.com/webinars/salesforce-document-management-2-0-smarter-faster-better/
Multiple Your Crypto Portfolio with the Innovative Features of Advanced Crypt...Hivelance Technology
Cryptocurrency trading bots are computer programs designed to automate buying, selling, and managing cryptocurrency transactions. These bots utilize advanced algorithms and machine learning techniques to analyze market data, identify trading opportunities, and execute trades on behalf of their users. By automating the decision-making process, crypto trading bots can react to market changes faster than human traders
Hivelance, a leading provider of cryptocurrency trading bot development services, stands out as the premier choice for crypto traders and developers. Hivelance boasts a team of seasoned cryptocurrency experts and software engineers who deeply understand the crypto market and the latest trends in automated trading, Hivelance leverages the latest technologies and tools in the industry, including advanced AI and machine learning algorithms, to create highly efficient and adaptable crypto trading bots
Software Engineering, Software Consulting, Tech Lead.
Spring Boot, Spring Cloud, Spring Core, Spring JDBC, Spring Security,
Spring Transaction, Spring MVC,
Log4j, REST/SOAP WEB-SERVICES.
Climate Science Flows: Enabling Petabyte-Scale Climate Analysis with the Eart...Globus
The Earth System Grid Federation (ESGF) is a global network of data servers that archives and distributes the planet’s largest collection of Earth system model output for thousands of climate and environmental scientists worldwide. Many of these petabyte-scale data archives are located in proximity to large high-performance computing (HPC) or cloud computing resources, but the primary workflow for data users consists of transferring data, and applying computations on a different system. As a part of the ESGF 2.0 US project (funded by the United States Department of Energy Office of Science), we developed pre-defined data workflows, which can be run on-demand, capable of applying many data reduction and data analysis to the large ESGF data archives, transferring only the resultant analysis (ex. visualizations, smaller data files). In this talk, we will showcase a few of these workflows, highlighting how Globus Flows can be used for petabyte-scale climate analysis.
We describe the deployment and use of Globus Compute for remote computation. This content is aimed at researchers who wish to compute on remote resources using a unified programming interface, as well as system administrators who will deploy and operate Globus Compute services on their research computing infrastructure.
In software engineering, the right architecture is essential for robust, scalable platforms. Wix has undergone a pivotal shift from event sourcing to a CRUD-based model for its microservices. This talk will chart the course of this pivotal journey.
Event sourcing, which records state changes as immutable events, provided robust auditing and "time travel" debugging for Wix Stores' microservices. Despite its benefits, the complexity it introduced in state management slowed development. Wix responded by adopting a simpler, unified CRUD model. This talk will explore the challenges of event sourcing and the advantages of Wix's new "CRUD on steroids" approach, which streamlines API integration and domain event management while preserving data integrity and system resilience.
Participants will gain valuable insights into Wix's strategies for ensuring atomicity in database updates and event production, as well as caching, materialization, and performance optimization techniques within a distributed system.
Join us to discover how Wix has mastered the art of balancing simplicity and extensibility, and learn how the re-adoption of the modest CRUD has turbocharged their development velocity, resilience, and scalability in a high-growth environment.
Providing Globus Services to Users of JASMIN for Environmental Data AnalysisGlobus
JASMIN is the UK’s high-performance data analysis platform for environmental science, operated by STFC on behalf of the UK Natural Environment Research Council (NERC). In addition to its role in hosting the CEDA Archive (NERC’s long-term repository for climate, atmospheric science & Earth observation data in the UK), JASMIN provides a collaborative platform to a community of around 2,000 scientists in the UK and beyond, providing nearly 400 environmental science projects with working space, compute resources and tools to facilitate their work. High-performance data transfer into and out of JASMIN has always been a key feature, with many scientists bringing model outputs from supercomputers elsewhere in the UK, to analyse against observational or other model data in the CEDA Archive. A growing number of JASMIN users are now realising the benefits of using the Globus service to provide reliable and efficient data movement and other tasks in this and other contexts. Further use cases involve long-distance (intercontinental) transfers to and from JASMIN, and collecting results from a mobile atmospheric radar system, pushing data to JASMIN via a lightweight Globus deployment. We provide details of how Globus fits into our current infrastructure, our experience of the recent migration to GCSv5.4, and of our interest in developing use of the wider ecosystem of Globus services for the benefit of our user community.
Code reviews are vital for ensuring good code quality. They serve as one of our last lines of defense against bugs and subpar code reaching production.
Yet, they often turn into annoying tasks riddled with frustration, hostility, unclear feedback and lack of standards. How can we improve this crucial process?
In this session we will cover:
- The Art of Effective Code Reviews
- Streamlining the Review Process
- Elevating Reviews with Automated Tools
By the end of this presentation, you'll have the knowledge on how to organize and improve your code review proces
Understanding Globus Data Transfers with NetSageGlobus
NetSage is an open privacy-aware network measurement, analysis, and visualization service designed to help end-users visualize and reason about large data transfers. NetSage traditionally has used a combination of passive measurements, including SNMP and flow data, as well as active measurements, mainly perfSONAR, to provide longitudinal network performance data visualization. It has been deployed by dozens of networks world wide, and is supported domestically by the Engagement and Performance Operations Center (EPOC), NSF #2328479. We have recently expanded the NetSage data sources to include logs for Globus data transfers, following the same privacy-preserving approach as for Flow data. Using the logs for the Texas Advanced Computing Center (TACC) as an example, this talk will walk through several different example use cases that NetSage can answer, including: Who is using Globus to share data with my institution, and what kind of performance are they able to achieve? How many transfers has Globus supported for us? Which sites are we sharing the most data with, and how is that changing over time? How is my site using Globus to move data internally, and what kind of performance do we see for those transfers? What percentage of data transfers at my institution used Globus, and how did the overall data transfer performance compare to the Globus users?
7. Azure Virtual Networks
• Cross-Premises Virtual Network
• Cloud-Only Virtual Network
• Allows to group Virtual Machines
• Active Directory Domain Services in VNet
• Can Azure AD help you?
10. High Availability
• Location
• 17 regions
• Service offerings – (http://tinyurl.com/nf6uwsk)
• Affinity groups
• Colocation of cloud services and storage
• Host for an affinity group VNet
• VNet (Regional)
• Upgraded in late 2014
• http://tinyurl.com/nesvcjn
• Cloud Service
• Availability Sets
13. Disks, disks & more disks
• Always create large disks – you pay for actual use
• Performance of disks
• Blob storage for storing VHD
• Blobs are limited to 500 IOPS per disk
• Basic tier 300 IOPS per disk
• Storage space - http://tinyurl.com/n5ljfr8
• SQL Specific
• Format the disks with 64K allocation
• SQL Server service account – “Perform Volume Maintenance Tasks”
• Do NOT store anything in Temp Disk
14. Minimal IOPS
• Configuration DB – 2GB
• 40Mb for each 50,000 site collections
• Recovery model – full to simple
• Central Administration – 1GB
15. Content DB
• Formula
• Content DB size = ((D × V) × S) + (10 KB × (L + (V × D)))
• http://tinyurl.com/k75suaz
• Recycle bins
• Auditing
• 2Kb per entry
• Best practice 0.5 IOPS/GB
• range from 0.05 IOPS/GB to around 0.2 IOPS/GB
• RAID 10
• 5 Content DBs per disk
• Read & Write caching for Data Disks – Set-AzureDataDisk
• http://tinyurl.com/n4ktw46
• SQL File groups (Data files on different disks)
• Disk striping if you need over 1TB
Speaking Points:
Here’s another way to look at the cloud services taxonomy and how this taxonomy maps to the components in an IT infrastructure.
Packaged Software
With packaged software a customer would be responsible for managing the entire stack – ranging from the network connectivity to the applications.
IaaS
With Infrastructure as a Service, the lower levels of the stack are managed by a vendor. Some of these components can be provided by traditional hosters – in fact most of them have moved to having a virtualized offering.
Very few actually provide an OS
The customer is still responsible for managing the OS through the Applications.
For the developer, an obvious benefit with IaaS is that it frees the developer from many concerns when provisioning physical or virtual machines.
This was one of the earliest and primary use cases for Amazon Web Services Elastic Cloud Compute (EC2).
Developers were able to readily provision virtual machines (AMIs) on EC2, develop and test solutions and, often, run the results ‘in production’.
The only requirement was a credit card to pay for the services.
PaaS
With Platform as a Service, everything from the network connectivity through the runtime is provided and managed by the platform vendor.
The Windows Azure best fits in this category today.
In fact because we don’t provide access to the underlying virtualization or operating system today, we’re often referred to as not providing IaaS.
PaaS offerings further reduce the developer burden by additionally supporting the platform runtime and related application services.
With PaaS, the developer can, almost immediately, begin creating the business logic for an application.
Potentially, the increases in productivity are considerable and, because the hardware and operational aspects of the cloud platform are also managed by the cloud platform provider, applications can quickly be taken from an idea to reality very quickly.
SaaS
Finally, with SaaS, a vendor provides the application and abstracts you from all of the underlying components.
Cloud Services
“Cloud Services” are used in Azure to “group” instances together. Cloud Services have features such as end-points, load balancing and auto-scale. For SharePoint and SQL don’t even think about auto-scale though.
Availability Sets
Availability sets are a logical construct that allows you to specify a group of instances/roles. Whenever Microsoft must reboot one of your VMs for maintenance or other purposes, they will respect the Availability Sets and make sure that only one of the machines within an Availability Set are down at a time. For instance grouping all SharePoint Web Servers into one Availability Set, makes sense.