This document provides an overview and best practices for storage technologies. It discusses factors that affect storage performance like interconnect bandwidth versus IOPS and command sizing. It covers tiering strategies and when auto-tiering may not be effective. It also discusses SSDs versus spinning disks, large VMDK and VMFS support, thin provisioning at the VM and LUN level, and architecting storage for failure including individual component failure, temporary and permanent site loss. It provides examples of how to implement a low-cost disaster recovery site using inexpensive hardware.
VMworld 2013: Storage DRS: Deep Dive and Best Practices to Suit Your Storage ...VMworld
VMworld 2013
Sachin Manpathak, VMware
Mustafa Uysal, VMware
Sunil Muralidhar, VMware
Learn more about VMworld and register at http://www.vmworld.com/index.jspa?src=socmed-vmworld-slideshare
This document discusses database deployment automation. It begins with introductions and an example of a problematic Friday deployment. It then reviews the concept of automation and different visions of it within an organization. Potential tools and frameworks for automation are discussed, along with common pitfalls. Basic deployment workflows using Oracle Cloud Control are demonstrated, including setting credentials, creating a proxy user, adding target properties, and using a job template. The document concludes by emphasizing that database deployment automation is possible but requires effort from multiple teams.
This document discusses best practices for virtualizing databases. It begins with an introduction of the presenters, Michael Corey and Jeff Szastak, who are experts in virtualizing Oracle and SQL Server databases. The document then covers reasons for virtualizing databases, including flexibility, efficiency of resources, and cost savings. It provides examples of large production databases that have been successfully virtualized. The document discusses performance results from testing that show virtualized database performance is typically within 5% of physical performance. It provides recommendations for right-sizing resources and avoiding configurations like BIOS settings that could negatively impact performance. The overall message is that databases can be successfully virtualized while meeting service level agreements by following best practices.
VMworld 2013: How SRP Delivers More Than Power to Their Customers VMworld
VMworld 2013
Sheldon Brown, SRP
Girish Manmadkar, VMware
Learn more about VMworld and register at http://www.vmworld.com/index.jspa?src=socmed-vmworld-slideshare
Best Practices of HA and Replication of PostgreSQL in Virtualized EnvironmentsJignesh Shah
This document discusses best practices for high availability (HA) and replication of PostgreSQL databases in virtualized environments. It covers enterprise needs for HA, technologies like VMware HA and replication that can provide HA, and deployment blueprints for HA, read scaling, and disaster recovery within and across datacenters. The document also discusses PostgreSQL's different replication modes and how they can be used for HA, read scaling, and disaster recovery.
The document discusses how QNAP Turbo NAS can help businesses by providing cloud storage solutions and file sharing capabilities. It highlights features like syncing files across devices, integrating cloud services like Google Drive and Dropbox, and providing a private cloud solution for sharing and synchronizing work files. The NAS also offers storage management features like snapshots, SSD caching, and auto-tiering to optimize performance. It aims to help businesses address challenges of growing storage needs, syncing work in the cloud, and sharing and protecting files and data.
This document provides an overview of JetStor's data storage platform. It introduces the JetStor SAN/NAS Platform which offers a single architecture for datastore, backup, disaster recovery, file storage and production storage. The platform includes various storage array models suited for hybrid-flash block storage, all-flash block storage, file storage and unified storage. Key features highlighted include RAID-EE for faster rebuild times, thin provisioning, snapshots, replication, tiering and a centralized management system. Performance comparisons show JetStor arrays outperforming other solutions. The document promotes JetStor's all-flash arrays for demanding workloads like VDI and virtualization clustering.
Why does my choice of storage matter with cassandra?Johnny Miller
The document discusses how the choice of storage is critical for Cassandra deployments. It summarizes that SSDs are generally the best choice as they have no moving parts, resulting in much faster performance compared to HDDs. Specifically, SSDs can eliminate issues caused by disk seeks and allow the use of compaction strategies like leveled compaction that require lower seek times. The document provides measurements showing SSDs are up to 100x faster than HDDs for read/write speeds and latency. It recommends choosing local SSD storage in a JBOD configuration when possible for best performance and manageability.
VMworld 2013: Storage DRS: Deep Dive and Best Practices to Suit Your Storage ...VMworld
VMworld 2013
Sachin Manpathak, VMware
Mustafa Uysal, VMware
Sunil Muralidhar, VMware
Learn more about VMworld and register at http://www.vmworld.com/index.jspa?src=socmed-vmworld-slideshare
This document discusses database deployment automation. It begins with introductions and an example of a problematic Friday deployment. It then reviews the concept of automation and different visions of it within an organization. Potential tools and frameworks for automation are discussed, along with common pitfalls. Basic deployment workflows using Oracle Cloud Control are demonstrated, including setting credentials, creating a proxy user, adding target properties, and using a job template. The document concludes by emphasizing that database deployment automation is possible but requires effort from multiple teams.
This document discusses best practices for virtualizing databases. It begins with an introduction of the presenters, Michael Corey and Jeff Szastak, who are experts in virtualizing Oracle and SQL Server databases. The document then covers reasons for virtualizing databases, including flexibility, efficiency of resources, and cost savings. It provides examples of large production databases that have been successfully virtualized. The document discusses performance results from testing that show virtualized database performance is typically within 5% of physical performance. It provides recommendations for right-sizing resources and avoiding configurations like BIOS settings that could negatively impact performance. The overall message is that databases can be successfully virtualized while meeting service level agreements by following best practices.
VMworld 2013: How SRP Delivers More Than Power to Their Customers VMworld
VMworld 2013
Sheldon Brown, SRP
Girish Manmadkar, VMware
Learn more about VMworld and register at http://www.vmworld.com/index.jspa?src=socmed-vmworld-slideshare
Best Practices of HA and Replication of PostgreSQL in Virtualized EnvironmentsJignesh Shah
This document discusses best practices for high availability (HA) and replication of PostgreSQL databases in virtualized environments. It covers enterprise needs for HA, technologies like VMware HA and replication that can provide HA, and deployment blueprints for HA, read scaling, and disaster recovery within and across datacenters. The document also discusses PostgreSQL's different replication modes and how they can be used for HA, read scaling, and disaster recovery.
The document discusses how QNAP Turbo NAS can help businesses by providing cloud storage solutions and file sharing capabilities. It highlights features like syncing files across devices, integrating cloud services like Google Drive and Dropbox, and providing a private cloud solution for sharing and synchronizing work files. The NAS also offers storage management features like snapshots, SSD caching, and auto-tiering to optimize performance. It aims to help businesses address challenges of growing storage needs, syncing work in the cloud, and sharing and protecting files and data.
This document provides an overview of JetStor's data storage platform. It introduces the JetStor SAN/NAS Platform which offers a single architecture for datastore, backup, disaster recovery, file storage and production storage. The platform includes various storage array models suited for hybrid-flash block storage, all-flash block storage, file storage and unified storage. Key features highlighted include RAID-EE for faster rebuild times, thin provisioning, snapshots, replication, tiering and a centralized management system. Performance comparisons show JetStor arrays outperforming other solutions. The document promotes JetStor's all-flash arrays for demanding workloads like VDI and virtualization clustering.
Why does my choice of storage matter with cassandra?Johnny Miller
The document discusses how the choice of storage is critical for Cassandra deployments. It summarizes that SSDs are generally the best choice as they have no moving parts, resulting in much faster performance compared to HDDs. Specifically, SSDs can eliminate issues caused by disk seeks and allow the use of compaction strategies like leveled compaction that require lower seek times. The document provides measurements showing SSDs are up to 100x faster than HDDs for read/write speeds and latency. It recommends choosing local SSD storage in a JBOD configuration when possible for best performance and manageability.
This document discusses tuning DB2 in a Solaris environment. It provides background on the presenters, Tom Bauch from IBM and Jignesh Shah from Sun Microsystems. The agenda covers general considerations, memory usage and bottlenecks, disk I/O considerations and bottlenecks, and tuning DB2 V8.1 specifically in Solaris 9. It discusses supported Solaris versions, kernel settings, required patches, installation methods, and the configuration wizard. Specific topics covered in more depth include the Data Partitioning Feature, DB2 Enterprise Server Edition, and analyzing and addressing potential memory bottlenecks.
San Francisco Cassadnra Meetup - March 2014: I/O Performance tuning on AWS fo...DataStax Academy
What You'll Learn at this Meetup
Tips and Tricks to achieve high performance when running Cassandra on AWS
• Configuration tuning for Cassandra
• Tools to benchmark raw filesystem IO
• AWS available AMIs to boost performance
• Stress testing on AWS i2 HVM instances
• Configuring AWS EC2 instances with SSDs and EBS storage with PIOPS
My experience with embedding PostgreSQLJignesh Shah
At my current company, we embed PostgreSQL based technologies in various applications shipped as shrink-wrapped software. In this session we talk about the experience of embedding PostgreSQL where it is not directly exposed to end-user and the issues encountered on how they were resolved.
We will talk about business reasons,technical architecture of deployments, upgrades, security processes on how to work with embedded PostgreSQL databases.
How to deploy SQL Server on an Microsoft Azure virtual machinesSolarWinds
Running apps on Microsoft Azure Virtual Machines is tempting; promising faster deployments and lower overall TCO. But how easy is it really to configure and run SQL Server in an Azure VM environment? Learn what you should know about tuning, optimizing, and key indicators for monitoring performance, as well as special considerations for High-Availability and Disaster Recovery.
Lightweight locks (LWLocks) in PostgreSQL provide mutually exclusive access to shared memory structures. They support both shared and exclusive locking modes. The LWLocks framework uses wait queues, semaphores, and spinlocks to efficiently manage acquiring and releasing locks. Dynamic monitoring of LWLock events is possible through special builds that incorporate statistics collection.
Ceph Community Talk on High-Performance Solid Sate Ceph Ceph Community
The document summarizes a presentation given by representatives from various companies on optimizing Ceph for high-performance solid state drives. It discusses testing a real workload on a Ceph cluster with 50 SSD nodes that achieved over 280,000 read and write IOPS. Areas for further optimization were identified, such as reducing latency spikes and improving single-threaded performance. Various companies then described their contributions to Ceph performance, such as Intel providing hardware for testing and Samsung discussing SSD interface improvements.
The best DBAs tune SQL Server for performance at the server, instance, and database layers. This allows for both the logical and physical database designs to meet performance expectations. But it can be difficult to know which configuration options are better than others. Learn expert tips from Microsoft Certified Masters Tim Chapman and Thomas LaRock.
Yesterday's thinking may still believe NVMe (NVM Express) is in transition to a production ready solution. In this session, we will discuss how the evolution of NVMe is ready for production, the history and evolution of NVMe and the Linux stack to address where NVMe has progressed today to become the low latency, highly reliable database key value store mechanism that will drive the future of cloud expansion. Examples of protocol efficiencies and types of storage engines that are optimizing for NVMe will be discussed. Please join us for an exciting session where in-memory computing and persistence have evolved.
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSCeph Community
This document summarizes the performance of an all-NVMe Ceph cluster using Intel P3700 NVMe SSDs. Key results include achieving over 1.35 million 4K random read IOPS and 171K 4K random write IOPS with sub-millisecond latency. Partitioning the NVMe drives into multiple OSDs improved performance and CPU utilization compared to a single OSD per drive. The cluster also demonstrated over 5GB/s of sequential bandwidth.
Thin provisioning allows storage arrays to provision more capacity than is physically available by only allocating space as it is used. This improves efficiency but can lead to issues if overprovisioned storage runs out. There are challenges to thin provisioning across different layers including file systems, virtualization, and storage arrays. For thin provisioning to be effective, all layers must work together to monitor capacity usage and free space accurately at a fine granularity.
VMworld 2013: VMware Virtual SAN Technical Best Practices VMworld
This document provides an overview of VMware Virtual SAN (VSAN) technical best practices. It discusses VSAN's key components, hardware considerations, use cases, management, and demo. VSAN is a software-defined storage solution that clusters direct-attached host storage and provides a virtual SAN datastore. It has integrated management with vSphere and uses capabilities and policies to enable VM-centric storage provisioning and automation. The document demonstrates how to configure VSAN, create VM storage policies, and deploy VMs according to policies and capabilities.
Doble CPU y doble sistema operativo: Una solución NAS, SAN perfecta para tareas informáticas de alto rendimiento y almacenamiento, utilizando el sistema operativo QES o QTS
This session covers the engineering strategies and lessons learned at IBM creating industry leading in-memory data warehousing technology for use with both cloud and on-premises software. Along with rich in-memory SQL support for OLAP, data mining, and data warehousing leveraging memory optimized parallel vector processing, we’ll showcase the in-database analytics for R, spatial, and the built-in synchronization with Cloudant JSON NoSQL. We'll take a closer look at the architectural strategy for treating RAM as the new disk (and worth avoiding access to), while dramatically constraining the potential cost pressures of in-memory technology. We’ll describe how we designed for super-simplicity with load-and-go no-tuning technology for any size system, and of course… a demo. Ridiculously easy to use and freakishly fast. Not your grandmother’s IBM database.
Ceph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance BarriersCeph Community
The document discusses a presentation about Ceph on all-flash storage using InfiniFlash systems to break performance barriers. It describes how Ceph has been optimized for flash storage and how InfiniFlash systems provide industry-leading performance of over 1 million IOPS and 6-9GB/s of throughput using SanDisk flash technology. The presentation also covers how InfiniFlash can provide scalable performance and capacity for large-scale enterprise workloads.
Flash for the Real World – Separate Hype from RealityHitachi Vantara
Join us for a live webcast and hear Hu Yoshida, Chief Technology Officer of Hitachi Data Systems, discuss the real world criteria for making an effective decision when evaluating flash storage. With all the noise in the market it can be difficult to separate fact from fiction in order to evaluate the performance, efficiency and economic trade-offs for flash storage.
Specifically, you’ll learn how to determine if flash storage will help you:
Actually achieve the performance you need as you compare technology options.
Realize efficiency gains that extend beyond the promise of flash performance.
Make the economic case for real-world business decisions before taking the leap.
This document discusses various options for deploying solid state drives (SSDs) in the data center to address storage performance issues. It describes all-flash arrays that use only SSDs, hybrid arrays that combine SSDs and hard disk drives, and server-side flash caching. Key points covered include the performance benefits of SSDs over HDDs, different types of SSDs, form factors, deployment architectures like all-flash arrays from vendors, hybrid arrays, server-side caching software, virtual storage appliances, and hyperconverged infrastructure systems. Choosing the best solution depends on factors like performance needs, capacity, data services required, and budget.
Developing a Ceph Appliance for Secure EnvironmentsCeph Community
Keeper Technology develops a Ceph appliance called keeperSAFE for secure storage environments. The keeperSAFE appliance provides a preconfigured Linux distribution, automated installation using Ansible, enclosure management tools, a graphical user interface for monitoring and configuration, data collection and analytics, encryption capabilities, and extensive testing. It is designed for environments that require high availability, no single points of failure, easy management, and auditability. The keeperSAFE appliance addresses the challenges of deploying and managing Ceph at scale in restricted, mission critical environments.
How to configure SQL Server for SSDs and VMsSolarWinds
Recent trends in virtualized converged infrastructures, coupled with rapidly decreasing costs of solid state storage (SSD), pose new challenges for DBAs and developers alike. Learn answers to to pressing questions like:
• If my array uses compression, are there benefits to using database compression?
• How should I arrange my data files on an SSD array?
• If I have limited SSD space, how should I use it?
• On a hybrid array (spinning disk and SSD) how should I structure policies?
• What changes from physical servers to VMs?
This document discusses migrating an Oracle Database Appliance (ODA) from a bare metal to a virtualized platform. It outlines the initial situation, desired target, challenges, and solution approach. The key challenges included system downtime during the migration, backup/restore processes, using external storage, and database reorganizations. The solution involved first converting to a virtual platform and then upgrading, using backup/restore, attaching an NGENSTOR Hurricane storage appliance for direct attached storage, and moving database reorganizations to a separate maintenance window. It also discusses the odaback-API tool created to help automate and standardize the migration process.
Storage and performance- Batch processing, WhiptailInternet World
Batch processing allows jobs to run without manual intervention by shifting processing to less busy times. It avoids idling computing resources and allows higher overall utilization. Batch processing provides benefits like prioritizing batch and interactive work. The document then discusses different approaches to batch processing like dedicating all resources to it or sharing resources. It outlines challenges like systems being unavailable during batch processing. The rest of the document summarizes Whiptail's flash storage solutions for accelerating workloads and reducing costs and resources compared to HDDs.
This document discusses tuning DB2 in a Solaris environment. It provides background on the presenters, Tom Bauch from IBM and Jignesh Shah from Sun Microsystems. The agenda covers general considerations, memory usage and bottlenecks, disk I/O considerations and bottlenecks, and tuning DB2 V8.1 specifically in Solaris 9. It discusses supported Solaris versions, kernel settings, required patches, installation methods, and the configuration wizard. Specific topics covered in more depth include the Data Partitioning Feature, DB2 Enterprise Server Edition, and analyzing and addressing potential memory bottlenecks.
San Francisco Cassadnra Meetup - March 2014: I/O Performance tuning on AWS fo...DataStax Academy
What You'll Learn at this Meetup
Tips and Tricks to achieve high performance when running Cassandra on AWS
• Configuration tuning for Cassandra
• Tools to benchmark raw filesystem IO
• AWS available AMIs to boost performance
• Stress testing on AWS i2 HVM instances
• Configuring AWS EC2 instances with SSDs and EBS storage with PIOPS
My experience with embedding PostgreSQLJignesh Shah
At my current company, we embed PostgreSQL based technologies in various applications shipped as shrink-wrapped software. In this session we talk about the experience of embedding PostgreSQL where it is not directly exposed to end-user and the issues encountered on how they were resolved.
We will talk about business reasons,technical architecture of deployments, upgrades, security processes on how to work with embedded PostgreSQL databases.
How to deploy SQL Server on an Microsoft Azure virtual machinesSolarWinds
Running apps on Microsoft Azure Virtual Machines is tempting; promising faster deployments and lower overall TCO. But how easy is it really to configure and run SQL Server in an Azure VM environment? Learn what you should know about tuning, optimizing, and key indicators for monitoring performance, as well as special considerations for High-Availability and Disaster Recovery.
Lightweight locks (LWLocks) in PostgreSQL provide mutually exclusive access to shared memory structures. They support both shared and exclusive locking modes. The LWLocks framework uses wait queues, semaphores, and spinlocks to efficiently manage acquiring and releasing locks. Dynamic monitoring of LWLock events is possible through special builds that incorporate statistics collection.
Ceph Community Talk on High-Performance Solid Sate Ceph Ceph Community
The document summarizes a presentation given by representatives from various companies on optimizing Ceph for high-performance solid state drives. It discusses testing a real workload on a Ceph cluster with 50 SSD nodes that achieved over 280,000 read and write IOPS. Areas for further optimization were identified, such as reducing latency spikes and improving single-threaded performance. Various companies then described their contributions to Ceph performance, such as Intel providing hardware for testing and Samsung discussing SSD interface improvements.
The best DBAs tune SQL Server for performance at the server, instance, and database layers. This allows for both the logical and physical database designs to meet performance expectations. But it can be difficult to know which configuration options are better than others. Learn expert tips from Microsoft Certified Masters Tim Chapman and Thomas LaRock.
Yesterday's thinking may still believe NVMe (NVM Express) is in transition to a production ready solution. In this session, we will discuss how the evolution of NVMe is ready for production, the history and evolution of NVMe and the Linux stack to address where NVMe has progressed today to become the low latency, highly reliable database key value store mechanism that will drive the future of cloud expansion. Examples of protocol efficiencies and types of storage engines that are optimizing for NVMe will be discussed. Please join us for an exciting session where in-memory computing and persistence have evolved.
Accelerating Cassandra Workloads on Ceph with All-Flash PCIE SSDSCeph Community
This document summarizes the performance of an all-NVMe Ceph cluster using Intel P3700 NVMe SSDs. Key results include achieving over 1.35 million 4K random read IOPS and 171K 4K random write IOPS with sub-millisecond latency. Partitioning the NVMe drives into multiple OSDs improved performance and CPU utilization compared to a single OSD per drive. The cluster also demonstrated over 5GB/s of sequential bandwidth.
Thin provisioning allows storage arrays to provision more capacity than is physically available by only allocating space as it is used. This improves efficiency but can lead to issues if overprovisioned storage runs out. There are challenges to thin provisioning across different layers including file systems, virtualization, and storage arrays. For thin provisioning to be effective, all layers must work together to monitor capacity usage and free space accurately at a fine granularity.
VMworld 2013: VMware Virtual SAN Technical Best Practices VMworld
This document provides an overview of VMware Virtual SAN (VSAN) technical best practices. It discusses VSAN's key components, hardware considerations, use cases, management, and demo. VSAN is a software-defined storage solution that clusters direct-attached host storage and provides a virtual SAN datastore. It has integrated management with vSphere and uses capabilities and policies to enable VM-centric storage provisioning and automation. The document demonstrates how to configure VSAN, create VM storage policies, and deploy VMs according to policies and capabilities.
Doble CPU y doble sistema operativo: Una solución NAS, SAN perfecta para tareas informáticas de alto rendimiento y almacenamiento, utilizando el sistema operativo QES o QTS
This session covers the engineering strategies and lessons learned at IBM creating industry leading in-memory data warehousing technology for use with both cloud and on-premises software. Along with rich in-memory SQL support for OLAP, data mining, and data warehousing leveraging memory optimized parallel vector processing, we’ll showcase the in-database analytics for R, spatial, and the built-in synchronization with Cloudant JSON NoSQL. We'll take a closer look at the architectural strategy for treating RAM as the new disk (and worth avoiding access to), while dramatically constraining the potential cost pressures of in-memory technology. We’ll describe how we designed for super-simplicity with load-and-go no-tuning technology for any size system, and of course… a demo. Ridiculously easy to use and freakishly fast. Not your grandmother’s IBM database.
Ceph Day Melbourne - Ceph on All-Flash Storage - Breaking Performance BarriersCeph Community
The document discusses a presentation about Ceph on all-flash storage using InfiniFlash systems to break performance barriers. It describes how Ceph has been optimized for flash storage and how InfiniFlash systems provide industry-leading performance of over 1 million IOPS and 6-9GB/s of throughput using SanDisk flash technology. The presentation also covers how InfiniFlash can provide scalable performance and capacity for large-scale enterprise workloads.
Flash for the Real World – Separate Hype from RealityHitachi Vantara
Join us for a live webcast and hear Hu Yoshida, Chief Technology Officer of Hitachi Data Systems, discuss the real world criteria for making an effective decision when evaluating flash storage. With all the noise in the market it can be difficult to separate fact from fiction in order to evaluate the performance, efficiency and economic trade-offs for flash storage.
Specifically, you’ll learn how to determine if flash storage will help you:
Actually achieve the performance you need as you compare technology options.
Realize efficiency gains that extend beyond the promise of flash performance.
Make the economic case for real-world business decisions before taking the leap.
This document discusses various options for deploying solid state drives (SSDs) in the data center to address storage performance issues. It describes all-flash arrays that use only SSDs, hybrid arrays that combine SSDs and hard disk drives, and server-side flash caching. Key points covered include the performance benefits of SSDs over HDDs, different types of SSDs, form factors, deployment architectures like all-flash arrays from vendors, hybrid arrays, server-side caching software, virtual storage appliances, and hyperconverged infrastructure systems. Choosing the best solution depends on factors like performance needs, capacity, data services required, and budget.
Developing a Ceph Appliance for Secure EnvironmentsCeph Community
Keeper Technology develops a Ceph appliance called keeperSAFE for secure storage environments. The keeperSAFE appliance provides a preconfigured Linux distribution, automated installation using Ansible, enclosure management tools, a graphical user interface for monitoring and configuration, data collection and analytics, encryption capabilities, and extensive testing. It is designed for environments that require high availability, no single points of failure, easy management, and auditability. The keeperSAFE appliance addresses the challenges of deploying and managing Ceph at scale in restricted, mission critical environments.
How to configure SQL Server for SSDs and VMsSolarWinds
Recent trends in virtualized converged infrastructures, coupled with rapidly decreasing costs of solid state storage (SSD), pose new challenges for DBAs and developers alike. Learn answers to to pressing questions like:
• If my array uses compression, are there benefits to using database compression?
• How should I arrange my data files on an SSD array?
• If I have limited SSD space, how should I use it?
• On a hybrid array (spinning disk and SSD) how should I structure policies?
• What changes from physical servers to VMs?
This document discusses migrating an Oracle Database Appliance (ODA) from a bare metal to a virtualized platform. It outlines the initial situation, desired target, challenges, and solution approach. The key challenges included system downtime during the migration, backup/restore processes, using external storage, and database reorganizations. The solution involved first converting to a virtual platform and then upgrading, using backup/restore, attaching an NGENSTOR Hurricane storage appliance for direct attached storage, and moving database reorganizations to a separate maintenance window. It also discusses the odaback-API tool created to help automate and standardize the migration process.
Storage and performance- Batch processing, WhiptailInternet World
Batch processing allows jobs to run without manual intervention by shifting processing to less busy times. It avoids idling computing resources and allows higher overall utilization. Batch processing provides benefits like prioritizing batch and interactive work. The document then discusses different approaches to batch processing like dedicating all resources to it or sharing resources. It outlines challenges like systems being unavailable during batch processing. The rest of the document summarizes Whiptail's flash storage solutions for accelerating workloads and reducing costs and resources compared to HDDs.
Data deduplication is a hot topic in storage and saves significant disk space for many environments, with some trade offs. We’ll discuss what deduplication is and where the Open Source solutions are versus commercial offerings. Presentation will lean towards the practical – where attendees can use it in their real world projects (what works, what doesn’t, should you use in production, etcetera).
This document provides guidance and best practices for migrating database workloads to infrastructure as a service (IaaS) in Microsoft Azure. It discusses choosing the appropriate virtual machine series and storage options to meet performance needs. The document emphasizes migrating the workload, not the hardware, and using cloud services to simplify management like automated patching and backup snapshots. It also recommends bringing existing monitoring and management tools to the cloud when possible rather than replacing them. The key takeaways are to understand the workload demands, choose optimal IaaS configurations, leverage cloud-enabled tools, and involve database experts when issues arise to address the root cause rather than just adding resources.
Ceph Day London 2014 - Best Practices for Ceph-powered Implementations of Sto...Ceph Community
This document discusses Dell's support for CEPH storage solutions and provides an agenda for a CEPH Day event at Dell. Key points include:
- Dell is a certified reseller of Red Hat-Inktank CEPH support, services, and training.
- The agenda covers why Dell supports CEPH, hardware recommendations, best practices shared with CEPH colleagues, and a concept for research data storage that is seeking input.
- Recommended CEPH architectures, components, configurations, and considerations are discussed for planning and implementing a CEPH solution. Dell server hardware options that could be used are also presented.
VMworld Europe 2014: Virtual SAN Best Practices and Use CasesVMworld
This document provides an overview and agenda for a presentation on VMware Virtual SAN. It discusses key features of Virtual SAN including its software-defined storage approach and hybrid storage using SSD and HDD. Several use cases are reviewed like virtual desktop infrastructure, remote office/branch office, and DMZ/isolated environments. Best practices are also covered for various use cases around sizing, policies, and ready nodes. The document aims to introduce attendees to Virtual SAN capabilities and considerations for different deployment scenarios.
The document discusses troubleshooting storage performance in a vSphere environment. It covers using vCenter performance charts and ESXTop to analyze disk and storage path latency. Acceptable latency values are provided for different storage protocols. Ensuring proper disk alignment between the guest operating system and storage array is also discussed as an important factor for storage performance. Maintaining an optimal queue depth is recommended over arbitrarily increasing it.
On X86 systems, using an Unbreakable Enterprise Kernel (UEK) is recommended over other enterprise distributions as it provides better hardware support, security patches, and testing from the larger Linux community. Key configuration recommendations include enabling maximum CPU performance in BIOS, using memory types validated by Oracle, ensuring proper NUMA and CPU frequency settings, and installing only Oracle-validated packages to avoid issues. Monitoring tools like top, iostat, sar and ksar help identify any CPU, memory, disk or I/O bottlenecks.
The document discusses various ways to tune Linux and MySQL for performance. It recommends measuring different aspects of the database, operating system, disk and application performance. Some specific tuning techniques discussed include testing different IO schedulers, increasing the number of InnoDB threads, reducing swapping by lowering the swappiness value, enabling interleave mode for NUMA systems, and potentially using huge pages, though noting the complexity of configuring huge pages. The key message is that default settings may not be optimal and testing is needed to understand each individual system's performance.
This document discusses Linux huge pages, including:
- What huge pages are and how they can reduce memory management overhead by allocating larger blocks of memory
- How to configure huge pages on Linux, including installing required packages, mounting the huge page filesystem, and setting kernel parameters
- When huge pages should be configured, such as for data-intensive or latency-sensitive applications like databases, but that testing is required due to disadvantages like reduced swappability
Virtualizing SQL Server workloads can provide high availability, flexibility, and portability while maximizing performance. Key considerations for virtualized SQL include properly sizing and configuring virtual CPUs, memory, storage, and networking. Features like SQL Always-On clustering allow for high availability without shared storage. Host-based backups and SQL maintenance plans are both important for backup strategies. Templates simplify deployment and updates of virtualized SQL servers.
VMworld 2014: Advanced SQL Server on vSphere Techniques and Best PracticesVMworld
This document provides an overview of advanced SQL Server techniques and best practices when running SQL Server in a virtualized environment on vSphere. It covers topics such as storage configuration including VMFS, block alignment, and I/O profiling. Networking techniques like jumbo frames and guest tuning are discussed. The document also reviews memory management and optimization, CPU sizing considerations, workload consolidation strategies, and high availability options for SQL Server on vSphere.
The document provides recommendations for optimizing an OpenStack cloud environment using Ceph storage. It discusses configuring Glance, Cinder, and Nova to integrate with Ceph, as well as recommendations for the Ceph cluster itself regarding OSDs, journals, networking, and failure domains. Performance was improved by converting image formats to raw, enabling SSD journals, bonding network interfaces, and adjusting scrubbing settings.
Best Practices & Performance Tuning - OpenStack Cloud Storage with Ceph - In this presentation, we discuss best practices and performance tuning for OpenStack cloud storage with Ceph to achieve high availability, durability, reliability and scalability at any point of time. Also discuss best practices for failure domain, recovery, rebalancing, backfilling, scrubbing, deep-scrubbing and operations
The document discusses best practices for running MySQL on Linux, covering choices for Linux distributions, hardware recommendations including using solid state drives, OS configuration such as tuning the filesystem and IO scheduler, and MySQL installation and configuration options. It provides guidance on topics like virtualization, networking, and MySQL variants to help ensure successful and high performance deployment of MySQL on Linux.
2015 deploying flash in the data centerHoward Marks
This document discusses deploying flash storage in the data center to improve storage performance. It begins with an overview of the performance gap between processors and disks. It then discusses all-flash arrays, hybrid arrays, server-side flash caching, and converged architectures as solutions. It provides details on flash memory types, form factors, and considerations for choosing a flash solution.
2015 deploying flash in the data centerHoward Marks
Deploying Flash in the Data Center discusses various ways to deploy flash storage in the data center to improve performance. It describes all-flash arrays that provide the highest performance but also more expensive options like hybrid arrays that combine flash and disk. It also covers using flash in servers or as a cache to accelerate storage arrays. Choosing the best approach depends on factors like workload, budget, and existing infrastructure.
DataStax: Extreme Cassandra Optimization: The SequelDataStax Academy
Al has been using Cassandra since version 0.6 and has spent the last few months doing little else but tune Cassandra clusters. In this talk, Al will show how to tune Cassandra for efficient operation using multiple views into system metrics, including OS stats, GC logs, JMX, and cassandra-stress.
Tuning Linux for your database FLOSSUK 2016Colin Charles
Some best practices about tuning Linux for your database workloads. The focus is not just on MySQL or MariaDB Server but also on understanding the OS from hardware/cloud, I/O, filesystems, memory, CPU, network, and resources.
Similar to VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0) (20)
VMworld 2016: vSphere 6.x Host Resource Deep DiveVMworld
1. This document provides an overview and agenda for a presentation on vSphere 6.x host resource deep dive topics including compute, storage, and network.
2. It introduces the presenters, Niels Hagoort and Frank Denneman, and provides background on their expertise.
3. The document outlines the topics to be covered under each section, including NUMA, CPU cache, DIMM configuration, I/O queue placement, driver considerations, RSS and NetQueue scaling for networking.
VMworld 2016: Troubleshooting 101 for HorizonVMworld
This document provides an overview of troubleshooting tools and techniques for Horizon. It begins with introductions and disclaimers. It then covers defining problems, identifying symptoms, gathering additional information, determining possible causes, identifying the root cause, resolving problems, and documenting solutions. Common troubleshooting tools are discussed, including ESXCLI commands, vSphere CLI commands, and log file locations and contents. Methods for collecting log files from Horizon components like desktops, clients, and servers are also provided.
VMworld 2016: Advanced Network Services with NSXVMworld
NSX provides network virtualization and security services including distributed firewalling, load balancing, and VPN connectivity. It reproduces traditional network and security functions in software throughout the virtual infrastructure for improved performance, agility, and security compared to physical appliances. Over 1700 customers use NSX across various industries, with growth of 100% year-over-year. NSX services can be distributed across hypervisors for massive scalability. The platform also integrates with security and application delivery partners to enhance its native capabilities.
VMworld 2016: How to Deploy VMware NSX with Cisco InfrastructureVMworld
This document provides an overview of how to deploy VMware NSX with Cisco infrastructure, including:
- NSX has minimal requirements of 1600 MTU and IP connectivity and is agnostic to the underlying network topology.
- When using Cisco Nexus switches, VLANs must be configured for various traffic types and SVIs created with consistent IP subnets. Jumbo MTU is required across all links.
- NSX is also compatible with Cisco ACI fabrics using Fabric Path or DFA topologies, with the VXLAN VLAN spanning multiple pods/clusters across the fabric.
VMworld 2016: Enforcing a vSphere Cluster Design with PowerCLI AutomationVMworld
This document discusses enforcing vSphere cluster designs using PowerCLI automation. It provides an overview of vSphere cluster design basics like HA and DRS configurations. It then discusses crafting declarative configurations to define the desired infrastructure state. Infrastructure as code principles are reviewed for managing configurations outside the endpoints. The presentation introduces the Vester project for declaratively configuring vSphere clusters using PowerCLI.
Horizon 7 introduces several new features including just-in-time desktops that instantly provision desktops and applications when users log in using VMware's instant clone technology. It also features smart policies that dynamically change desktop configurations based on user location or device. Infrastructure updates improve scalability and failover capabilities. The user experience is enhanced with support for 3D graphics, new protocols like Blast Extreme for optimized mobile access, and expanded capabilities for hosted applications and RDS desktops.
VMworld 2016: Virtual Volumes Technical Deep DiveVMworld
Virtual Volumes provide a more efficient operational model for external storage management in vSphere. They integrate storage capabilities directly into virtual machines at the individual disk level through Storage Policy-Based Management. This simplifies operations by removing the need for static LUN/volume provisioning and allows storage services to be applied non-disruptively on a per-virtual machine basis according to policies. A key component is the VASA Provider, which is used to publish an array's storage capabilities and manage the creation of VM-level objects called Virtual Volumes on behalf of vSphere.
VMworld 2016: The KISS of vRealize Operations! VMworld
This presentation introduces new features in vRealize Operations 6.3 that simplify operations management. It begins with an overview of the vRealize Operations architecture and dashboard. New features are then demonstrated, including a recommended actions page, cluster resource dashboard, data collection notifications, workload balancing through rebalancing containers, guided remediation through alerts, integration with vRealize Log Insight, capacity management of clusters and projections, and extensibility with management packs. Finally, related VMworld sessions are listed that provide further information on capacity planning, troubleshooting, intelligent operations management, log insight, and network insight.
VMworld 2016: Getting Started with PowerShell and PowerCLI for Your VMware En...VMworld
This document provides an overview and introduction to PowerShell and PowerCLI for managing VMware environments. It discusses what PowerShell and PowerCLI are, important terminology like modules and functions, how to set them up and configure profiles, and examples of how to start coding with PowerShell including gathering data, writing logic statements, and using cmdlets safely. The presenters are introduced and an agenda is provided covering these topics at a high level to get started with PowerShell and PowerCLI.
VMworld 2016: Ask the vCenter Server Exerts PanelVMworld
This document is a disclaimer stating that the presentation may include features still under development and not committed to be delivered in final products. Any features discussed are subject to change based on technical feasibility and market demand, and pricing and packaging have not been determined for any new technologies presented. The document is confidential.
VMworld 2016: Virtualize Active Directory, the Right Way! VMworld
Virtualizing Active Directory domain controllers provides benefits like increased availability, scalability, and manageability. However, there are some technical challenges to address like ensuring proper time synchronization. This presentation provides best practices for virtualizing domain controllers including using host-guest affinity rules, disabling time synchronization settings, and ensuring the ESXi host clock is correct. It also introduces new "safety" features in Windows Server 2012 like VM GenerationID that help address issues from restoring or reverting snapshots like USN rollback.
VMworld 2016: Migrating from a hardware based firewall to NSX to improve perf...VMworld
Iain Leiter from A.T. Still University discussed their organization's migration from a hardware-based firewall to NSX to improve performance and compliance. Some key advantages of NSX include distributed firewalling for high performance and scalability, pay-as-you-grow flexibility, and advanced security features like microsegmentation. Their deployment process involved installing NSX, defining security groups, building security policies using syslog data from "recon rules", and applying a common services policy. Discoveries included many backdoors, application architecture issues, and the security benefits of microsegmentation.
VMworld 2015: Troubleshooting for vSphere 6VMworld
The document provides an overview of troubleshooting tools and techniques for vSphere 6. It discusses gathering diagnostic information, identifying potential causes, and resolving problems. The vSphere ESXi Shell and vCLI commands can be used to troubleshoot issues locally or remotely via SSH. An example troubleshooting process is provided to demonstrate defining a vMotion failure problem, gathering logs, testing connectivity, and resolving an incorrect VMkernel interface IP address.
VMworld 2015: Monitoring and Managing Applications with vRealize Operations 6...VMworld
This year VMware vSphere 6 combined with vRealize Operations 6.1 (vR Ops 6) adds critical features to increase technical agility in the infrastructure, and reduce Mean time to Repair. With a new Automated remediation action framework in vR Ops, vSphere 6’s ability to vMotion Physical Raw Device mappings (RDMs), and a complete Management Pack Ecosystem for monitoring Infrastructure to applications, administrators have the tools needed to get to maintain 5 9’s uptime, shorten Mean Time to Repair (MTTR), and predict capacity requirements as and when the business requires.. This session will be a deep technical explanation, and live demonstration of these tools. It will give administrators a solid understanding of how they can use these tools to monitor and manage their application clusters, keep applications running during Infrastructure maintenance, and get deep holistic visibility into the entire Application ecosystem, from Storage to Networking.
VMworld 2015: Advanced SQL Server on vSphereVMworld
Microsoft SQL Server is one of the most widely deployed “apps” in the market today and is used as the database layer for a myriad of applications, ranging from departmental content repositories to large enterprise OLTP systems. Typical SQL Server workloads are somewhat trivial to virtualize; however, business critical SQL Servers require careful planning to satisfy performance, high availability, and disaster recovery requirements. It is the design of these business critical databases that will be the focus of this breakout session. You will learn how build high-performance SQL Server virtual machines through proper resource allocation, database file management, and use of all-flash storage like XtremIO. You will also learn how to protect these critical systems using a combination of SQL Server and vSphere high availability features. For example, did you know you can vMotion shared-disk Windows Failover Cluster nodes? You can in vSphere 6! Finally, you will learn techniques for rapid deployment, backup, and recovery of SQL Server virtual machines using an all-flash array.
VMworld 2015: Virtualize Active Directory, the Right Way!VMworld
Active Directory Domain Services (ADDS) allows organizations to deploy a scalable and secure directory service for managing users, resources and applications. Virtualization of ADDS has been supported for many years now, however has required careful management to avoid pitfalls around replication, time management, and access. Windows Server 2012 provides greater support for virtualization by including virtualization-safe features and support for rapid domain controller deployment.
VMworld 2015: Site Recovery Manager and Policy Based DR Deep Dive with Engine...VMworld
Policy based management greatly simplifies the work of IT Administrators making it easy to ensure that applications and VMs receive the resources, protection and functionality required. Learn about the latest enhancements of Site Recovery Manager in this space, which represent a huge step towards providing policy based DR. In this session we'll dive deep into how this approach works and how to work with them.
VMworld 2015: Building a Business Case for Virtual SANVMworld
This presentation discusses building a business case for VMware Virtual SAN. It provides an overview of Virtual SAN and its benefits for customers like choice, integration, cost savings and performance. A case study is presented of how Dominos Pizza implemented Virtual SAN which resulted in roughly 40% lower costs compared to a traditional storage array. The presentation concludes by demonstrating the Virtual SAN assessment tool and various ways customers can try Virtual SAN.
Not content to simply describe the Virtual Volume (VVOL) framework, this session instead examines practical use cases: How different configurations and workloads benefit from VVOLs. Learn how Storage Policy Based Management (SPBM) couples with VVOLs to provide VM configuration options not previously available. We demonstrate a handful of real-life scenarios, specifically covering how VVOLs benefits oversubscribed systems, disaster recovery preparation and multi-tenant requirements for customers. Specific configuration options and constraints are covered in detail, including how they work with underlying storage.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Pushing the limits of ePRTC: 100ns holdover for 100 days
VMworld 2013: Just Because You Could, Doesn't Mean You Should: Lessons Learned in Storage Best Practices (v2.0)
1. Just Because You Could, Doesn't Mean You Should:
Lessons Learned in Storage Best Practices (v2.0)
Patrick Carmichael, VMware
STO4791
#STO4791
2. 22
Just Because You Could, Doesn’t Mean You SHOULD.
Storage Performance and Technology
Interconnect vs IOP
Command Sizing.
PDU Sizing
SSD vs Magnetic Disk
Tiering
SSD vs Spinning Media, Capacity vs Performance
Thin Provisioning
Architecting for Failure
Planning for the failure from the initial design
Complete Site Failure
Backup RTO
DR RTO
The low-cost DR site
3. 33
Storage Performance – Interconnect vs IOP
Significant advances in interconnect performance
FC 2/4/8/16GB
iSCSI 1G/10G/40G
NFS 1G/10G/40G
Differences in performance between technologies
• None – NFS, iSCSI and FC are effectively interchangeable
• IO Footprint from array perspective.
Despite advances, performance limit is still hit at the media itself
90% of storage performance cases seen by GSS that are not config related,
are media related
Payload (throughput) is fundamentally different from IOP (cmd/s)
IOP performance is always lower than throughput
4. 44
Factors That Affect Performance
Performance versus Capacity
Disk performance does not scale with drive size
Larger drives generally equate lower performance
IOPS(I/Os per second) is crucial
How many IOPS does this number of disks provide?
How many disks are required to achieve a
required number of IOPS?
More spindles generally equals greater performance
5. 55
Factors That Affect Performance - I/O Workload and RAID
Understanding workload is a crucial consideration when designing
for optimal performance
Workload is characterized by IOPS and write % vs read %
Design choice is usually a question of:
How many IOPs can I achieve with a given number of disks?
• Total Raw IOPS = Disk IOPS * Number of disks
• Functional IOPS = (Raw IOPS * Write%)/(Raid Penalty) + (Raw IOPS * Read %)
How many disks are required to achieve a required IOPS value?
• Disks Required = ((Read IOPS) + (Write IOPS*Raid Penalty))/ Disk IOPS
6. 66
IOPS Calculations – Fixed Number of disks
Calculating IOPS for a given number of disks
• 8 x 146GB 15K RPM SAS drives
• ~150 IOPS per disk
• RAID 5
• 150 * 8 = 1200 Raw IOPS
• Workload is 80% Write, 20% Read
• (1200*0.8)/4 + (1200*0.2) = 480 Functional IOPS
Raid Level IOPS(80%Read 20%Write) IOPS(20%Read 80%Write)
5 1020 480
6 1000 400
10 1080 720
7. 77
IOPS Calculations – Minimum IOPS Requirement
Calculating number of disks for a required IOPS value
1200 IOPS required
15K RPM SAS drives. ~150 IOPS per disk
Workload is 80% Write, 20% Read
RAID 5
Disks Required = (240 + (960*4))/150 IOPS
27 Disks required
Raid Level Disks(80%Read 20%Write) Disks (20%Read 80%Write)
5 13 27
6 16 40
10 10 14
8. 88
Storage Performance – Interconnect vs IOP – contd.
While still true for 90% of workloads, this is starting to shift
SCSI commands passed at native sizes
• Windows 2k8/2k12 (8MB copy)
• Linux (native applications, up to 8MB+)
• NoSQL Databases (Especially Key/Value or Document stores)
• Cassandra (128k)
• Mongo (64-256k)
ESX will gladly pass whatever size command the guest requests.
New applications are starting to run into some protocol limitations
• iSCSI PDU (64k/256k)
• NFS Buffer/Request Size
• Write Amplification!
Offloading these processes offers significant performance
improvements
10. 1010
Tiering
Very old concept
• Primarily a manual process in the past
• Based on RAID groups / Disk types
• Second generation arrays offer batch mode / scheduled auto-movement
system of blocks under IO pressure
12. 1212
Tiering – contd.
New arrays from several partners now offer near real-time
block movement
• Dot Hill AssuredSAN
• X-IO ISE2
Or, batch mode
• Compellent
• VNX
Auto-tiering will get you better performance, faster
• ~2 hours for 150G working set to be fully moved to SSD.
However…
13. 1313
Tiering – contd.
Remember – Just because you could, doesn’t mean you should!
There are several places where an auto-tiering array may not be
able to help
• Working Set
• Total Data > SSD Space
• Second tier of cache to worry about
• Fully utilized
• May not have spare IOPS with which to move data between layers.
• Rapidly changing datasets
• VDI
• In-memory databases.
• Often hard to predict when they’re fully utilized
• “Never ending” performance bucket.
14. 1414
Tiering – contd.
These platforms have amazing and very capable technology
Have the vendor show you, with your workload, what their platform
is capable of
Workloads differ massively between users and vendors – validate!
15. 1515
SSD vs Spinning Disk
SSDs offer extremely high performance per single “spindle”
• Exceptionally great for “L2 Cache”, especially on auto-tiering arrays.
• Capable of 1,000,000+ IOPS in pure SSD arrays
• Exceptionally dense consolidation ratios (Excellent for VDI/HPC).
However…
Ultra-high performance arrays have other limitations
• Surrounding infrastructure
• Switching (packets/frames a second)
• CPU (packaging frames in software initiators)
• Command sizing
• Some SSDs do not work well outside of certain command sizes
• Architectural limitations
• Overhead loss for scratch space
• Garbage Collect
16. 1616
VMFS/VMDK Updates for the Next Release
VMFS5 (ESX5) introduced support for LUNs larger than 2TB
• Still limited to 2TB VMDK, regardless of LUN sizing
• 64TB RDM
Major update in this release
• Max size: 64TB for both VMDK and VMFS volumes.
• Lets you avoid spanning disks inside of a guest, or using RDMs for large
workloads
17. 1717
VMDK Contd.
VMFS5, in combination with ATS, will
allow consolidation of ever-larger
number of VMs onto single VMFS
datastores
Alternatively, a couple of very large
VMs can occupy a single lun
• While potentially easier for management, this
means that significantly more data is now
stored in one location.
• When defining a problem space, you’ve now
expanded greatly the amount of data stored
in one place.
Just because you could, does that
mean you should?
18. 1818
Thin Provisioning
Thin provisioning offers very unique opportunities to manage your
storage “after” provisioning
• Workloads that require a certain amount of space, but don’t actually use it
• Workloads that may grow over time, and can be managed/moved if they do
• Providing space for disparate groups that have competing requirements
Many arrays ~only~ offer Thin Provisioning, by design
The question is, where should you thin provision, and what are the
ramifications?
19. 1919
Thin Provisioning – VM Level
VM disk is created @ 0b, until used, and then grows @ VMFS Block
size as needed
Minor performance penalty for provisioning / zeroing
Disk cannot currently be shrunk – once grown, it stays grown
• There are some workarounds for this, but they are not guaranteed
What happens when you finally run out of space?
• All VMs stun until space is created
• Production impacting, but potentially a quick fix (shut down VMs, adjust
memory reservation, etc)
• Extremely rare to see data loss of any kind
20. 2020
Thin Provisioning – LUN Level
Allows your array to seem bigger than it actually is
Allows you to share resources between groups (the whole goal
of virtualization)
Some groups may not use all or much of what they’re allocated,
allowing you to utilize the space they’re not using
Standard sized or process defined luns may waste significant
amounts of space, and space being wasted is $$ being wasted
Significant CapEX gains can be seen with thin luns
21. 2121
Thin Provisioning – LUN Level - contd
What happens when you finally run out of space?
• New VAAI T10 primitives, for compatible arrays, will let ESX know that the
underlying storage has no free blocks (related to NULL_UNMAP)
• If VAAI works, and your array is compatible, and you’re on a supported version of
ESX, this will result in the same as a thin VMDK running out of space – All VMs will
stun (that are waiting on blocks). VMs not waiting on blocks will continue as normal.
• Cleanup will require finding additional space on the array, as VSWP files / etc will
already be allocated blocks at the lun level. Depending on your utilization, this may
not be possible, unless UNMAP also works (very limited support at this time).
• If NULL_UNMAP is not available for your environment, or does not work
correctly, then what?
• On a good day, the VMs will simply crash with a write error, or the application inside
will fail (depends on array and how it handles a full filesystem).
• Databases and the like are worst affected, will most likely require rebuild/repair.
• And on a bad day?
23. 2323
Thin Provisioning
There are many reasons to use Thin Provisioning, at both the VM
and the LUN level
Thin provisioning INCREASES the management workload of
maintaining your environment. You cannot just ignore it
24. 2424
Freeing space on thin provisioned LUNs
ESX 5.0 introduced NULL_UNMAP, a T10 command to mark blocks
as “freed” for an underlying storage device.
• Disabled in P01 due to major performance issues
• Re-enabled in U1, manual only, to prevent sudden spikes in load, or corruption
from array firmware issues.
Vmkfstools –y command, set amount of space to try and reclaim.
*VALIDATE IF VMFSTOOLS / Auto unmap made it into 5.5*
25. 2525
Just Because You Could, Doesn’t Mean You Should
Everything we’ve covered so far is based on new technologies
What about the existing environments?
The ultimate extension of “Just because you could, doesn’t mean
you should,” is what I call “Architecting for Failure”
26. 2626
Architecting for Failure
The ultimate expression of “Just because you could, doesn’t mean
you should.”
• Many core infrastructure designs are built with tried and true hardware and
software, and people assume that things will always work
• We all know this isn’t true – Murphy’s law
Architect for the failure.
• Consider all of your physical infrastructure.
• If any component failed, how would you recover?
• If everything failed, how would you recover?
• Consider your backup/DR plan as well
27. 2727
Black Box Testing
Software engineering concept
• Consider your design, all of the inputs, and all of the expected outputs
• Feed it good entries, bad entries, and extreme entries, find the result, and
make sure it is sane
• If not, make it sane
This can be applied before, or after, you build your environment
28. 2828
Individual Component Failure
Black box: consider each step your IO takes, from VM to physical
media
Physical hardware components are generally easy to compensate
for
• VMware HA and FT both make it possible for a complete system to fail with
little/no downtime to the guests in question
• Multiple hardware components add redundancy and eliminate single points of
failure
• Multiple NICs
• Multiple storage paths
• Traditional hardware (multiple power supplies, etc)
• Even with all of this, many environments are not taking advantage of these
features
Sometimes, the path that IO takes passes through a single point of
failure that you don’t realize is one
29. 2929
What About a Bigger Problem?
You’re considering all the different ways to make sure individual
components don’t ruin your day
What if your problem is bigger?
30. 3030
Temporary Total Site Loss
Consider your entire infrastructure during a temporary complete
failure
• What would happen if you had to bootstrap it cold?
• This actually happens more often than would be expected
• Hope for the best, prepare for the worst
• Consider what each component relies on – do you have any circular dependencies?
• Also known as the “Chicken and the Egg” problem, these can increase your
RTO significantly
• Example: Storage mounted via DNS, all DNS servers on the same storage
devices. Restoring VC from backup when all networking is via DVS
• What steps are required to bring your environment back to life?
• How long will it take? Is that acceptable?
31. 3131
Permanent Site Loss
Permanent site loss is not always an “Act of God” type event
• Far more common is a complete loss of a major, critical component
• Site infrastructure (power, networking, etc) may be intact, but your data is not
• Array failures (controller failure, major filesystem corruption, RAID failure)
• Array disaster (thermal event, fire, malice)
• Human error – yes, it happens!
• Multiple recovery options – which do you have?
• Backups
• Tested and verified?
• What’s your RTO for a full failure?
• Array based replication
• Site Recovery Manager
• Manual DR
• How long for a full failover?
• Host based replication
32. 3232
Permanent Site Loss – contd.
Consider the RTO of your choice of disaster recovery technology
• It equates directly to the amount of time you will be without your virtual
machines
• How long can you, and your business, be without those services?
• A perfectly sound backup strategy is useless, if it cannot return you to
operation quickly enough
Architect for the Failure – make sure every portion of your
environment can withstand a total failure, and recovery is possible
in a reasonable amount of time
33. 3333
Let’s Be Honest for a Moment…
We would all love to have a DR site that is equally capable as your
production site.
• I’d also like a horsey. Neither of us have all the budget we’d really like for that.
There’s a key DR strategy that is often missed: Data existence is
far more important than full operation
• Many places decline to implement a DR strategy due to budget, time, or
believing that they have to build an equal DR site as their production
environment.
• For many of these places, having key services ~only~ online, with the rest of
the data being protected at the DR site and present if needed, is more than
enough.
34. 3434
How to Do It On a Budget of Almost $0
VMware GSS set out to prove that you can build a “protection” site
with a lot less than you’d expect
Services required were true tier 1 apps:
• Store front end (based on DVD Store)
• Exchange
• SQL
• Oracle RAC
• JVM
35. 3535
Hardware
8x Dell D630 Laptops (4g, 150G internal drives)
2x Dell Inspiron 5100 Laptops (8g, 200G internal drives)
4x Dell Precision 390 Workstations (8G/16G, 5x 1TB internal drives)
1 Intel Atom Supermicro (2G / 5x1TB Internal Drives)
• Found by the side of the road
• No joke
1 Dumb Ethernet Switch (Netgear), purchased from Best Buy
Duct tape, 1 roll (boot drive for one of the 390s is taped to the
chassis)
4x Broadcom dual-port 1G cards (borrowed from R710 servers)
36. 3636
Software
ESX 5.1 U1
SRM 5.1.1
Ubuntu + knfsd (low performance storage)
• Host based replication target
Datacore San Symphony V
• 4x Nodes, high performance, array based replication
Starwind iSCSI
• 2x Nodes, high performance, management software.
Lefthand VSA
• 4x Nodes, medium/high performance, array based replication
37. 3737
End results
While we were certainly unable to bring up the full environment on
the DR site, we were able to replicate all 6TB of data successfully.
Critical services for managing a disaster-caused outage were
online
• 1/5 of the Exchange mailboxes (executives and IT personnel)
• Web front end
• SQL database for minimal customer queries/interactions
• No major processing
• RAC
• Online for queries, no processing
• JVM
• Offline (RAC processing engine).
This certainly isn’t ideal, but everything was intact, we could
“manage” the situation, and the “company” hadn’t vanished.
40. 4040
Conclusion
You can make a DR site out of almost anything
• Retired servers
• Spare parts
• Out of warranty equipment
You don’t need high end equipment – hardware a generation or two
old is more than enough for basic, or even partial operation
Existing software solutions make it very easy to turn servers alone
into extremely capable shared storage solutions
42. 4242
Other VMware Activities Related to This Session
HOL:
HOL-SDC-1304
vSphere Performance Optimization
Group Discussions:
STO1000-GD
Overall Storage with Patrick Carmichael