SharePoint Performance: Best Practices from the Field by Jason Himmelstein - ...SPTechCon
This document provides contact information for Jason Himmelstein and links to resources about SharePoint performance tuning and troubleshooting. It also includes recommendations for hardware configurations for small, medium, and large SharePoint farms and guidelines for optimizing database files through appropriate RAID levels.
The document discusses different components used in building Taobao storefronts including HTML, CSS, JavaScript, and PHP as well as frameworks like seed.js and tb.js. It also outlines the structure of CSS including base, head, and skin files. Key elements like regions, boxes, and reflow are explained at a high level.
confessions of a dba: worst and best things I've done in production - Open So...emilyslocombe
Emily Slocombe shares her experiences as a DBA over 15 years, including times where she broke things in production. Some highlights include using the Blackhole storage engine to offload replication traffic, a week-long data migration with no sleep sustained on energy drinks, and mistakes made with RAID configurations and MySQL tuning scripts. The overall message is that breaking things can lead to learning opportunities for DBAs.
This document discusses using PowerShell to manage SharePoint. It provides an overview of PowerShell and what it can do, including managing servers and SharePoint. It demonstrates PowerShell usage through examples like monitoring server status, importing/exporting files, and adding an announcement to a SharePoint site. The document encourages use of PowerShell for automating administrative tasks on Microsoft platforms.
The document discusses Microsoft System Center 2012 and its capabilities for cloud and datacenter management. It highlights products within System Center 2012 like Operations Manager, Virtual Machine Manager, Service Manager, and Configuration Manager that provide monitoring, virtualization management, service desk, and client management capabilities. The document also discusses how System Center 2012 allows for unified management of physical, virtual, and cloud environments from a single console.
SharePoint Administration with PowerShellEric Kraus
Why limit yourself to STSADM? Discover the power of PowerShell 2.0 as it is used to perform advanced administrative & development tasks. This session will start with a brief introduction to PowerShell scripting and continue with a look into helpful SharePoint scripts including: filtering event and ULS logs, managing sites and users, streamlining feature development, working with the object model, and much more! Both administrators and developers will benefit from this powerful discussion.
This document summarizes Microsoft's Windows Azure cloud platform and Sentri's partnership with Microsoft. It highlights how Sentri has won multiple Microsoft Partner of the Year awards for its work innovating cloud solutions using Windows Azure. It also provides an overview of Sentri's cloud consulting services for envisioning, proof of concept testing, and deploying hybrid cloud solutions using Windows Azure and System Center.
SharePoint Performance: Best Practices from the Field by Jason Himmelstein - ...SPTechCon
This document provides contact information for Jason Himmelstein and links to resources about SharePoint performance tuning and troubleshooting. It also includes recommendations for hardware configurations for small, medium, and large SharePoint farms and guidelines for optimizing database files through appropriate RAID levels.
The document discusses different components used in building Taobao storefronts including HTML, CSS, JavaScript, and PHP as well as frameworks like seed.js and tb.js. It also outlines the structure of CSS including base, head, and skin files. Key elements like regions, boxes, and reflow are explained at a high level.
confessions of a dba: worst and best things I've done in production - Open So...emilyslocombe
Emily Slocombe shares her experiences as a DBA over 15 years, including times where she broke things in production. Some highlights include using the Blackhole storage engine to offload replication traffic, a week-long data migration with no sleep sustained on energy drinks, and mistakes made with RAID configurations and MySQL tuning scripts. The overall message is that breaking things can lead to learning opportunities for DBAs.
This document discusses using PowerShell to manage SharePoint. It provides an overview of PowerShell and what it can do, including managing servers and SharePoint. It demonstrates PowerShell usage through examples like monitoring server status, importing/exporting files, and adding an announcement to a SharePoint site. The document encourages use of PowerShell for automating administrative tasks on Microsoft platforms.
The document discusses Microsoft System Center 2012 and its capabilities for cloud and datacenter management. It highlights products within System Center 2012 like Operations Manager, Virtual Machine Manager, Service Manager, and Configuration Manager that provide monitoring, virtualization management, service desk, and client management capabilities. The document also discusses how System Center 2012 allows for unified management of physical, virtual, and cloud environments from a single console.
SharePoint Administration with PowerShellEric Kraus
Why limit yourself to STSADM? Discover the power of PowerShell 2.0 as it is used to perform advanced administrative & development tasks. This session will start with a brief introduction to PowerShell scripting and continue with a look into helpful SharePoint scripts including: filtering event and ULS logs, managing sites and users, streamlining feature development, working with the object model, and much more! Both administrators and developers will benefit from this powerful discussion.
This document summarizes Microsoft's Windows Azure cloud platform and Sentri's partnership with Microsoft. It highlights how Sentri has won multiple Microsoft Partner of the Year awards for its work innovating cloud solutions using Windows Azure. It also provides an overview of Sentri's cloud consulting services for envisioning, proof of concept testing, and deploying hybrid cloud solutions using Windows Azure and System Center.
SharePoint 2010 enables people to work together, share information & manage content! But if your SharePoint environment is not set up properly, you may impact the overall performance and ultimately the satisfaction of your users.
2014-02-22 - IT Pro Camp - SharePoint 2013, A Brief Overview of CapabilityDan Usher
So SharePoint 2013 has been out for just over a year now with Release to Manufacture in October 2012. Interesting in learning what’s new and different? Then come and learn more about new capabilities in the product such as Shredded Storage and Distributed Caching among others as well as how the migration story changes for SharePoint 2013.
Development to Production with Sharded MongoDB ClustersSeveralnines
Severalnines presentation at MongoDB Stockholm Conference.
Presentation covers:
- mongoDB sharding/clustering concepts
- recommended dev/test/prod setups
- how to verify your deployment
- how to avoid downtime
- what MongoDB metrics to watch
- when to scale
This document discusses various performance-related topics in SharePoint including latency, throughput, resource throttling, monitoring, and hardware requirements. It provides definitions of latency and throughput. It discusses tools for monitoring like the SharePoint Log Viewer. It also lists minimum hardware requirements for SharePoint 2010 and SQL Server.
This document summarizes a presentation about Ceph, an open-source distributed storage system. It discusses Ceph's introduction and components, benchmarks Ceph's block and object storage performance on Intel architecture, and describes optimizations like cache tiering and erasure coding. It also outlines Intel's product portfolio in supporting Ceph through optimized CPUs, flash storage, networking, server boards, software libraries, and contributions to the open source Ceph community.
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Odinot Stanislas
Après la petite intro sur le stockage distribué et la description de Ceph, Jian Zhang réalise dans cette présentation quelques benchmarks intéressants : tests séquentiels, tests random et surtout comparaison des résultats avant et après optimisations. Les paramètres de configuration touchés et optimisations (Large page numbers, Omap data sur un disque séparé, ...) apportent au minimum 2x de perf en plus.
23 October 2013 - AWS 201 - A Walk through the AWS Cloud: Introduction to Ama...Amazon Web Services
Amazon Redshift is the new data warehouse service from Amazon Web Services. Redshift offers you fast query performance when analyzing data sets from a few hundred gigabytes to over a petabyte at a fraction of the cost of traditional solutions. In this webinar, we will take a detailed look at Redshift, including a live demonstration. This webinar is ideal for anyone looking to gain deeper insight into their data, without the usual challenges of time, cost and effort.
Paris Spark Meetup Oct 26, 2015 - Spark After Dark v1.5 - Best of Advanced Ap...Chris Fregly
* Title *
Spark After Dark 1.5: Deep Dive Into Latest Perf and Scale Improvements in Spark Ecosystem
* Abstract *
Combining the most popular and technically-deep material from his wildly popular Advanced Apache Spark Meetup, Chris Fregly will provide code-level deep dives into the latest performance and scalability advancements within the Apache Spark Ecosystem by exploring the following:
1) Building a Scalable and Performant Spark SQL/DataFrames Data Source Connector such as Spark-CSV, Spark-Cassandra, Spark-ElasticSearch, and Spark-Redshift
2) Speeding Up Spark SQL Queries using Partition Pruning and Predicate Pushdowns with CSV, JSON, Parquet, Avro, and ORC
3) Tuning Spark Streaming Performance and Fault Tolerance with KafkaRDD and KinesisRDD
4) Maintaining Stability during High Scale Streaming Ingestion using Approximations and Probabilistic Data Structures from Spark, Redis, and Twitter's Algebird
5) Building Effective Machine Learning Models using Feature Engineering, Dimension Reduction, and Natural Language Processing with MLlib/GraphX, ML Pipelines, DIMSUM, Locality Sensitive Hashing, and Stanford's CoreNLP
6) Tuning Core Spark Performance by Acknowledging Mechanical Sympathy for the Physical Limitations of OS and Hardware Resources such as CPU, Memory, Network, and Disk with Project Tungsten, Asynchronous Netty, and Linux epoll
* Demos *
This talk features many interesting and audience-interactive demos - as well as code-level deep dives into many of the projects listed above.
All demo code is available on Github at the following link: https://github.com/fluxcapacitor/pipeline/wiki
In addition, the entire demo environment has been Dockerized and made available for download on Docker Hub at the following link: https://hub.docker.com/r/fluxcapacitor/pipeline/
* Speaker Bio *
Chris Fregly is a Principal Data Solutions Engineer for the newly-formed IBM Spark Technology Center, an Apache Spark Contributor, a Netflix Open Source Committer, as well as the Organizer of the global Advanced Apache Spark Meetup and Author of the Upcoming Book, Advanced Spark.
Previously, Chris was a Data Solutions Engineer at Databricks and a Streaming Data Engineer at Netflix.
When Chris isn’t contributing to Spark and other open source projects, he’s creating book chapters, slides, and demos to share knowledge with his peers at meetups and conferences throughout the world.
The document discusses stackArmor's presentation on using MongoDB and Solr on AWS for a customer's ETL processing needs. The customer receives large amounts of data monthly and needed to process it faster at lower cost. StackArmor designed a solution using MongoDB shards, SolrCloud, and Chef automation to meet the customer's goals of processing 5 million records per hour within budget and compliance requirements. Through tuning the design over time, they were able to improve throughput to over 3 million records per hour sustained.
DevOps for ETL processing at scale with MongoDB, Solr, AWS and ChefGaurav "GP" Pal
Large scale data processing for Extract Transform and Loading (ETL) jobs is a very common practice. The stackArmor DevOps team developed a Chef based automation solution to automate the AWS environment provisioning, code deployment and data ingestion processing to ingest and process over 2 TB of Data.
This presentation covers the technologies used, the planning phase, AWS instance selection and optimizing the ETL processing for not only performance but also cost.
The target was to process 500 million rows within 72 hours with a processing rate of 5 million transactions per hour.
The presentation also provides pitfalls and automation optimizations performed to accomplish the targeted processing rates.
The presentation was delivered at the DevOpsDC Meetup on May 17, 2016
Taking the Performance of your Data Warehouse to the Next Level with Amazon R...Amazon Web Services
Amazon Redshift gives you fast SQL query performance on large data sets. We will discuss optimisation from end to end, all the way from loading through to querying to ensure your end users get the data they need, when they need it.
Speaker: Russell Nash, Solutions Architect, Amazon Web Services
Featured Customer - Domain
Sql Health in a SharePoint environmentEnrique Lima
This document discusses how to maintain a healthy SharePoint environment. It emphasizes the importance of properly configuring and managing the SQL Server database that SharePoint runs on. It provides guidance on capacity planning, hardware sizing, maintenance best practices, and understanding SharePoint limitations and thresholds. The goal is to ensure the SQL Server infrastructure can support the SharePoint implementation and meet performance requirements.
Brussels Spark Meetup Oct 30, 2015: Spark After Dark 1.5: Real-time, Advanc...Chris Fregly
Combining the most popular and technically-deep material from his wildly popular Advanced Apache Spark Meetup, Chris Fregly will provide code-level deep dives into the latest performance and scalability advancements within the Apache Spark Ecosystem by exploring the following:
1) Building a Scalable and Performant Spark SQL/DataFrames Data Source Connector such as Spark-CSV, Spark-Cassandra, Spark-ElasticSearch, and Spark-Redshift
2) Speeding Up Spark SQL Queries using Partition Pruning and Predicate Pushdowns with CSV, JSON, Parquet, Avro, and ORC
3) Tuning Spark Streaming Performance and Fault Tolerance with KafkaRDD and KinesisRDD
4) Maintaining Stability during High Scale Streaming Ingestion using Approximations and Probabilistic Data Structures from Spark, Redis, and Twitter's Algebird
5) Building Effective Machine Learning Models using Feature Engineering, Dimension Reduction, and Natural Language Processing with MLlib/GraphX, ML Pipelines, DIMSUM, Locality Sensitive Hashing, and Stanford's CoreNLP
6) Tuning Core Spark Performance by Acknowledging Mechanical Sympathy for the Physical Limitations of OS and Hardware Resources such as CPU, Memory, Network, and Disk with Project Tungsten, Asynchronous Netty, and Linux epoll
* Demos *
This talk features many interesting and audience-interactive demos - as well as code-level deep dives into many of the projects listed above.
All demo code is available on Github at the following link: https://github.com/fluxcapacitor/pipeline/wiki
In addition, the entire demo environment has been Dockerized and made available for download on Docker Hub at the following link: https://hub.docker.com/r/fluxcapacitor/pipeline/
* Speaker Bio *
Chris Fregly is a Principal Data Solutions Engineer for the newly-formed IBM Spark Technology Center, an Apache Spark Contributor, a Netflix Open Source Committer, as well as the Organizer of the global Advanced Apache Spark Meetup and Author of the Upcoming Book, Advanced Spark.
Previously, Chris was a Data Solutions Engineer at Databricks and a Streaming Data Engineer at Netflix.
When Chris isn’t contributing to Spark and other open source projects, he’s creating book chapters, slides, and demos to share knowledge with his peers at meetups and conferences throughout the world.
Know thy cost (or where performance problems lurk)Oren Eini
Performance happens. Whether you're designed for it or not it doesn’t matter, she is always invited to the party (and you better find her in a good mood). Knowing the cost of every operation, and how it distributes on every subsystem will ensure that when you are building that proof-of-concept (that always ends up in production) or designing the latest’s enterprise-grade application; you will know where those pesky performance bugs like to inhabit. In this session, we will go deep into the inner working of every performance sensitive subsystem. From the relative safety of the client to the binary world of Voron.
Hadoop Hardware @Twitter: Size does matter.Michael Zhang
This document discusses Twitter's experience scaling their Hadoop clusters and evaluating different hardware configurations. It describes how Twitter developed the "Twitter Hadoop Server" (THS) to optimize for different workloads like backups, processing, and cold storage. The THS uses single-socket servers with fewer disks optimized for cost and density. Testing showed the THS outperformed baseline dual-socket servers on processing benchmarks while providing higher storage density and lower costs per node. The document concludes that for large clusters, specialized hardware can improve performance and efficiency compared to a single size-fits-all approach.
Effective SharePoint Architecture - SharePoint Saturday Stockholm 2016Alistair Pugin
The document discusses effective SharePoint architecture and provides recommendations for server roles, hardware specifications, and database configuration. It recommends a farm architecture with two web front end servers, two application servers, and a SQL Server 2012 cluster with two nodes. The web front end and application servers should each have 16GB RAM, 4 CPU cores, and two hard drives. SQL servers should have 32GB RAM, 8 CPU cores, and five hard drives configured for data, tempdb, and backup files. It also provides tips for securing SharePoint such as implementing firewalls, running security analyzers, and following hardening guidance.
The document discusses Ceph performance on all-flash storage systems. It describes optimizations made to Ceph's OSD architecture and write path that have led to significant performance improvements when deployed on SanDisk's InfiniFlash all-flash storage. These include reducing CPU utilization and improving throughput and latency. Example performance metrics are provided showing random read IOPS over 1.5M and latency under 5ms for most operations. The document also outlines the InfiniFlash hardware architecture and roadmap for further Ceph optimizations including new storage backends like BlueStore.
The document discusses Ceph storage performance on all-flash storage systems. It describes how SanDisk optimized Ceph for all-flash environments by tuning the OSD to handle the high performance of flash drives. The optimizations allowed over 200,000 IOPS per OSD using 12 CPU cores. Testing on SanDisk's InfiniFlash storage system showed it achieving over 1.5 million random read IOPS and 200,000 random write IOPS at 64KB block size. Latency was also very low, with 99% of operations under 5ms for reads. The document outlines reference configurations for the InfiniFlash system optimized for small, medium and large workloads.
Uncover the Benefits of Office 365 & Windows AzureSentri
Polycom has acquired Sentri. The acquisition will allow Polycom to expand its offerings beyond video equipment to include solution capabilities. Sentri's technology and team will help Polycom grow its business.
The document announces that Sentri has been acquired by Polycom. It discusses Polycom's financial strength, strategic partnerships, customer momentum, and industry leadership in video conferencing. The acquisition of Sentri will help Polycom expand its offerings from equipment to solution capabilities.
SharePoint 2010 enables people to work together, share information & manage content! But if your SharePoint environment is not set up properly, you may impact the overall performance and ultimately the satisfaction of your users.
2014-02-22 - IT Pro Camp - SharePoint 2013, A Brief Overview of CapabilityDan Usher
So SharePoint 2013 has been out for just over a year now with Release to Manufacture in October 2012. Interesting in learning what’s new and different? Then come and learn more about new capabilities in the product such as Shredded Storage and Distributed Caching among others as well as how the migration story changes for SharePoint 2013.
Development to Production with Sharded MongoDB ClustersSeveralnines
Severalnines presentation at MongoDB Stockholm Conference.
Presentation covers:
- mongoDB sharding/clustering concepts
- recommended dev/test/prod setups
- how to verify your deployment
- how to avoid downtime
- what MongoDB metrics to watch
- when to scale
This document discusses various performance-related topics in SharePoint including latency, throughput, resource throttling, monitoring, and hardware requirements. It provides definitions of latency and throughput. It discusses tools for monitoring like the SharePoint Log Viewer. It also lists minimum hardware requirements for SharePoint 2010 and SQL Server.
This document summarizes a presentation about Ceph, an open-source distributed storage system. It discusses Ceph's introduction and components, benchmarks Ceph's block and object storage performance on Intel architecture, and describes optimizations like cache tiering and erasure coding. It also outlines Intel's product portfolio in supporting Ceph through optimized CPUs, flash storage, networking, server boards, software libraries, and contributions to the open source Ceph community.
Ceph: Open Source Storage Software Optimizations on Intel® Architecture for C...Odinot Stanislas
Après la petite intro sur le stockage distribué et la description de Ceph, Jian Zhang réalise dans cette présentation quelques benchmarks intéressants : tests séquentiels, tests random et surtout comparaison des résultats avant et après optimisations. Les paramètres de configuration touchés et optimisations (Large page numbers, Omap data sur un disque séparé, ...) apportent au minimum 2x de perf en plus.
23 October 2013 - AWS 201 - A Walk through the AWS Cloud: Introduction to Ama...Amazon Web Services
Amazon Redshift is the new data warehouse service from Amazon Web Services. Redshift offers you fast query performance when analyzing data sets from a few hundred gigabytes to over a petabyte at a fraction of the cost of traditional solutions. In this webinar, we will take a detailed look at Redshift, including a live demonstration. This webinar is ideal for anyone looking to gain deeper insight into their data, without the usual challenges of time, cost and effort.
Paris Spark Meetup Oct 26, 2015 - Spark After Dark v1.5 - Best of Advanced Ap...Chris Fregly
* Title *
Spark After Dark 1.5: Deep Dive Into Latest Perf and Scale Improvements in Spark Ecosystem
* Abstract *
Combining the most popular and technically-deep material from his wildly popular Advanced Apache Spark Meetup, Chris Fregly will provide code-level deep dives into the latest performance and scalability advancements within the Apache Spark Ecosystem by exploring the following:
1) Building a Scalable and Performant Spark SQL/DataFrames Data Source Connector such as Spark-CSV, Spark-Cassandra, Spark-ElasticSearch, and Spark-Redshift
2) Speeding Up Spark SQL Queries using Partition Pruning and Predicate Pushdowns with CSV, JSON, Parquet, Avro, and ORC
3) Tuning Spark Streaming Performance and Fault Tolerance with KafkaRDD and KinesisRDD
4) Maintaining Stability during High Scale Streaming Ingestion using Approximations and Probabilistic Data Structures from Spark, Redis, and Twitter's Algebird
5) Building Effective Machine Learning Models using Feature Engineering, Dimension Reduction, and Natural Language Processing with MLlib/GraphX, ML Pipelines, DIMSUM, Locality Sensitive Hashing, and Stanford's CoreNLP
6) Tuning Core Spark Performance by Acknowledging Mechanical Sympathy for the Physical Limitations of OS and Hardware Resources such as CPU, Memory, Network, and Disk with Project Tungsten, Asynchronous Netty, and Linux epoll
* Demos *
This talk features many interesting and audience-interactive demos - as well as code-level deep dives into many of the projects listed above.
All demo code is available on Github at the following link: https://github.com/fluxcapacitor/pipeline/wiki
In addition, the entire demo environment has been Dockerized and made available for download on Docker Hub at the following link: https://hub.docker.com/r/fluxcapacitor/pipeline/
* Speaker Bio *
Chris Fregly is a Principal Data Solutions Engineer for the newly-formed IBM Spark Technology Center, an Apache Spark Contributor, a Netflix Open Source Committer, as well as the Organizer of the global Advanced Apache Spark Meetup and Author of the Upcoming Book, Advanced Spark.
Previously, Chris was a Data Solutions Engineer at Databricks and a Streaming Data Engineer at Netflix.
When Chris isn’t contributing to Spark and other open source projects, he’s creating book chapters, slides, and demos to share knowledge with his peers at meetups and conferences throughout the world.
The document discusses stackArmor's presentation on using MongoDB and Solr on AWS for a customer's ETL processing needs. The customer receives large amounts of data monthly and needed to process it faster at lower cost. StackArmor designed a solution using MongoDB shards, SolrCloud, and Chef automation to meet the customer's goals of processing 5 million records per hour within budget and compliance requirements. Through tuning the design over time, they were able to improve throughput to over 3 million records per hour sustained.
DevOps for ETL processing at scale with MongoDB, Solr, AWS and ChefGaurav "GP" Pal
Large scale data processing for Extract Transform and Loading (ETL) jobs is a very common practice. The stackArmor DevOps team developed a Chef based automation solution to automate the AWS environment provisioning, code deployment and data ingestion processing to ingest and process over 2 TB of Data.
This presentation covers the technologies used, the planning phase, AWS instance selection and optimizing the ETL processing for not only performance but also cost.
The target was to process 500 million rows within 72 hours with a processing rate of 5 million transactions per hour.
The presentation also provides pitfalls and automation optimizations performed to accomplish the targeted processing rates.
The presentation was delivered at the DevOpsDC Meetup on May 17, 2016
Taking the Performance of your Data Warehouse to the Next Level with Amazon R...Amazon Web Services
Amazon Redshift gives you fast SQL query performance on large data sets. We will discuss optimisation from end to end, all the way from loading through to querying to ensure your end users get the data they need, when they need it.
Speaker: Russell Nash, Solutions Architect, Amazon Web Services
Featured Customer - Domain
Sql Health in a SharePoint environmentEnrique Lima
This document discusses how to maintain a healthy SharePoint environment. It emphasizes the importance of properly configuring and managing the SQL Server database that SharePoint runs on. It provides guidance on capacity planning, hardware sizing, maintenance best practices, and understanding SharePoint limitations and thresholds. The goal is to ensure the SQL Server infrastructure can support the SharePoint implementation and meet performance requirements.
Brussels Spark Meetup Oct 30, 2015: Spark After Dark 1.5: Real-time, Advanc...Chris Fregly
Combining the most popular and technically-deep material from his wildly popular Advanced Apache Spark Meetup, Chris Fregly will provide code-level deep dives into the latest performance and scalability advancements within the Apache Spark Ecosystem by exploring the following:
1) Building a Scalable and Performant Spark SQL/DataFrames Data Source Connector such as Spark-CSV, Spark-Cassandra, Spark-ElasticSearch, and Spark-Redshift
2) Speeding Up Spark SQL Queries using Partition Pruning and Predicate Pushdowns with CSV, JSON, Parquet, Avro, and ORC
3) Tuning Spark Streaming Performance and Fault Tolerance with KafkaRDD and KinesisRDD
4) Maintaining Stability during High Scale Streaming Ingestion using Approximations and Probabilistic Data Structures from Spark, Redis, and Twitter's Algebird
5) Building Effective Machine Learning Models using Feature Engineering, Dimension Reduction, and Natural Language Processing with MLlib/GraphX, ML Pipelines, DIMSUM, Locality Sensitive Hashing, and Stanford's CoreNLP
6) Tuning Core Spark Performance by Acknowledging Mechanical Sympathy for the Physical Limitations of OS and Hardware Resources such as CPU, Memory, Network, and Disk with Project Tungsten, Asynchronous Netty, and Linux epoll
* Demos *
This talk features many interesting and audience-interactive demos - as well as code-level deep dives into many of the projects listed above.
All demo code is available on Github at the following link: https://github.com/fluxcapacitor/pipeline/wiki
In addition, the entire demo environment has been Dockerized and made available for download on Docker Hub at the following link: https://hub.docker.com/r/fluxcapacitor/pipeline/
* Speaker Bio *
Chris Fregly is a Principal Data Solutions Engineer for the newly-formed IBM Spark Technology Center, an Apache Spark Contributor, a Netflix Open Source Committer, as well as the Organizer of the global Advanced Apache Spark Meetup and Author of the Upcoming Book, Advanced Spark.
Previously, Chris was a Data Solutions Engineer at Databricks and a Streaming Data Engineer at Netflix.
When Chris isn’t contributing to Spark and other open source projects, he’s creating book chapters, slides, and demos to share knowledge with his peers at meetups and conferences throughout the world.
Know thy cost (or where performance problems lurk)Oren Eini
Performance happens. Whether you're designed for it or not it doesn’t matter, she is always invited to the party (and you better find her in a good mood). Knowing the cost of every operation, and how it distributes on every subsystem will ensure that when you are building that proof-of-concept (that always ends up in production) or designing the latest’s enterprise-grade application; you will know where those pesky performance bugs like to inhabit. In this session, we will go deep into the inner working of every performance sensitive subsystem. From the relative safety of the client to the binary world of Voron.
Hadoop Hardware @Twitter: Size does matter.Michael Zhang
This document discusses Twitter's experience scaling their Hadoop clusters and evaluating different hardware configurations. It describes how Twitter developed the "Twitter Hadoop Server" (THS) to optimize for different workloads like backups, processing, and cold storage. The THS uses single-socket servers with fewer disks optimized for cost and density. Testing showed the THS outperformed baseline dual-socket servers on processing benchmarks while providing higher storage density and lower costs per node. The document concludes that for large clusters, specialized hardware can improve performance and efficiency compared to a single size-fits-all approach.
Effective SharePoint Architecture - SharePoint Saturday Stockholm 2016Alistair Pugin
The document discusses effective SharePoint architecture and provides recommendations for server roles, hardware specifications, and database configuration. It recommends a farm architecture with two web front end servers, two application servers, and a SQL Server 2012 cluster with two nodes. The web front end and application servers should each have 16GB RAM, 4 CPU cores, and two hard drives. SQL servers should have 32GB RAM, 8 CPU cores, and five hard drives configured for data, tempdb, and backup files. It also provides tips for securing SharePoint such as implementing firewalls, running security analyzers, and following hardening guidance.
The document discusses Ceph performance on all-flash storage systems. It describes optimizations made to Ceph's OSD architecture and write path that have led to significant performance improvements when deployed on SanDisk's InfiniFlash all-flash storage. These include reducing CPU utilization and improving throughput and latency. Example performance metrics are provided showing random read IOPS over 1.5M and latency under 5ms for most operations. The document also outlines the InfiniFlash hardware architecture and roadmap for further Ceph optimizations including new storage backends like BlueStore.
The document discusses Ceph storage performance on all-flash storage systems. It describes how SanDisk optimized Ceph for all-flash environments by tuning the OSD to handle the high performance of flash drives. The optimizations allowed over 200,000 IOPS per OSD using 12 CPU cores. Testing on SanDisk's InfiniFlash storage system showed it achieving over 1.5 million random read IOPS and 200,000 random write IOPS at 64KB block size. Latency was also very low, with 99% of operations under 5ms for reads. The document outlines reference configurations for the InfiniFlash system optimized for small, medium and large workloads.
Similar to SharePoint Performance at SPS Philly (20)
Uncover the Benefits of Office 365 & Windows AzureSentri
Polycom has acquired Sentri. The acquisition will allow Polycom to expand its offerings beyond video equipment to include solution capabilities. Sentri's technology and team will help Polycom grow its business.
The document announces that Sentri has been acquired by Polycom. It discusses Polycom's financial strength, strategic partnerships, customer momentum, and industry leadership in video conferencing. The acquisition of Sentri will help Polycom expand its offerings from equipment to solution capabilities.
Explore Microsoft Lync & Exchange 2013 WebinarSentri
The document announces that Sentri has been acquired by Polycom. Key details include that Sentri provides solutions for Microsoft technologies like Lync and Exchange, and that the acquisition will allow Polycom to expand its offerings beyond video equipment to full communications solutions.
Relationship Management for Property Investment Management webinar 2.5.13Sentri
The document discusses Relationship Management for Property & Investment Management and introduces Property Hub, a relationship management solution designed for professional property managers and building owners to manage their portfolio of properties and relationships. Property Hub provides a 360 degree view of property information, tools to manage investor portfolios and deals, critical date reminders, document management, marketing analytics, and is priced at $44 per user per month.
SharePoint 2013 Communities Overview for SPS PhillySentri
The document discusses SharePoint communities and provides an overview of key topics:
- It describes the different types of community sites, roles, and experiences for visitors, members, owners, and moderators.
- Functionality for communities such as discussions, profiles, activities are highlighted.
- Best practices for community setup and moderation are covered.
- The presentation emphasizes community integration with social networks and outlines scenarios for different collaboration solutions.
Move Storage to the Cloud with Windows Azure webinar 2.21.2013Sentri
This document discusses the business value of moving workloads and services to the cloud using Microsoft Office 365 and Azure. Key benefits highlighted include agility, scalability, reduced costs, high availability, disaster recovery, and the ability to focus on core projects rather than infrastructure maintenance. The document provides examples showing lower total cost of ownership for Office 365 compared to on-premises deployments over multi-year periods. Common cloud computing patterns and scenarios that can be implemented using Azure services are also reviewed.
The document discusses extranets and customer portals using cloud hybrid solutions. It describes how Sentri helps companies implement extranets and portals using Microsoft technologies like SharePoint. Sentri offers professional services, managed services, and cloud services to design, deploy, and manage extranets and customer portals. The document also discusses best practices for building extranets and portals with SharePoint 2013.
Lync Intro Exchange User Group New York 2_12_2013Sentri
This document discusses Microsoft Lync 2013 and its integration with Exchange. Lync 2013 provides a unified communications platform that allows users to communicate using instant messaging, presence, audio/video calls, and meetings from any device. It integrates with Exchange for archiving chat histories and enables single sign-on. The presentation emphasizes how Lync improves productivity and communication through an easy to use interface available across Windows, web, mobile clients and its federation capabilities.
The document discusses Microsoft Lync and how it enables presence, instant messaging, and communication across devices. It highlights how Lync improves productivity by enabling persistent chat and easy contact through instant messaging. It also describes features like seamless escalation between chat and calls, archiving of chat history to Exchange, and integration across the Lync client and server. The document promotes next steps like downloading the Lync preview, planning a Lync deployment, conducting a proof of concept, or managing Lync services through Sentri.
WEBINAR: Uncover the Benefits of Office 365 and Windows Azure Sentri
The document discusses cloud computing strategies and Microsoft's cloud offerings. Some key points:
- The cloud can provide agility, maximize infrastructure, and focus on projects rather than maintenance.
- Microsoft offers Office 365, Azure, and hybrid solutions that allow using public and private clouds together.
- Storage strategies are evolving from dispersed local storage to consolidated cloud-integrated storage.
- Solutions like StorSimple and Azure help simplify storage management and reduce costs by leveraging the cloud.
The document discusses Microsoft's new SharePoint 2013 product. It highlights many new features including improved social and mobile capabilities, enhanced document management and collaboration tools, and the ability to build modern websites and apps. It also discusses options for deploying SharePoint 2013 on-premises or in the cloud through Office 365 and managing the product through tools like System Center 2012.
Webinar: Understanding the System Center suite & Windows Server 2012 Sentri
This document discusses datacenter management and provides options for managing mobile users, virtualization, security, and costs. It compares traditional and virtualized datacenters and approaches for on-premises and off-premises dynamic data centers. The System Center suite is presented as providing a productive, predictable, and flexible infrastructure that can manage applications across private and public clouds using common tools.
Sentri Webinar: Microsoft Lync Exchange 2013Sentri
The document provides an overview of Microsoft Lync and Exchange 2013. It highlights key features such as unified communications, collaboration tools, mobility access, and security and compliance capabilities. The document also discusses integration with Office apps and upcoming features for Lync and Exchange. It concludes by presenting several next steps for learning more about, planning for, and deploying Lync and Exchange 2013.
The document discusses Microsoft's Windows Azure cloud platform. It describes Azure as an open and flexible platform that allows users to focus on their projects instead of infrastructure. Azure provides agility through services like virtual machines, databases, storage and networking that can be accessed on demand over the internet.
The document explores Microsoft Exchange 2013 and its features. It summarizes that Exchange 2013 allows users to manage communications across multiple devices, work collaboratively in teams, and helps organizations meet compliance requirements. It also highlights new features including an improved interface, enhanced security and compliance tools, and the ability to work across Outlook and Outlook Web App.
The document discusses Microsoft's new SharePoint 2013 product. It highlights many new features including improved social and mobile capabilities, enhanced document management and collaboration tools, and the ability to build modern websites and apps. It also discusses options for deploying SharePoint 2013 on-premises or in the cloud through Office 365 and managing the product through tools like System Center 2012.
Empowering Collaboration through Self-Service BISentri
The document discusses empowering collaboration through self-service business intelligence (BI). It introduces Jason Himmelstein and Kristen Maraghy and provides information about their roles and backgrounds. Several quotes are included that discuss how big data is becoming increasingly important for competition and growth. The document then outlines Microsoft's BI and analytics portfolio including Excel, PowerPivot, Power View and SharePoint and how they work together.
Empowering Collaboration through Self-Service BISentri
The document discusses empowering collaboration through self-service business intelligence (BI). It introduces Jason Himmelstein and Kristen Maraghy and provides information about their roles and backgrounds. Several quotes are included that discuss how big data is becoming increasingly important for competition and growth. The document then outlines Microsoft's BI and analytics portfolio including Excel, PowerPivot, Power View and SharePoint and how they work together.
Exchange 2013 allows users to access email from any device. It provides security, compliance, and insights through monitoring and analytics. The document outlines Exchange 2013 features like unified communications, eDiscovery, data loss prevention, and retention policies that help users work productively from anywhere while keeping the organization safe and compliant.
The document provides information about Microsoft Lync 2013 Preview. It highlights key features of Lync 2013 such as a unified communications platform, HD video conferencing, mobile client experiences, integration with Office apps, federation with Skype, and deployment flexibility. It also discusses benefits of Lync like improved productivity and mobility, connected experiences, and simplified management. Finally, it lists next steps one can take including downloading the Lync 2013 preview, planning an upgrade, and working with Sentri on proof of concepts or managed services.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
AI 101: An Introduction to the Basics and Impact of Artificial IntelligenceIndexBug
Imagine a world where machines not only perform tasks but also learn, adapt, and make decisions. This is the promise of Artificial Intelligence (AI), a technology that's not just enhancing our lives but revolutionizing entire industries.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Have you ever been confused by the myriad of choices offered by AWS for hosting a website or an API?
Lambda, Elastic Beanstalk, Lightsail, Amplify, S3 (and more!) can each host websites + APIs. But which one should we choose?
Which one is cheapest? Which one is fastest? Which one will scale to meet our needs?
Join me in this session as we dive into each AWS hosting service to determine which one is best for your scenario and explain why!
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Programming Foundation Models with DSPy - Meetup SlidesZilliz
Prompting language models is hard, while programming language models is easy. In this talk, I will discuss the state-of-the-art framework DSPy for programming foundation models with its powerful optimizers and runtime constraint system.
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
2. Israel - Development &
Business since 1999 Engineering
Partner of the Year
Microsoft 2010/2011
Northeast
Partner of the Year
Microsoft 2011 Northeast
VOICE
7. Developer or Evaluation environments Production in Single Server or farm environments
CPU: 4 cores, 64-bit required CPU: 4 cores, 64-bit required
RAM: 4GB RAM: 8GB
Hard Drive space: 80GB Hard Drive space: 80GB
Small Farm Medium Farm Large Farm
CPU: 4 cores, 64-bit required CPU: 8 cores, 64-bit required Up to 2TB Content DBS
RAM: 8GB RAM: 16GB RAM: 32 GB
Hard Drive space: 80GB Hard Drive space: 80GB From 2TB to 5TB Content DBS
RAM: 64 GB
8.
9.
10.
11.
12.
13.
14.
15. DB Files RAID Level Optimization
1 TempDB data 10 Write
2 TempDB logs 10 Write
3 ContentDB data 10 ReadWrite
4 ContentDB logs 10 Write
5 Crawl DB logs 10 Write
6 Crawl DB data 10 ReadWrite
7 Property DB logs 10 Write
8 Property DB data 10 Write
9 Services DB logs 10 Write
10 Services DB data 5/10 ReadWrite
11 Archive Content DB 5 Read
12 Publishing Site Content DB 5 Read