Datalagring för AI
Vad bör man att tänka på, hur bygger man och vilken skillnad kan IBM's infrastruktur göra.
Talare: Christofer Jensen, Storage Technical Specialist, IBM
Presentationen hölls på Watson Kista Summit 2018
In this video from the Stanford HPC Conference, Liran Zvibel from Weka.IO presents: Making Machine Learning Compute Bound Again.
"GPUs are getting faster on a yearly cycle. Networking was able to catch up and support linear scaling of models that fit in memory. Traditional storage has not caught up to the condensed performance needed by GPU-filled servers. The amount of concurrent clients and the sheer amount of data required to effectively scale modern deep learning models keeps growing.
We are going to present WekaIO, the lowest latency, highest throughput file system solution that scales to 100s of PB in a single namespace supporting the most challenging deep learning projects that run today. We will present real life benchmarks comparing WekaIO performance to a local SSD file system, showing that we are the only coherent shared storage that is even faster than the current caching solutons, while allowing customers to linearly scale performance by adding more GPU servers. Also, we will view the complete ML project lifecycle, from collecting data, cleaning, tagging, exploring, training, validating, and finally archiving, and how customers can use cloud bursting to leverage public cloud infrastructure for improved economics."
Learn more: https://weka.io
and
http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
HORTONWORKS DATA PLATFORM AND IBM SYSTEMS – A COMPLETE SOLUTION FOR COGNITIVE BUSINESS
SynerScope has been helping European organizations across industries unlock competitive business value from data for almost a decade. Now, by leveraging state-of-the-art access control and audit mechanisms from Hortonworks combined with the latest generation high-performance computing and storage solutions from IBM, SynerScope can connect and correlate enterprise data at a scale not previously possible. SynerScope will demonstrate end-to-end analytics workflows including deep-learning based automation using new integrated solutions from Hortonworks and IBM.
In this video from the Stanford HPC Conference, Liran Zvibel from Weka.IO presents: Making Machine Learning Compute Bound Again.
"GPUs are getting faster on a yearly cycle. Networking was able to catch up and support linear scaling of models that fit in memory. Traditional storage has not caught up to the condensed performance needed by GPU-filled servers. The amount of concurrent clients and the sheer amount of data required to effectively scale modern deep learning models keeps growing.
We are going to present WekaIO, the lowest latency, highest throughput file system solution that scales to 100s of PB in a single namespace supporting the most challenging deep learning projects that run today. We will present real life benchmarks comparing WekaIO performance to a local SSD file system, showing that we are the only coherent shared storage that is even faster than the current caching solutons, while allowing customers to linearly scale performance by adding more GPU servers. Also, we will view the complete ML project lifecycle, from collecting data, cleaning, tagging, exploring, training, validating, and finally archiving, and how customers can use cloud bursting to leverage public cloud infrastructure for improved economics."
Learn more: https://weka.io
and
http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
HORTONWORKS DATA PLATFORM AND IBM SYSTEMS – A COMPLETE SOLUTION FOR COGNITIVE BUSINESS
SynerScope has been helping European organizations across industries unlock competitive business value from data for almost a decade. Now, by leveraging state-of-the-art access control and audit mechanisms from Hortonworks combined with the latest generation high-performance computing and storage solutions from IBM, SynerScope can connect and correlate enterprise data at a scale not previously possible. SynerScope will demonstrate end-to-end analytics workflows including deep-learning based automation using new integrated solutions from Hortonworks and IBM.
Backup and Archive Doesn't Have to be Complicated and Expensivespectralogic
Spectra and CommVault® have joined forces to significantly lower the cost of managing and storing data virtually forever. Organizations worldwide face continuing challenges in managing and protecting their data, particularly with the prevailing desire to retain data for longer, often indefinite periods of time. Combining CommVault Simpana software, which allows organizations to protect, manage and access information regardless of where the data resides, with Spectra's simply affordable nTier Verde disk and T-Series tape library platforms, provides customers with very affordable storage and effective, easy to manage data protection, backup and archive.
In this deck, Gilad Shainer from the HPC Advisory Council kicks off the 2018 Stanford HPC Conference.
Watch the video: https://youtu.be/2ya9Ougcz9c
Learn more:
http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Lacking the technology to directly leverage Hadoop, some companies are foregoing its full benefits opting to treat Hadoop as just another data source for their legacy BI tools. But storage is only one benefit of Hadoop and ignores its linear scalability and data flexibility across all data types. Using Hadoop natively for both storage and computation in an analytic capacity has already led to dramatic increases in business benefits. Hadoop analytics has already identified over $2B in potential fraud at one of the world’s largest credit card companies. Sears has already reduced reporting times over traditional BI from 12 weeks to 3 days. A major internet security company increased customer conversion by 60% and revenue by $20 million. Meaningful returns are spread across Fortune 100 enterprises and fast growing startups with the common thread being self-service big data analytics leveraging Hadoop’s native capabilities. In this talk, we’ll highlight the core value proposition of building analytics natively on Hadoop, share real-world use cases that resulted in dramatic ROI, and reveal the next major step in visual big data analytics.
Learn about how to reduce public cloud storage costs on the AWS and Azure marketplaces with SoftNAS Senior Director of Product Marketing, John Bedrick.
Introduction to Big Data Hadoop Training Online by www.itjobzone.bizITJobZone.biz
Want to learn Hadoop online? This PPT give you Introduction to Big Data Hadoop Training Online by expert trainers at ITJobZone.biz - Start your Hadoop Online training with this Presentation.
Threat Detection and Response at Scale with Dominique BrezinskiDatabricks
Security monitoring and threat response has diverse processing demands on large volumes of log and telemetry data. Processing requirements span from low-latency stream processing to interactive queries over months of data. To make things more challenging, we must keep the data accessible for a retention window measured in years. Having tackled this problem before in a massive-scale environment using Apache Spark, when it came time to do it again, there were a few things I knew worked and a few wrongs I wanted to right.
We approached Databricks with a set of challenges to collaborate on: provide a stable and optimized platform for Unified Analytics that allows our team to focus on value delivery using streaming, SQL, graph, and ML; leverage decoupled storage and compute while delivering high performance over a broad set of workloads; use S3 notifications instead of list operations; remove Hive Metastore from the write path; and approach indexed response times for our more common search cases, without hard-to-scale index maintenance, over our entire retention window. This is about the fruit of that collaboration.
Decoupling Compute and Storage for Data WorkloadsAlluxio, Inc.
This was presented by Carlos Quieroz, Head of Data Platform at Development Bank of Singapore, at the Data Transformation in Financial Services meetup in Singapore jointly hosted by Accenture, Talend, BigDataSG Hadoop, and Alluxio.
Introducing Big Data concepts & Hadoop to those who wish to begin their journey in the future of Information Technology. It is certain that data is going to play a major role in days to come, from our daily lives to the biggest of the venture we might undertake.And hence, knowing about Big data and technologies to work with the same is going to be essential for IT professionals.
We will start with some simple presentations and then will go on building upon it. For more intense, focused introduction & training on Big Data and related technologies, visit our website or write to us.
Alluxio 2.0 Deep Dive – Simplifying data access for cloud workloadsAlluxio, Inc.
Alluxio Tech Talk
Aug 7, 2019
Speaker:
Dipti Borkar, Alluxio
Alluxio 2.0 is the most ambitious platform upgrade since the inception of Alluxio with greatly expanded capabilities to empower users to run analytics and AI workloads on private, public or hybrid cloud infrastructures leveraging valuable data wherever it might be stored.
This release, now available for download, includes many advancements that will allow users to push the limits of their data-workloads in the cloud.
In this tech talk, we will introduce the key new features and enhancements such as:
- Support for hyper-scale data workloads with tiered metadata storage, distributed cluster services, and adaptive replication for increased data locality
- Machine learning and deep learning workloads on any storage with the improved POSIX API
- Better storage abstraction with support for HDFS clusters across different versions & active sync with Hadoop
Running cost effective big data workloads with Azure Synapse and Azure Data L...Michael Rys
The presentation discusses how to migrate expensive open source big data workloads to Azure and leverage latest compute and storage innovations within Azure Synapse with Azure Data Lake Storage to develop a powerful and cost effective analytics solutions. It shows how you can bring your .NET expertise with .NET for Apache Spark to bear and how the shared meta data experience in Synapse makes it easy to create a table in Spark and query it from T-SQL.
Meetup at AI NextCon 2019: In-Stream data process, Data Orchestration & MoreAlluxio, Inc.
Alluxio - Data Orchestration for Analytics and AI in the Cloud
Oct 8, 2019
Speakers:
Haoyuan Li & Bin Fan, Alluxio
Visit https://www.alluxio.io/events/ for more Alluxio events.
Using Spark and Riak for IoT Apps—Patterns and Anti-Patterns: Spark Summit Ea...Spark Summit
Everybody agrees that IoT is changing the world… and creates new challenges for software developers, architects and DevOps. How can we build efficient and highly scalable distributed applications using open-source technologies? What are characteristics of data generated by IoT devices and how it differs from traditional enterprise or Big Data problems? Which architectural patterns are beneficial for IoT use cases and why some trusted methods eventually turn out to be “anti-patterns”? This talk will show how to combine best-of-breed open-source technologies, like Apache Spark, Riak and Mesos to build scalable IoT pipelines to ingest, store and analyze huge amounts of data, while keeping operational complexity and costs under control. We will discuss cons and pros of using relational, NoSQL and object storage products for storing and archiving IoT data. Then we cover best practices how to use Spark with Riak NoSQL database. Will describe how Apache Spark advanced modules (Spark SQL, Spark Streaming and MLlib) can solve the problems common to IoT apps, while using Riak for fast and scalable persistence. At the end, will explain why Structured Spark Streaming is a godsend for IoT data and make a case for Time Series databases deserving a separate category in NoSQL classification.
MaaS (Model as a Service): Modern Streaming Data Science with Apache Metron (...DataWorks Summit
Apache Metron (Incubating) is a streaming cybersecurity application
built on Apache Storm and Hadoop. One of its core missions is to enable
advanced analytics through machine learning and data science to the
users. Because of the relative immaturity of data science platform
infrastructure integrated into Hadoop that is oriented to streaming
analytics applications, we have been forced to create the requisite
platform components out of necessity, utilizing many of the pieces of
the Hadoop ecosystem.
In this talk, we will speak about the Metron analytics architecture and
how it utilizes a custom data science model deployment and autodiscovery
service that is tightly integrated with Hadoop via Yarn and Zookeeper.
We will discuss how we interact with the models deployed there via a
custom domain specific language that can query models as data streams
past. We will generally discuss the full-stack data science tooling that
has been created to enable data science at scale on an advanced analytics
streaming application.
Evolution from Apache Hadoop to the Enterprise Data Hub by Cloudera - ArabNet...ArabNet ME
A new foundation for the Modern Information Architecture.
Speaker: Amr Awadallah, CTO & Cofounder, Cloudera
Our legacy information architecture is not able to cope with the realities of today's business. This is because it is not able to scale to meet our SLAs due to separation of storage and compute, economically store the volumes and types of data we currently confront, provide the agility necessary for innovation, and most importantly, provide a full 360 degree view of our customers, products, and business. In this talk Dr. Amr Awadallah will present the Enterprise Data Hub (EDH) as the new foundation for the modern information architecture. Built with Apache Hadoop at the core, the EDH is an extremely scalable, flexible, and fault-tolerant, data processing system designed to put data at the center of your business.
Data analytics, Spark, Hadoop and AI have become fundamental tools to drive digital transformation. A critical challenge is moving from isolated experiments to an organizational or enterprise production infrastructure. In this talk, we break apart the modern data analytics workflow to focus on the data challenges across different phases of the analytics and AI life cycle. By presenting a unified approach to data storage for AI and Analytics, organizations can reduce costs, modernize their data strategy and build a sustainable enterprise data lake. By anticipating how Hadoop, Spark, Tensorflow, Caffe and traditional analytics like SAS, HPC can share data, IT departments and data science practitioners can not only co-exist, but speed time to insight. We will present the tangible benefits of a Reference Architecture using real-world installations that span proprietary and open-source frameworks. Using intelligent software-defined shared storage, users are able to eliminate silos, reduce multiple data copies, and improve time to insight.PALLAVI GALGALI, Offering Manager,IBM and DOUGLAS O'FLAHERTY, Portfolio Product Manager, IBM
Backup and Archive Doesn't Have to be Complicated and Expensivespectralogic
Spectra and CommVault® have joined forces to significantly lower the cost of managing and storing data virtually forever. Organizations worldwide face continuing challenges in managing and protecting their data, particularly with the prevailing desire to retain data for longer, often indefinite periods of time. Combining CommVault Simpana software, which allows organizations to protect, manage and access information regardless of where the data resides, with Spectra's simply affordable nTier Verde disk and T-Series tape library platforms, provides customers with very affordable storage and effective, easy to manage data protection, backup and archive.
In this deck, Gilad Shainer from the HPC Advisory Council kicks off the 2018 Stanford HPC Conference.
Watch the video: https://youtu.be/2ya9Ougcz9c
Learn more:
http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Lacking the technology to directly leverage Hadoop, some companies are foregoing its full benefits opting to treat Hadoop as just another data source for their legacy BI tools. But storage is only one benefit of Hadoop and ignores its linear scalability and data flexibility across all data types. Using Hadoop natively for both storage and computation in an analytic capacity has already led to dramatic increases in business benefits. Hadoop analytics has already identified over $2B in potential fraud at one of the world’s largest credit card companies. Sears has already reduced reporting times over traditional BI from 12 weeks to 3 days. A major internet security company increased customer conversion by 60% and revenue by $20 million. Meaningful returns are spread across Fortune 100 enterprises and fast growing startups with the common thread being self-service big data analytics leveraging Hadoop’s native capabilities. In this talk, we’ll highlight the core value proposition of building analytics natively on Hadoop, share real-world use cases that resulted in dramatic ROI, and reveal the next major step in visual big data analytics.
Learn about how to reduce public cloud storage costs on the AWS and Azure marketplaces with SoftNAS Senior Director of Product Marketing, John Bedrick.
Introduction to Big Data Hadoop Training Online by www.itjobzone.bizITJobZone.biz
Want to learn Hadoop online? This PPT give you Introduction to Big Data Hadoop Training Online by expert trainers at ITJobZone.biz - Start your Hadoop Online training with this Presentation.
Threat Detection and Response at Scale with Dominique BrezinskiDatabricks
Security monitoring and threat response has diverse processing demands on large volumes of log and telemetry data. Processing requirements span from low-latency stream processing to interactive queries over months of data. To make things more challenging, we must keep the data accessible for a retention window measured in years. Having tackled this problem before in a massive-scale environment using Apache Spark, when it came time to do it again, there were a few things I knew worked and a few wrongs I wanted to right.
We approached Databricks with a set of challenges to collaborate on: provide a stable and optimized platform for Unified Analytics that allows our team to focus on value delivery using streaming, SQL, graph, and ML; leverage decoupled storage and compute while delivering high performance over a broad set of workloads; use S3 notifications instead of list operations; remove Hive Metastore from the write path; and approach indexed response times for our more common search cases, without hard-to-scale index maintenance, over our entire retention window. This is about the fruit of that collaboration.
Decoupling Compute and Storage for Data WorkloadsAlluxio, Inc.
This was presented by Carlos Quieroz, Head of Data Platform at Development Bank of Singapore, at the Data Transformation in Financial Services meetup in Singapore jointly hosted by Accenture, Talend, BigDataSG Hadoop, and Alluxio.
Introducing Big Data concepts & Hadoop to those who wish to begin their journey in the future of Information Technology. It is certain that data is going to play a major role in days to come, from our daily lives to the biggest of the venture we might undertake.And hence, knowing about Big data and technologies to work with the same is going to be essential for IT professionals.
We will start with some simple presentations and then will go on building upon it. For more intense, focused introduction & training on Big Data and related technologies, visit our website or write to us.
Alluxio 2.0 Deep Dive – Simplifying data access for cloud workloadsAlluxio, Inc.
Alluxio Tech Talk
Aug 7, 2019
Speaker:
Dipti Borkar, Alluxio
Alluxio 2.0 is the most ambitious platform upgrade since the inception of Alluxio with greatly expanded capabilities to empower users to run analytics and AI workloads on private, public or hybrid cloud infrastructures leveraging valuable data wherever it might be stored.
This release, now available for download, includes many advancements that will allow users to push the limits of their data-workloads in the cloud.
In this tech talk, we will introduce the key new features and enhancements such as:
- Support for hyper-scale data workloads with tiered metadata storage, distributed cluster services, and adaptive replication for increased data locality
- Machine learning and deep learning workloads on any storage with the improved POSIX API
- Better storage abstraction with support for HDFS clusters across different versions & active sync with Hadoop
Running cost effective big data workloads with Azure Synapse and Azure Data L...Michael Rys
The presentation discusses how to migrate expensive open source big data workloads to Azure and leverage latest compute and storage innovations within Azure Synapse with Azure Data Lake Storage to develop a powerful and cost effective analytics solutions. It shows how you can bring your .NET expertise with .NET for Apache Spark to bear and how the shared meta data experience in Synapse makes it easy to create a table in Spark and query it from T-SQL.
Meetup at AI NextCon 2019: In-Stream data process, Data Orchestration & MoreAlluxio, Inc.
Alluxio - Data Orchestration for Analytics and AI in the Cloud
Oct 8, 2019
Speakers:
Haoyuan Li & Bin Fan, Alluxio
Visit https://www.alluxio.io/events/ for more Alluxio events.
Using Spark and Riak for IoT Apps—Patterns and Anti-Patterns: Spark Summit Ea...Spark Summit
Everybody agrees that IoT is changing the world… and creates new challenges for software developers, architects and DevOps. How can we build efficient and highly scalable distributed applications using open-source technologies? What are characteristics of data generated by IoT devices and how it differs from traditional enterprise or Big Data problems? Which architectural patterns are beneficial for IoT use cases and why some trusted methods eventually turn out to be “anti-patterns”? This talk will show how to combine best-of-breed open-source technologies, like Apache Spark, Riak and Mesos to build scalable IoT pipelines to ingest, store and analyze huge amounts of data, while keeping operational complexity and costs under control. We will discuss cons and pros of using relational, NoSQL and object storage products for storing and archiving IoT data. Then we cover best practices how to use Spark with Riak NoSQL database. Will describe how Apache Spark advanced modules (Spark SQL, Spark Streaming and MLlib) can solve the problems common to IoT apps, while using Riak for fast and scalable persistence. At the end, will explain why Structured Spark Streaming is a godsend for IoT data and make a case for Time Series databases deserving a separate category in NoSQL classification.
MaaS (Model as a Service): Modern Streaming Data Science with Apache Metron (...DataWorks Summit
Apache Metron (Incubating) is a streaming cybersecurity application
built on Apache Storm and Hadoop. One of its core missions is to enable
advanced analytics through machine learning and data science to the
users. Because of the relative immaturity of data science platform
infrastructure integrated into Hadoop that is oriented to streaming
analytics applications, we have been forced to create the requisite
platform components out of necessity, utilizing many of the pieces of
the Hadoop ecosystem.
In this talk, we will speak about the Metron analytics architecture and
how it utilizes a custom data science model deployment and autodiscovery
service that is tightly integrated with Hadoop via Yarn and Zookeeper.
We will discuss how we interact with the models deployed there via a
custom domain specific language that can query models as data streams
past. We will generally discuss the full-stack data science tooling that
has been created to enable data science at scale on an advanced analytics
streaming application.
Evolution from Apache Hadoop to the Enterprise Data Hub by Cloudera - ArabNet...ArabNet ME
A new foundation for the Modern Information Architecture.
Speaker: Amr Awadallah, CTO & Cofounder, Cloudera
Our legacy information architecture is not able to cope with the realities of today's business. This is because it is not able to scale to meet our SLAs due to separation of storage and compute, economically store the volumes and types of data we currently confront, provide the agility necessary for innovation, and most importantly, provide a full 360 degree view of our customers, products, and business. In this talk Dr. Amr Awadallah will present the Enterprise Data Hub (EDH) as the new foundation for the modern information architecture. Built with Apache Hadoop at the core, the EDH is an extremely scalable, flexible, and fault-tolerant, data processing system designed to put data at the center of your business.
Data analytics, Spark, Hadoop and AI have become fundamental tools to drive digital transformation. A critical challenge is moving from isolated experiments to an organizational or enterprise production infrastructure. In this talk, we break apart the modern data analytics workflow to focus on the data challenges across different phases of the analytics and AI life cycle. By presenting a unified approach to data storage for AI and Analytics, organizations can reduce costs, modernize their data strategy and build a sustainable enterprise data lake. By anticipating how Hadoop, Spark, Tensorflow, Caffe and traditional analytics like SAS, HPC can share data, IT departments and data science practitioners can not only co-exist, but speed time to insight. We will present the tangible benefits of a Reference Architecture using real-world installations that span proprietary and open-source frameworks. Using intelligent software-defined shared storage, users are able to eliminate silos, reduce multiple data copies, and improve time to insight.PALLAVI GALGALI, Offering Manager,IBM and DOUGLAS O'FLAHERTY, Portfolio Product Manager, IBM
Modernizing upstream workflows with aws storage - john malloryAmazon Web Services
Modernizing Upstream Workflows with AWS Storage
Accelerating seismic data retrieval, getting better data protection and reliability, and providing a common AWS data platform for compute and graphic intensive processing, simulation and visualization workloads.
Modernizing and transforming exploration and production workflows with AWS Storage services
Accelerating seismic data retrieval, getting better data protection and reliability, and providing a common AWS data platform for compute and graphic intensive processing, simulation and visualization workloads.
Capturing and processing streaming sensor data from remote oil rigs with Snowball Edge
Providing a Data Lake foundation for a next generation Digital Oilfield IoT analytics platform with Amazon S3
Speaker: John Mallory - AWS Storage Business Development Manager
One of the toughest decisions when selecting a new storage system is deciding between an All-Flash Array and a Hybrid Array. Do you go with the predictable high performance of the all flash architecture or the attractive price per GB of the mixed flash and hard disk architecture? What if you can have both? In this webinar learn how to get predictable performance from your Hybrid Arrays so it will perform like an All-Flash Array.
HPE Hadoop Solutions - From use cases to proposalDataWorks Summit
Hadoop is now doing a lot more than just storage and Map/Reduce and always improving and innovating. It brings near real time, interactive and cost efficient features to do Big Data.
Join us to hear about solutions based on Hadoop, how they responds to specific customer needs, with what component(s) from the Hadoop ecosystem, based on what HPE Reference Architecture(s) for the platform.
Hadoop solutions like, ETL offloading, Predictive Analytics, Ad hoc query, Complex Event processing, Stream processing, Search, Machine learning, Deep learning, …
Based on software components like, Spark, Hive, HBase, Kafka, Storm, Flume, Impala and Elastic Search.
Speaker
John Osborn, SA, Hewlett Packard Enterprise
Is Software Defined Storage (SDS) getting hijacked? It seems every vendor, old and new, is claiming that their storage is “software defined”. The original intent was to create software-only solutions that could be deployed on the customer’s choice of servers. But that original intent has evolved, and now hardware vendors are providing what they claim to be software defined storage solutions too. In addition, SDS is being combined with an embedded compute function to create hyper-converged solutions as well.
In this webinar we will discuss the differences in these approaches and you will learn what the four key deliverables of a SDS solution should be so you can decide which makes the most sense for your organization.
If you're like most of the world, you're on an aggressive race to implement machine learning applications and on a path to get to deep learning. If you can give better service at a lower cost, you will be the winners in 2030. But infrastructure is a key challenge to getting there. What does the technology infrastructure look like over the next decade as you move from Petabytes to Exabytes? How are you budgeting for more colossal data growth over the next decade? How do your data scientists share data today and will it scale for 5-10 years? Do you have the appropriate security, governance, back-up and archiving processes in place? This session will address these issues and discuss strategies for customers as they ramp up their AI journey with a long term view.
ADV Slides: The Evolution of the Data Platform and What It Means to Enterpris...DATAVERSITY
Thirty years is a long time for a technology foundation to be as active as relational databases. Are their replacements here?
In this webinar, we look at this foundational technology for modern Data Management and show how it evolved to meet the workloads of today, as well as when other platforms make sense for enterprise data.
Semiconductor design companies, electronic design automation (EDA) vendors, and foundries remain competitive by innovating and reducing time to market. AWS is deeply invested in semiconductor use cases, including EDA, emulation, and smart manufacturing, including data lake and IoT/AI. We care about this because Amazon depends on faster semiconductor innovation from our suppliers and in our own silicon teams. We have a wide breadth of services that will directly benefit the entire industry. In this session, learn how to achieve the maximum possible performance and throughput from design and engineering workloads running on AWS. We demonstrate specific optimization techniques and share architectures to accelerate batch and interactive workloads on AWS. We also demonstrate how to extend and migrate on-premises, high performance compute workloads with AWS, and use a combination of On-Demand Instances, Reserved Instances, and Spot Instances to minimize costs. Learn how semiconductor customers address security as they move to the cloud as they discuss the AWS capabilities and controls available to secure sensitive design IP and offer strategies for data classification, management, and transfer to third parties.
Antoine Genereux takes us on a detailed overview of the Database solutions available on the AWS Cloud, addressing the needs and requirements of customers at all levels. He also discusses Business Intelligence and Analytics solutions.
Medical imaging for active data and archive
Digital simulations in pharmaceutical, automotive, aerospace
Rich content records in insurance, construction, realty
Video capture for security, process management, education
Content distribution in Media & Entertainment
Rich text E-mail, Web 2.0 and Social Networking
Analytics in Financial services
Grow smarter project kista watson summit 2018_tommy auoja-1IBM Sverige
Avicii på Tele2 arena, Drake på Globen och AIK - Luleå på Hovet bäddar för en trång lördagseftermiddag i Globenområdet... (SVT Nyheter, 1 mars 2014) ...och problemen kvarstår än idag
Talare: Tommy Auoja, Kundansvarig för Offentlig Sektor, Kontaktperson i EU projektet GrowSmarter, IBM
Presentation från Watson Kista Summit 2018
Bemanningsplanering axfood och houston finalIBM Sverige
Automatiserad budgetering – låt matematiken göra grovgörat för att säkerställa en optimerad bemanning
Talare: Niklas Westerholm, Axfood & Robert Moberg, Chief Analyst, Houston Analytics
Presentation från Watson Kista Summit 2018
File share and sync (bara) är så 2017!
Att dela filer bekvämt och säkert var bara början. Box har gått vidare till att integrera delade filer i applikationer och processflöden, och revolutionera både internt och externt arbete. Hur kan det revolutionera för dig?
Talare: Jan Hygstedt, Director Nordic, Box
Presentation från Watson Kista Summit 2018
Watson kista summit 2018 en bättre arbetsdag för de många människornaIBM Sverige
Först tvingades vi anpassa oss efter datorerna. Sedan använde vi dem för att samarbeta med varandra. Nu är det dags för datorerna att förstå oss. Vad innebär det för vår arbetsvardag?
Talare och moderator: Peter Bjellerup, Executive Consultant - Social Business, Collaboration & Knowledge Sharing, IBM
Presentation från Watson Kista Summit 2018
Iwcs and cisco watson kista summit 2018 v2IBM Sverige
Samarbeta både över tid och i realtid
Cisco Spark och IBM Connections – tillsammans! Kombinera ledaren för konversationer i realtid – text, video, individuellt och i team med branschledaren sedan sju år för internt samarbete, transparens och nätverk.
Talare: Bo Holtemann, Solution Specialist, IBM Collaboration Solutions
Presentation från Watson Kista Summit 2018
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Welocme to ViralQR, your best QR code generator.ViralQR
Welcome to ViralQR, your best QR code generator available on the market!
At ViralQR, we design static and dynamic QR codes. Our mission is to make business operations easier and customer engagement more powerful through the use of QR technology. Be it a small-scale business or a huge enterprise, our easy-to-use platform provides multiple choices that can be tailored according to your company's branding and marketing strategies.
Our Vision
We are here to make the process of creating QR codes easy and smooth, thus enhancing customer interaction and making business more fluid. We very strongly believe in the ability of QR codes to change the world for businesses in their interaction with customers and are set on making that technology accessible and usable far and wide.
Our Achievements
Ever since its inception, we have successfully served many clients by offering QR codes in their marketing, service delivery, and collection of feedback across various industries. Our platform has been recognized for its ease of use and amazing features, which helped a business to make QR codes.
Our Services
At ViralQR, here is a comprehensive suite of services that caters to your very needs:
Static QR Codes: Create free static QR codes. These QR codes are able to store significant information such as URLs, vCards, plain text, emails and SMS, Wi-Fi credentials, and Bitcoin addresses.
Dynamic QR codes: These also have all the advanced features but are subscription-based. They can directly link to PDF files, images, micro-landing pages, social accounts, review forms, business pages, and applications. In addition, they can be branded with CTAs, frames, patterns, colors, and logos to enhance your branding.
Pricing and Packages
Additionally, there is a 14-day free offer to ViralQR, which is an exceptional opportunity for new users to take a feel of this platform. One can easily subscribe from there and experience the full dynamic of using QR codes. The subscription plans are not only meant for business; they are priced very flexibly so that literally every business could afford to benefit from our service.
Why choose us?
ViralQR will provide services for marketing, advertising, catering, retail, and the like. The QR codes can be posted on fliers, packaging, merchandise, and banners, as well as to substitute for cash and cards in a restaurant or coffee shop. With QR codes integrated into your business, improve customer engagement and streamline operations.
Comprehensive Analytics
Subscribers of ViralQR receive detailed analytics and tracking tools in light of having a view of the core values of QR code performance. Our analytics dashboard shows aggregate views and unique views, as well as detailed information about each impression, including time, device, browser, and estimated location by city and country.
So, thank you for choosing ViralQR; we have an offer of nothing but the best in terms of QR code services to meet business diversity!
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
3. What is AI?
• Machine learning
• Deep learning
• Artificial intelligence ”Set framework”
• Four legs
• Narrow eyes
• Sharp teeth
• Tail
• Etc…
”Tell the system”
”Take action”
4. Three ways how IT uses data … today
Procedural (if…then)
Statistical (big data)
Artificial Intelligence
”One truth” ”Qualified guess” ”Learning Systems”
5. … and in 10 years
Procedural
(if…then)
Statistical
(big data)
AI
6. Current examples
shopping, profiling,
fraud detection …
autonomous driving,
image classification,
chatbots, gaming…
Structured processing
plausible, credible data
Accumulation of data
not 100% precise is ok
(e.g. Recommendations)
Training data
true and false examples
+ independent test data
business as usual
classic / legacy IT
7. Why this will happen
Procedural Statistical AI
Amount of
data used
Manual modeling
Accumulation
of examples
Automatic modeling
Legacy systems
Structured models
Data generation
”Just store the data”
New gen programmers
Automatic consumption
”Set the system free”
8. Procedural:
Archive for
auditing
Statistical:
Store all data for
parallel processing
Machine Learning:
Train sample data, then
offer for data trade
How is data stored?
if…then…else
GB/s
1
2
Structured Unstructured Unstructured + structured
9. What is important for
Image: Business over Broadway
GB/s
• Collected data is analyzed in parallel
• Number of analyzes / second is important
• Data must be close to the CPU
• Transaction latency is irrelevant
• Data consistency is irrelevant
10. What is important for
• Sample data is trained and then archived
• Short training = many training cycles, high quality
• The better the data, the better result
• High throughput at 1 point in the life cycle [1]
• As low as possible maintenance cost after [2]
1
2
11. Storage requirements summary
Primary:
• High throughput for analysis and training
• Scalable due to high data growth
• Low cost long term storage
Secondary:
• Automated archiving
• Data rescillency
• Availabilty
How does IBM solve this???
13. Automotive Industry generates large amounts of data
Sensors
Video
CAN
FlexRay
Radar
LiDAR
Etc etc
Data must be synchronously captured, stored, modified and executed
14. Dev / test is challenging
Test Drives
50TB / day / car
R&D Lab: tagging
R&D Labs: developing
& testing
> 5PB / car model (project)
> 200h / 1h driving
16. Major IT Challenges
4. How to analyze the data – esp.
sensor and video data analytics
2. How to distribute data globally
within an enterprise
1. How to implement & operate an
efficient storage, workflow and
management system
„The Foundation“
3. How to preserve digital data
for decades
6. How to embed analytics/data
management into R&D
Environment
5. How to do efficient IT workload
and resource scheduling?
17. Summary – Solution Elements ADAS/AD
AREMA AgentsAREMA EngineAREMA Interfaces
<
SOAP REST OSLC
Elektrobit ADTF and other
ADTF and testing tools
AREMA Clients
Spectrum Scale client OS
IBM Video Analytics
IBM Reserach
HiL Station(s)
IBM Spectrum
Protec
Job Management, Media Portal
Automatic Video
Tagging/Labelling
ArchiveStorage & DistributionTest Execution
Test- & Lab Management
+ linkages to Development
Manage & Control Video & Testing Workflow
IBM Spectrum
Archive
LTFS Tape
Library
<
other
MiL / SiL
HPC environments IBM Spectrum
Scale
IBM Cloud Object
Storage
The
foundation
Orchestration
Intelligence
20. Recap
Primary:
• High throughput for analysis and training
• Scalable due to high data growth
• Low cost long term storage
Secondary:
• Automated archiving
• Data rescillency
• Availabilty
GB/s
Flexible
Commodity components
Built in intelligence
Data integrity check
Multi sites
21. First thing to consider, storage virtualisation
A B C D
SAN / LAN
Virtualisation
Virtualisation
• Availability
• Reliability
• Performance
• Ease of use
• Automation
• Consolidation
• Hardware agnostic
• Utilisation
• ”Built in AI”
Client
Users and
applicationsCompute
Big Data
Analytics
22. IBM Spectrum Storage Family
FlashSystem
Any Storage
Private, Public or Hybrid Cloud
Spectrum LSF
Spectrum Symphony
Spectrum Conductor
Analytics-driven data management to reduce costs
by up to 50 percent
Optimized data protection to reduce backup costs
by up to 53 percent
Fast data retention that reduces TCO for active
archive data by up to 90%
Virtualization of mixed environments stores up to
5x more data
Enterprise storage for cloud deployed in minutes
instead of months
High-performance, highly scalable storage for
unstructured data
Web-scale secure Object Storage
Data Where And When You Need It
Copy Data Management For Modern IT
Platform computing
23. Spectrum Scale topology
Global namespace
IBM Spectrum Scale
Automated encrypted data placement and data migration
SMB/CIFSNFSPOSIX HDFS Controller
Disk Tape Storage Rich
Servers
Flash
On/Off Premise
OpenStack
Cinder Swift
Glance Manila
Transparent
Cloud Tiering
Site B
Site A
Site C
Cloud Data
Sharing Users and
applications
iSCSI
GB/s
24. Software Only Solution Bundles Off-premises
Software license
Can be deployed on standard hardware
Pre-packaged with IBM Spectrum Scale Software,
Spectrum Scale RAID, I/O servers, drives, support &
subscription
Deploy Spectrum Scale in
IBM Softlayer (Whitepaper)
High Performance Computing offerings with
Spectrum Scale
Spectrum Scale Deployment Options
+
30. IBM Spectrum Storage Family
FlashSystem
Any Storage
Private, Public or Hybrid Cloud
Spectrum LSF
Spectrum Symphony
Spectrum Conductor
Analytics-driven data management to reduce costs
by up to 50 percent
Optimized data protection to reduce backup costs
by up to 53 percent
Fast data retention that reduces TCO for active
archive data by up to 90%
Virtualization of mixed environments stores up to
5x more data
Enterprise storage for cloud deployed in minutes
instead of months
High-performance, highly scalable storage for
unstructured data
Web-scale secure Object Storage
Data Where And When You Need It
Copy Data Management For Modern IT
Platform computing
31. FILE STORAGE OBJECT STORAGE
• Stores hundreds of millions of files
• File system hierarchy
• Can be complex to scale
• Best for file based workflows
• I/O Performance
• Low Latency access
• Structured to be understood by humans
• File system maintains metadata
• Stores hundreds of billions of objects
• One storage pool, Object IDs
• Scales uniformly
• Low TCO
• High Latency access
• Structured to be understood by applications
• Application maintains metadata
32
What is object storage?
S3
Data Object ID
Put
Get
1
2