The Effectiveness, Efficiency and Legitimacy of Outsourcing Your Data DataCentred
Presentation given by our CEO Mike Kelly at this year's Excellence in Policing conference talking about the benefits of cloud computing and the Effectiveness, Efficiency and Legitimacy of outsourcing data. The presentation looks at the long term trends supporting the adoption of cloud technologies and dispels some of the myths and reasons why not to adopt cloud.
The presentation concludes with an examination of the benefits of utilising cloud technology and examines how best to adopt a cloud approach.
An Introduction to Red Hat Enterprise Linux OpenStack PlatformYandex
In this presentation, Rhys Oxenham will introduce attendees to Red Hat Enterprise Linux OpenStack Platform, Red Hat’s OpenStack distribution, covering the technology, the use-cases, and existing customer success stories. Attendees will gain valuable insight into how OpenStack works and how organisations can make the most of OpenStack technology.
Coding Secure Infrastructure in the Cloud using the PIE frameworkJames Wickett
At National Instruments, we have developed an automation and provisioning framework called PIE (Programmable Infrastructure Environment) that we use daily on our devops team. Similar tools are available such as chef or puppet, but what makes PIE unique is its ability to work in multi-cloud deployments (Azure and AWS) along with multiple node OS types (linux and windows). It uses zookeeper to keep state and track dependencies across nodes and services.
When building PIE we actively considered how to implement it in a Rugged way for a DevOps team. As noted in the deck on slide 68, we are Rugged by Design and Devops by Culture. We see these as intersecting domains that have the ability to impact each other. For more info see ruggeddevops.org
AWS re:Invent 2016: Bringing Deep Learning to the Cloud with Amazon EC2 (CMP314)Amazon Web Services
Algorithmia is a startup with a mission to make state of the art machine learning discoverable by everyone&emdash;they offer the largest algorithm marketplace in the world, with over 2500 algorithms supporting tens of thousands of application developers. Algorithma is the first company to make deep learning, one of the most conceptually difficult areas of computing, accessible to any company via microservices. In this session, you learn how this startup has selected and optimized Amazon EC2 instances for various algorithms (including the latest generation of GPU optimized instances), to create a flexible and scalable platform. They also share their architecture and best practices for getting any computationally-intensive application started quickly.
HPC and cloud distributed computing, as a journeyPeter Clapham
Introducing an internal cloud brings new paradigms, tools and infrastructure management. When placed alongside traditional HPC the new opportunities are significant But getting to the new world with micro-services, autoscaling and autodialing is a journey that cannot be achieved in a single step.
The Effectiveness, Efficiency and Legitimacy of Outsourcing Your Data DataCentred
Presentation given by our CEO Mike Kelly at this year's Excellence in Policing conference talking about the benefits of cloud computing and the Effectiveness, Efficiency and Legitimacy of outsourcing data. The presentation looks at the long term trends supporting the adoption of cloud technologies and dispels some of the myths and reasons why not to adopt cloud.
The presentation concludes with an examination of the benefits of utilising cloud technology and examines how best to adopt a cloud approach.
An Introduction to Red Hat Enterprise Linux OpenStack PlatformYandex
In this presentation, Rhys Oxenham will introduce attendees to Red Hat Enterprise Linux OpenStack Platform, Red Hat’s OpenStack distribution, covering the technology, the use-cases, and existing customer success stories. Attendees will gain valuable insight into how OpenStack works and how organisations can make the most of OpenStack technology.
Coding Secure Infrastructure in the Cloud using the PIE frameworkJames Wickett
At National Instruments, we have developed an automation and provisioning framework called PIE (Programmable Infrastructure Environment) that we use daily on our devops team. Similar tools are available such as chef or puppet, but what makes PIE unique is its ability to work in multi-cloud deployments (Azure and AWS) along with multiple node OS types (linux and windows). It uses zookeeper to keep state and track dependencies across nodes and services.
When building PIE we actively considered how to implement it in a Rugged way for a DevOps team. As noted in the deck on slide 68, we are Rugged by Design and Devops by Culture. We see these as intersecting domains that have the ability to impact each other. For more info see ruggeddevops.org
AWS re:Invent 2016: Bringing Deep Learning to the Cloud with Amazon EC2 (CMP314)Amazon Web Services
Algorithmia is a startup with a mission to make state of the art machine learning discoverable by everyone&emdash;they offer the largest algorithm marketplace in the world, with over 2500 algorithms supporting tens of thousands of application developers. Algorithma is the first company to make deep learning, one of the most conceptually difficult areas of computing, accessible to any company via microservices. In this session, you learn how this startup has selected and optimized Amazon EC2 instances for various algorithms (including the latest generation of GPU optimized instances), to create a flexible and scalable platform. They also share their architecture and best practices for getting any computationally-intensive application started quickly.
HPC and cloud distributed computing, as a journeyPeter Clapham
Introducing an internal cloud brings new paradigms, tools and infrastructure management. When placed alongside traditional HPC the new opportunities are significant But getting to the new world with micro-services, autoscaling and autodialing is a journey that cannot be achieved in a single step.
UI Dev in Big data world using open sourceTech Triveni
He will be sharing his last 10 years of experience in UI Development for Big Data Analytics & ML world using available open-source plethora in the market. How 'UI dev' needs to target big data problems?
Key points to consider while choosing any open-source framework/library for the big data world.
Do you need to write a custom framework or use ready-made open source, when what to choose?
How dev can leverage open source frameworks like Angular, REACT to making big data apps faster?
How you can extend open-source BI tools like Kibana, superset graphana to make UI development tool?
How to show network big data using open source graph libraries?
How to deal with real-time data in Big data UI?
Why use & contribute to open source?
Design UI for future as in Big data world customer problems keep changing with time. Showcasing demo for our real customer's problems, how we achieved using these open source libraries.
Supporting Research through "Desktop as a Service" models of e-infrastructure...David Wallom
Keynote presentation given 13/9/16 @ ESA Earth Observation Open Science workshop 2016.
"The rise in cloud computing as an e-infrastructure model is one that has the power to democratise access to computational and data resources throughout the research communities. We have seen the difference that Infrastructure as a Service (IaaS) has made for different communities and are now only beginning to understand what different models further up the stack can make. It is also becoming clear that with the increase in research data volumes, the number of sources and the possibility of utilising data from different regulatory regimes that a different model of how analysis is performed on the data is possible. Utilising a "Desktop as a Service" model, with community focused applications installed on a common and well understood virtual system image that is directly connected to community relevant data allows the researcher to no longer have to consider moving data but only the final analysed results. This massively simplifies both the user model and the data and resource owner model. We will consider the specific example of the Environmental Ecomics Synthesis Cloud and how it could easily be generalised to other areas."
Application Virtualization, University of New HampshireTony Austwick
Presentation to EDUCAUSE, NERCOMP by David Blezzard from University of New Hampshire.
More here http://www.educause.edu/nercomp-annual-conference/2016/2016/streamlining-application-deployment-and-management-through-virtualization
We will report how using the Application Jukebox application virtualization solution at the University of New Hampshire has saved time and simplified management of public and department-specific computer labs and our VDI environment. Traditionally, we used a combination of imaging and then various scripted deployments to install applications. With application virtualization, we have been able to move over a hundred Windows applications into virtual packages that can be "installed" by users on demand. We will discuss the general problem of managing large numbers of complex applications across a range of computer environments while also trying to be flexible and responsive to users.
Outcomes: Understand how application virtualization can overcome issues with deploying software * Appreciate the capabilities of Application Jukebox to be able to assess its utility to your campus * Obtain a model for evaluating a tool's potential savings
Brian Brownlow is an experienced senior analyst programmer for Mayo Clinic. He is made a workshop presentation at the 2014 BDPA Technology Conference on the topic, 'Big Data Implementation - Mayo Clinic Case Study'. This presentation will show part of the Mayo Clinic story on the embarking of an exploration of the application of `Big Data' technologies. `Big Data' is seen as one set of tools that can be used to enhance medical research, medical education and practice management. Mayo Clinic is always searching for better, faster and cheaper ways to use its data to improve patient care and sustain financial outcomes in a challenging reimbursement environment. Our approach uses several components that are open source and combines them with data from various sources to provide information to decision makers in near real time. We have created a center of `Big Data' excellence using in-house staff and vendor engagements. `Big Data' is one element of our Enterprise Data Trust framework.
Interested in how you can get a Private Cloud solution that just works efficiently? With Platform9 and Solidfire, users can get the self-service automation of OpenStack combined with the incredible speed of flash storage. Learn more here!
Sanger, upcoming Openstack for Bio-informaticiansPeter Clapham
Delivery of a new Bio-informatics infrastructure at the Wellcome Trust Sanger Center. We include how to programatically create, manage and provide providence for images used both at Sanger and elsewhere using open source tools and continuous integration.
How does one choose to architect a system that has a Microservice / REST API endpoints? There are many solutions out there. Some are better than others. Should state be held in a server side component, or externally? Generally we are told this is not a good practice for a Cloud Native system, when the 12-factor guidelines seem to be all about stateless containers, but is it? It’s unclear and this confusion may lead to poor technology stack choices that are impossible or extremely hard to change later on as your system evolves in terms of demand and performance.
While stateless systems are easier to work with, the reality is that we live in a stateful world, so we have to handle the state of data accordingly to ensure data integrity beyond securing it.
We will examine and demonstrate the fundamentals of a Cloud Native system with Stateful Microservices that’s built with Open Liberty and MicroProfile in Kubernetes.
This talk was given at a workshop entitled "Cybersecurity Engagement in a Research Environment" at Rady School of Management at UCSD. The workshop was organized by Michael Corn, the UCSD CISO. It tries to provoke discussion around the cybersecurity features and requirements of international science collaborations, as well as more generally, federated cyberinfrastructure systems.
UI Dev in Big data world using open sourceTech Triveni
He will be sharing his last 10 years of experience in UI Development for Big Data Analytics & ML world using available open-source plethora in the market. How 'UI dev' needs to target big data problems?
Key points to consider while choosing any open-source framework/library for the big data world.
Do you need to write a custom framework or use ready-made open source, when what to choose?
How dev can leverage open source frameworks like Angular, REACT to making big data apps faster?
How you can extend open-source BI tools like Kibana, superset graphana to make UI development tool?
How to show network big data using open source graph libraries?
How to deal with real-time data in Big data UI?
Why use & contribute to open source?
Design UI for future as in Big data world customer problems keep changing with time. Showcasing demo for our real customer's problems, how we achieved using these open source libraries.
Supporting Research through "Desktop as a Service" models of e-infrastructure...David Wallom
Keynote presentation given 13/9/16 @ ESA Earth Observation Open Science workshop 2016.
"The rise in cloud computing as an e-infrastructure model is one that has the power to democratise access to computational and data resources throughout the research communities. We have seen the difference that Infrastructure as a Service (IaaS) has made for different communities and are now only beginning to understand what different models further up the stack can make. It is also becoming clear that with the increase in research data volumes, the number of sources and the possibility of utilising data from different regulatory regimes that a different model of how analysis is performed on the data is possible. Utilising a "Desktop as a Service" model, with community focused applications installed on a common and well understood virtual system image that is directly connected to community relevant data allows the researcher to no longer have to consider moving data but only the final analysed results. This massively simplifies both the user model and the data and resource owner model. We will consider the specific example of the Environmental Ecomics Synthesis Cloud and how it could easily be generalised to other areas."
Application Virtualization, University of New HampshireTony Austwick
Presentation to EDUCAUSE, NERCOMP by David Blezzard from University of New Hampshire.
More here http://www.educause.edu/nercomp-annual-conference/2016/2016/streamlining-application-deployment-and-management-through-virtualization
We will report how using the Application Jukebox application virtualization solution at the University of New Hampshire has saved time and simplified management of public and department-specific computer labs and our VDI environment. Traditionally, we used a combination of imaging and then various scripted deployments to install applications. With application virtualization, we have been able to move over a hundred Windows applications into virtual packages that can be "installed" by users on demand. We will discuss the general problem of managing large numbers of complex applications across a range of computer environments while also trying to be flexible and responsive to users.
Outcomes: Understand how application virtualization can overcome issues with deploying software * Appreciate the capabilities of Application Jukebox to be able to assess its utility to your campus * Obtain a model for evaluating a tool's potential savings
Brian Brownlow is an experienced senior analyst programmer for Mayo Clinic. He is made a workshop presentation at the 2014 BDPA Technology Conference on the topic, 'Big Data Implementation - Mayo Clinic Case Study'. This presentation will show part of the Mayo Clinic story on the embarking of an exploration of the application of `Big Data' technologies. `Big Data' is seen as one set of tools that can be used to enhance medical research, medical education and practice management. Mayo Clinic is always searching for better, faster and cheaper ways to use its data to improve patient care and sustain financial outcomes in a challenging reimbursement environment. Our approach uses several components that are open source and combines them with data from various sources to provide information to decision makers in near real time. We have created a center of `Big Data' excellence using in-house staff and vendor engagements. `Big Data' is one element of our Enterprise Data Trust framework.
Interested in how you can get a Private Cloud solution that just works efficiently? With Platform9 and Solidfire, users can get the self-service automation of OpenStack combined with the incredible speed of flash storage. Learn more here!
Sanger, upcoming Openstack for Bio-informaticiansPeter Clapham
Delivery of a new Bio-informatics infrastructure at the Wellcome Trust Sanger Center. We include how to programatically create, manage and provide providence for images used both at Sanger and elsewhere using open source tools and continuous integration.
How does one choose to architect a system that has a Microservice / REST API endpoints? There are many solutions out there. Some are better than others. Should state be held in a server side component, or externally? Generally we are told this is not a good practice for a Cloud Native system, when the 12-factor guidelines seem to be all about stateless containers, but is it? It’s unclear and this confusion may lead to poor technology stack choices that are impossible or extremely hard to change later on as your system evolves in terms of demand and performance.
While stateless systems are easier to work with, the reality is that we live in a stateful world, so we have to handle the state of data accordingly to ensure data integrity beyond securing it.
We will examine and demonstrate the fundamentals of a Cloud Native system with Stateful Microservices that’s built with Open Liberty and MicroProfile in Kubernetes.
This talk was given at a workshop entitled "Cybersecurity Engagement in a Research Environment" at Rady School of Management at UCSD. The workshop was organized by Michael Corn, the UCSD CISO. It tries to provoke discussion around the cybersecurity features and requirements of international science collaborations, as well as more generally, federated cyberinfrastructure systems.
Connector Corner: Automate dynamic content and events by pushing a buttonDianaGray10
Here is something new! In our next Connector Corner webinar, we will demonstrate how you can use a single workflow to:
Create a campaign using Mailchimp with merge tags/fields
Send an interactive Slack channel message (using buttons)
Have the message received by managers and peers along with a test email for review
But there’s more:
In a second workflow supporting the same use case, you’ll see:
Your campaign sent to target colleagues for approval
If the “Approve” button is clicked, a Jira/Zendesk ticket is created for the marketing design team
But—if the “Reject” button is pushed, colleagues will be alerted via Slack message
Join us to learn more about this new, human-in-the-loop capability, brought to you by Integration Service connectors.
And...
Speakers:
Akshay Agnihotri, Product Manager
Charlie Greenberg, Host
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Accelerate your Kubernetes clusters with Varnish CachingThijs Feryn
A presentation about the usage and availability of Varnish on Kubernetes. This talk explores the capabilities of Varnish caching and shows how to use the Varnish Helm chart to deploy it to Kubernetes.
This presentation was delivered at K8SUG Singapore. See https://feryn.eu/presentations/accelerate-your-kubernetes-clusters-with-varnish-caching-k8sug-singapore-28-2024 for more details.
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 4
Panel: Open Infrastructure for an Open Society: OSG, Commercial Clouds, and Bring-Your-Own-Resources
1. Open Infrastructure for an Open Society:
OSG, Commercial Clouds, and
Bring-Your-Own-Resources
4NRP
February 9th, 2023
2. • James Deaton
• Executive Director, Great Plains Network
• Derek Weitzel
• Research Assistant Professor, University of Nebraska-Lincoln,
OSG, PATh, PNRP
• Jeremy Evert
• Associate Professor, Computer Science, Southwestern
Oklahoma State University
• Igor Sfiligoi
• Lead Scientific Software Developer and Researcher, San Diego
Supercomputer Center
3. Open Infrastructure
Derek Weitzel – University of Nebraska-Lincoln
(Strictly Derek’s Opinions)
This project is supported by the National Science Foundation under Cooperative
Agreements OAC-2112167,. Any opinions, findings, conclusions or
recommendations expressed in this material are those of the authors and do not
necessarily reflect the views of the National Science Foundation.
5. How is NRP “Open Infrastructure”
•All components are Open Source
• Kubernetes and containers
•Anyone can contribute resources
•Anyone can use the resources
•Documented Interfaces
•Resources were ”seeded” through various grants
• But grew with contributions from users
7. How is OSG “Open Infrastructure”
•All components are Open Source
• HTCondor and various tools
•Anyone can contribute resources
•Anyone can utilize the resources
•Interfaces are documented: osg-htc.org/docs
•Resources are ”seeded” by organizations such as LHC, and now CC*.
• But have grown through contributions of users
9. How is OSDF “Open Infrastructure”
• All components are Open Source
• Anyone can contribute resources
• Interfaces are documented
• Resources were “seeded” by various grants and Internet2
• But have grown by contributions from users, and soon CC*
10. Leveraging NRP on a smaller
campus
Jeremy Evert
Associate Professor, Southwestern Oklahoma State University
February 9th, 2023
11. About Southwestern Oklahoma State University
● 10th in the state in enrollment behind 2 community colleges
● 5,000 students across two campuses
○ Formerly a teaching college
○ Formerly a tribal serving institution
● Non-PhD Granting
● Serves a portion of the minorities in the area
● Around 200 full time faculty and about 60% hold a terminal degree
12. Bringing Our Own Resource
● SWOSU had: 200 Sq. Ft. Server closet, 5 ton A/C, 42U rack
○ NSF CC* switch
● Dell Server, 96 AMD cores, some memory, spinning disk, small gpu
● San Diego team guided SWOSU through NVMe storage upgrade
● Faculty installed Ubuntu for a base OS
● OneNet (State ISP) helped troubleshoot network
● San Diego deployed Nautilus node
● James Deaton enabled user authentication through
github.com/SWOSU
● OneNet (state ISP) and SWOSU central IT provided an alias for
jupyter.swosu.edu
13. Engage and empower every SWOSU student
● SWOSU Computer Science Discrete Structures assignment: join
GitHub.com/swosu
● Students are pointed to our server as soon as they start running codes
that heat up their laptop
● Promoted on every syllabus I have
14. Engage and empower every elementary and high
school student and researcher
● SWOSU invites area technology teachers for a weeklong camp
○ Esports, graphic design, Microsoft, and programming
● Full day on teaching programming
● Teachers run jobs on jupyter.swosu.edu
15. Supporting SWOSU for the next 10 years
● Enable more science drivers
○ Physics, Math, Biology, and other Compute Science faculty
● Partner with SWOSU Education Department to integrate more of the
Campus Champions / Carpentries type trainings into new primary
education curriculum
● Leverage mentors from NRP / Great Plains Network / OneNet /
OneOklahoma Cyber Infrastructure Initiative to keep growing
○ Look to NSF CC* or small school MRI to expand current platform
16. Please consider a weekly statewide call
● Set up a email list
● Encourage key players to join
● Allowing staff to show up and make connections
● Look for ways to add value to the individuals and larger community
● Connection to a larger community enables faculty at smaller schools
17. Open Infrastructure for an Open Society:
Commercial Clouds
Igor Sfiligoi
University of California San Diego
San Diego Supercomputer Center
Fourth National Research Platform (4NRP) – Feb 9th, 2023 1
18. Who cares about Commercial Clouds?
• Seems like everyone in industry is moving there!
• Not really, but it does look like it
• The big players have huge compute capacity
• Personally verified I can access 50k GPUs
• Others demonstrated access to several million CPU cores
• They have a large variety of compute resources
• Many x86 variants and several ARM CPUs
• Many GPU variants
• AI accelerators and FPGAs
• Great networking setups (both WAN and HPC-class LAN/Infiniband)
2
20. Who cares about Commercial Clouds?
• Seems like everyone in industry is moving there!
• Not really, but it does look like it
• The big players have huge compute capacity
• Personally verified I can access 50k GPUs
• Others demonstrated access to several million CPU cores
• They have a large variety of compute resources
• Many x86 variants and several ARM CPUs
• Many GPU variants
• AI accelerators and FPGAs
• Great networking setups (both WAN and HPC-class LAN/Infiniband)
4
21. Often have new HW
available before
you can buy it
5
Also,
Cloud-exclusive
HW variants
• CPUs
• INTEL Saphire Rapids available
on Google Cloud now
• AMD EPYC Milan-X available on
Azure now
• AMD EPYC Genoa in preview
• NVIDIA GPUs
• A10s were available in AWS
in 2021
• ARM CPUs
• AWS has its own ARM CPU
• Azure and Google regular one
• AI Accelerators
• AWS has Inferentia
• Google has TPUs
• AWS also offers Habana Gaudi
• FPGAs
• AWS had FPGAs since forever
22. Who cares about Commercial Clouds?
• Seems like everyone in industry is moving there!
• Not really, but it does look like it
• The big players have huge compute capacity
• Personally verified I can access 50k GPUs
• Others demonstrated access to several million CPU cores
• They have a large variety of compute resources
• Many x86 variants and several ARM CPUs
• Many GPU variants
• AI accelerators and FPGAs
• Great networking setups (both WAN and HPC-class LAN/Infiniband)
6
24. Pros and cons of Commercial Clouds
• Pros:
• See previous slide
• No need to go through allocation processes… all you need is money
• Cons:
• You need money
• And lots of it
• ”Regular”, on-demand Cloud computing is expensive
• Anywhere between 3x and 10x what you would pay on-prem on 24/7 basis
• Spot pricing is almost comparable to on-prem, but only useful for preemptible work
• Easy to get in, hard to get out
• Pricing optimized to let data get in cheaply, but expensive to move out
• No automatic price caps, easy to overspend
8
26. 10
vs
Private jet
Commercial airline
Ticket bought 2 months in advance
in economy class
through your travel department
Both will get you from A to B
Which one would you pick?
27. Who should consider Commercial Cloud?
• Flexible/urgent computing
• Hard to beat the scalability of the clouds
• Costs acceptable for short spikes
• Prototyping, R&D
• The variety of HW available in the clouds is hard to match
• Instant access, no-contention drastically raises productivity
• Ultra-High-Availability services
• Hard to beat the breath of Cloud deployments
• Many large datacenters, proven track record
11
28. Is Commercial Cloud easy to use?
• Yes and no
• Provide enormous flexibility
• You can do virtually everything you could do with your personal server
• But that can be daunting for non-IT users
• Lots of support services
• No need to reinvent the wheel, just pick one
• Finding what you need can be a challenge, lots of competing options
• Cloud providers invest a lot in the user interfaces
• More intuitive than anything you will find on-prem
• But each provider has its own flavor
• How do you mix on-prem and Cloud resources?
12
29. Facilitating Cloud access for science users
• CloudBank
• Account management and monitoring (I love their spend/budget tracking!)
• Extensive documentation/training
• Integrate with OSG/PATh/HTCondor ecosystem
• IT-savvy support staff can easily add cloud resources to a HTCondor pool
• Users see only HTCondor, cloud HW no different that on-prem HW
• Kubernetes (k8s) to the rescue
• All Cloud Providers expose a Kubernetes interface, too
• Cloud k8s feels like on-prem k8s (at least for compute)
• Kubernetes federation can make it completely transparent, e.g. from Nautilus
13