The document discusses the transition from workgroup-centric computing architectures to network-centric architectures. It describes how Google has developed a network-centric model based on 4 key principles: 1) building a protected information network, 2) offering universal connectivity, 3) maintaining shared data models, and 4) requiring continued upgrading and innovation. The network-centric model moves labor, capital and control to centralized networks rather than individual workgroups and users. This allows for more efficient, scalable and secure systems.
Jeff Hammerbacher gave a talk on evolving global information platforms. He discussed Facebook's data infrastructure evolution from using MySQL to adding a Hadoop tier. He then talked about challenges of managing large data systems and how the cloud can help. However, clouds are not optimized for data-intensive workloads and have issues around pricing, security, and data transfers. Future evolutions may include tiered storage, streaming analytics, tighter integration between data collection and reporting, and accommodating more workload types.
Learn about Combining IBMReal-timeCompression andIBM ProtecTIER Deduplication where combining storage optimization technologies helps achieve compelling results. For more information on IBM Systems, visit http://ibm.co/RKEeMO.
Visit the official Scribd Channel of IBM India Smarter Computing at http://bit.ly/VwO86R to get access to more documents.
This document discusses increasing datacenter efficiency with Fujitsu and NetApp technologies. It provides an overview of Fujitsu's PRIMERGY servers and storage solutions, such as the PRIMERGY BX400 data center in a box. It also summarizes NetApp's storage portfolio including FAS storage systems, V-Series controllers, and data efficiency technologies like deduplication, thin provisioning, snapshots, and thin replication. Case studies on BT show how NetApp helped consolidate servers, improve utilization, reduce power consumption, and improve backup times.
The document discusses a project called Our City, Our Music from 2006-2008 that aimed to have an impact on how people experience a city. It embedded narrative and extended social space by combining blogs, mobile phones, social mapping and GPS to interact with the public. The project allowed people to share stories about the city related to roads, love, philosophy, heartbreak and music.
(A2 Media Studies - Megan Smith) The Chosen Locations (Planning) Megan Smith
The document discusses potential filming locations for a school-based film trailer, including photographs of areas in the author's school that could be used. It describes plans to film scenes in the school courtyard, in front of the stage, in a classroom used for studying, the music room where the band practices, the common room where the main character's friend group hangs out, and outside a local music shop relevant to one of the film's themes. The locations were selected because they fit the storylines and themes of the film and capture the realistic school environment.
Evaluation Question 1 - Megan Smith A2 MediaMegan Smith
This document discusses the conventions of teaser and theatrical film trailers, and how the student's media product for their coursework either challenges or follows these conventions. For their coursework, the student created a theatrical film trailer for their independent film "Fleeting Youth". The student made the decision to create a theatrical rather than teaser trailer because independent films typically do not have teaser trailers. The student aimed to follow most conventions of a theatrical trailer, such as giving viewers a sense of the general story and leaving them wanting more, while their trailer was slightly shorter than typical. The student also wanted their trailer and other promotional materials to match the conventions of a social realism genre.
"Keeping House in Nepal" by Megan Smith for New Mobility magazine, February 2011. Used with permission. For accessible version, visit http://www.miusa.org/ncde/stories/smithnm.
Jeff Hammerbacher gave a talk on evolving global information platforms. He discussed Facebook's data infrastructure evolution from using MySQL to adding a Hadoop tier. He then talked about challenges of managing large data systems and how the cloud can help. However, clouds are not optimized for data-intensive workloads and have issues around pricing, security, and data transfers. Future evolutions may include tiered storage, streaming analytics, tighter integration between data collection and reporting, and accommodating more workload types.
Learn about Combining IBMReal-timeCompression andIBM ProtecTIER Deduplication where combining storage optimization technologies helps achieve compelling results. For more information on IBM Systems, visit http://ibm.co/RKEeMO.
Visit the official Scribd Channel of IBM India Smarter Computing at http://bit.ly/VwO86R to get access to more documents.
This document discusses increasing datacenter efficiency with Fujitsu and NetApp technologies. It provides an overview of Fujitsu's PRIMERGY servers and storage solutions, such as the PRIMERGY BX400 data center in a box. It also summarizes NetApp's storage portfolio including FAS storage systems, V-Series controllers, and data efficiency technologies like deduplication, thin provisioning, snapshots, and thin replication. Case studies on BT show how NetApp helped consolidate servers, improve utilization, reduce power consumption, and improve backup times.
The document discusses a project called Our City, Our Music from 2006-2008 that aimed to have an impact on how people experience a city. It embedded narrative and extended social space by combining blogs, mobile phones, social mapping and GPS to interact with the public. The project allowed people to share stories about the city related to roads, love, philosophy, heartbreak and music.
(A2 Media Studies - Megan Smith) The Chosen Locations (Planning) Megan Smith
The document discusses potential filming locations for a school-based film trailer, including photographs of areas in the author's school that could be used. It describes plans to film scenes in the school courtyard, in front of the stage, in a classroom used for studying, the music room where the band practices, the common room where the main character's friend group hangs out, and outside a local music shop relevant to one of the film's themes. The locations were selected because they fit the storylines and themes of the film and capture the realistic school environment.
Evaluation Question 1 - Megan Smith A2 MediaMegan Smith
This document discusses the conventions of teaser and theatrical film trailers, and how the student's media product for their coursework either challenges or follows these conventions. For their coursework, the student created a theatrical film trailer for their independent film "Fleeting Youth". The student made the decision to create a theatrical rather than teaser trailer because independent films typically do not have teaser trailers. The student aimed to follow most conventions of a theatrical trailer, such as giving viewers a sense of the general story and leaving them wanting more, while their trailer was slightly shorter than typical. The student also wanted their trailer and other promotional materials to match the conventions of a social realism genre.
"Keeping House in Nepal" by Megan Smith for New Mobility magazine, February 2011. Used with permission. For accessible version, visit http://www.miusa.org/ncde/stories/smithnm.
Tape continues to be an important storage solution due to its low cost and high capacity. It plays a key role in backup and archiving, especially for large amounts of cold data. Recent technology demonstrations by IBM show that tape capacity can continue to increase significantly at around 40% per year, with a demonstration of 220TB per cartridge. Tape provides unmatched security as an offline storage medium and is much more reliable than disk storage. With innovations like LTFS, tape is also easier to use for a wider range of applications.
The document discusses the benefits of using tape storage for backup and archiving large amounts of data. Tape provides low cost, high capacity storage when compared to disk and flash alternatives. Features such as air gaps between live systems and offline tape backups provide strong protection against ransomware and other cyber threats. With continued improvements in areal density, a single tape cartridge can now hold over 200 terabytes of data, growing cheaper and more scalable over time. Tape remains a critical technology for cost-effectively storing the massive amounts of cold and archived data being generated.
1) The document discusses IBM's Big Data platform which provides solutions for analyzing massive amounts of structured and unstructured data.
2) It includes offerings for Hadoop, streaming analytics, data warehousing, in-memory databases, and tools for information integration and governance.
3) The platform aims to help customers accelerate insights from their data, simplify big data adoption, and implement enterprise-grade big data systems.
RAMCloud: Scalable Datacenter Storage Entirely in DRAMGeorge Ang
The document summarizes the RAMCloud project which aims to create large-scale storage systems entirely in DRAM. RAMCloud would provide storage for datacenters with thousands of commodity servers, storing all data in RAM for high throughput and low latency access of under 10 microseconds. It motivates the project by discussing limitations of disk-based storage and benefits of low latency for enabling new applications.
This document discusses new technologies that are emerging beyond Hadoop and the original "Google canon" to process big data more efficiently. It introduces Percolator for incremental processing, Dremel for ad-hoc queries on nested data, and Pregel for large-scale graph processing. These new systems provide benefits like lower latency, easier analysis of diverse data types, and scalability to analyze graphs with billions of vertices.
Why z/OS is a Great Platform for Developing and Hosting APIsTeodoro Cipresso
z/OS Connect Enterprise Edition makes it possible to create new value by enabling the creation of APIs that bring together multiple, disparate, z subsystem assets.
The document discusses MapReduce, including its programming model, internal framework, and improvements. It describes MapReduce as a programming model and framework that allows parallel processing of large datasets across commodity machines. The map function processes input key-value pairs to generate intermediate pairs, and the reduce function combines values for each key. The framework automatically parallelizes jobs and provides fault tolerance.
S de2784 footprint-reduction-edge2015-v2Tony Pearson
Data footprint reduction is the umbrella term for technologies like Thin Provisioning, Space-efficient snapshots, Data deduplication, and Real-time Compression.
Flash Ahead: IBM Flash System Selling PointCTI Group
IBM's FlashSystem storage is designed to radically accelerate critical applications by providing consistent low latency flash performance. It can integrate with existing disk arrays to offload I/O-intensive workloads while improving overall performance. FlashSystem utilizes IBM's flash technology and software to deliver microsecond response times for applications such as databases, virtual infrastructures, and cloud computing. The FlashSystem family includes the all-flash 710, 720, 810, and 820 models that are optimized for performance, capacity, and mixed workloads.
The document summarizes four presentations from the USENIX NSDI 2016 conference session on resource sharing:
1. "Ernest: Efficient Performance Prediction for Large-Scale Advanced Analytics" proposes a framework that uses results from small training jobs to efficiently predict performance of data analytics workloads in cloud environments and reduce the number of required training jobs.
2. "Cliffhanger: Scaling Performance Cliffs in Web Memory Caches" presents algorithms to dynamically allocate memory across queues in Memcached to smooth out performance cliffs and potentially save memory usage.
3. "FairRide: Near-Optimal, Fair Cache Sharing" introduces a caching policy that provides isolation guarantees, prevents strategic behavior, and
Snapshots have been a key feature of primary storage infrastructures that IT professionals have relied on for years. But storage systems have traditionally been able to support only a limited number of active snapshots. And snapshots, being pointers and not actual data, are also susceptible to a primary storage system failure. As a result, most IT professionals use snapshots sparingly for protecting data. In this webinar Storage Switzerland and Nexenta show you how primary storage can be architected so that snapshots are able to meet almost all of the data protection requirements an organization has.
Compare Clustering Methods for MS SQL ServerAlexDepo
Clustering is very important technology for High Availability and it is important for DBA to understand benefits and pitfalls. With very few available techniques and a lot of gray areas right decision might help to avoid extra costs. Presentation is unveiling clustering basics, reviews and compares clustering technologies including Microsoft, XCOTO Gridscale, and HP PolyserveMatrix. This presentation can be helpful not only to beginners but to intermediate level DBAs and infrastructure managers.
New Business Applications Powered by In-Memory Technology @MIT Forum for Supp...Paul Hofmann
Talk about 4 leading edge projects:
1) Optimal pricing for energy management, online pricing, and truck scheduling @Princeton University
2) Infinite DRAM - RAMCloud @Stanford University
Applications: Extremely low latency and very high bandwidth
a) Facebook like problems with high read AND write rate,
b) advanced analytics, c) what-if scenarios for demand planning
3) Hybrid In-Memory Store @MIT CSAIL
4) Multithreading Real Time Event Platform @MIT Auto-ID Lab
500k events/s and millions of threads in-memory or distributed used for automatic meter reading, online billing, mobile billing and Smart Grid
Scalable Relational Databases with Amazon Aurora. Madrid Summit 2019javier ramirez
Amazon Aurora is a MySQL/PostgreSQL compatible managed database created by Amazon Web Services. Apart from being a managed service, it has better performance, better availability, and some nice extra features. Learn in this presentation some of the internals that make Aurora possible.
Research and technology explosion in scale-out storageJeff Spencer
A view of the directions storage is taking in science & technology from Ryan Sayre, technical strategist in the office of the CTO for EMC Isilon, using examples from recent work in life science genomics and other industries taking advantage of the combination of extreme computing (HPC) and big data. As presented at the Bull sponsored Science & Innovation 2013 conference Westminster.
This is the presentation I delivered on Hadoop User Group Ireland meetup in Dublin on Nov 28 2015. It covers at glance the architecture of GPDB and most important its features. Sorry for the colors - Slideshare is crappy with PDFs
Theresa Melvin, HP Enterprise - IOT/AI/ML at Hyperscale - how to go faster wi...Aerospike
The document discusses hyperscale AI and how to achieve faster performance with a smaller footprint. It outlines an end-to-end hyperscale design from HPE that spans from ingestion at the edge to analytics in the core and cloud. The design utilizes technologies like persistent memory, high-performance networking and hardware, and distributed databases optimized for AI workloads. It provides examples of how this design has been implemented and tested on real-world AI and analytics problems.
Must Know Postgres Extension for DBA and Developer during MigrationMydbops
Mydbops Opensource Database Meetup 16
Topic: Must-Know PostgreSQL Extensions for Developers and DBAs During Migration
Speaker: Deepak Mahto, Founder of DataCloudGaze Consulting
Date & Time: 8th June | 10 AM - 1 PM IST
Venue: Bangalore International Centre, Bangalore
Abstract: Discover how PostgreSQL extensions can be your secret weapon! This talk explores how key extensions enhance database capabilities and streamline the migration process for users moving from other relational databases like Oracle.
Key Takeaways:
* Learn about crucial extensions like oracle_fdw, pgtt, and pg_audit that ease migration complexities.
* Gain valuable strategies for implementing these extensions in PostgreSQL to achieve license freedom.
* Discover how these key extensions can empower both developers and DBAs during the migration process.
* Don't miss this chance to gain practical knowledge from an industry expert and stay updated on the latest open-source database trends.
Mydbops Managed Services specializes in taking the pain out of database management while optimizing performance. Since 2015, we have been providing top-notch support and assistance for the top three open-source databases: MySQL, MongoDB, and PostgreSQL.
Our team offers a wide range of services, including assistance, support, consulting, 24/7 operations, and expertise in all relevant technologies. We help organizations improve their database's performance, scalability, efficiency, and availability.
Contact us: info@mydbops.com
Visit: https://www.mydbops.com/
Follow us on LinkedIn: https://in.linkedin.com/company/mydbops
For more details and updates, please follow up the below links.
Meetup Page : https://www.meetup.com/mydbops-databa...
Twitter: https://twitter.com/mydbopsofficial
Blogs: https://www.mydbops.com/blog/
Facebook(Meta): https://www.facebook.com/mydbops/
Tape continues to be an important storage solution due to its low cost and high capacity. It plays a key role in backup and archiving, especially for large amounts of cold data. Recent technology demonstrations by IBM show that tape capacity can continue to increase significantly at around 40% per year, with a demonstration of 220TB per cartridge. Tape provides unmatched security as an offline storage medium and is much more reliable than disk storage. With innovations like LTFS, tape is also easier to use for a wider range of applications.
The document discusses the benefits of using tape storage for backup and archiving large amounts of data. Tape provides low cost, high capacity storage when compared to disk and flash alternatives. Features such as air gaps between live systems and offline tape backups provide strong protection against ransomware and other cyber threats. With continued improvements in areal density, a single tape cartridge can now hold over 200 terabytes of data, growing cheaper and more scalable over time. Tape remains a critical technology for cost-effectively storing the massive amounts of cold and archived data being generated.
1) The document discusses IBM's Big Data platform which provides solutions for analyzing massive amounts of structured and unstructured data.
2) It includes offerings for Hadoop, streaming analytics, data warehousing, in-memory databases, and tools for information integration and governance.
3) The platform aims to help customers accelerate insights from their data, simplify big data adoption, and implement enterprise-grade big data systems.
RAMCloud: Scalable Datacenter Storage Entirely in DRAMGeorge Ang
The document summarizes the RAMCloud project which aims to create large-scale storage systems entirely in DRAM. RAMCloud would provide storage for datacenters with thousands of commodity servers, storing all data in RAM for high throughput and low latency access of under 10 microseconds. It motivates the project by discussing limitations of disk-based storage and benefits of low latency for enabling new applications.
This document discusses new technologies that are emerging beyond Hadoop and the original "Google canon" to process big data more efficiently. It introduces Percolator for incremental processing, Dremel for ad-hoc queries on nested data, and Pregel for large-scale graph processing. These new systems provide benefits like lower latency, easier analysis of diverse data types, and scalability to analyze graphs with billions of vertices.
Why z/OS is a Great Platform for Developing and Hosting APIsTeodoro Cipresso
z/OS Connect Enterprise Edition makes it possible to create new value by enabling the creation of APIs that bring together multiple, disparate, z subsystem assets.
The document discusses MapReduce, including its programming model, internal framework, and improvements. It describes MapReduce as a programming model and framework that allows parallel processing of large datasets across commodity machines. The map function processes input key-value pairs to generate intermediate pairs, and the reduce function combines values for each key. The framework automatically parallelizes jobs and provides fault tolerance.
S de2784 footprint-reduction-edge2015-v2Tony Pearson
Data footprint reduction is the umbrella term for technologies like Thin Provisioning, Space-efficient snapshots, Data deduplication, and Real-time Compression.
Flash Ahead: IBM Flash System Selling PointCTI Group
IBM's FlashSystem storage is designed to radically accelerate critical applications by providing consistent low latency flash performance. It can integrate with existing disk arrays to offload I/O-intensive workloads while improving overall performance. FlashSystem utilizes IBM's flash technology and software to deliver microsecond response times for applications such as databases, virtual infrastructures, and cloud computing. The FlashSystem family includes the all-flash 710, 720, 810, and 820 models that are optimized for performance, capacity, and mixed workloads.
The document summarizes four presentations from the USENIX NSDI 2016 conference session on resource sharing:
1. "Ernest: Efficient Performance Prediction for Large-Scale Advanced Analytics" proposes a framework that uses results from small training jobs to efficiently predict performance of data analytics workloads in cloud environments and reduce the number of required training jobs.
2. "Cliffhanger: Scaling Performance Cliffs in Web Memory Caches" presents algorithms to dynamically allocate memory across queues in Memcached to smooth out performance cliffs and potentially save memory usage.
3. "FairRide: Near-Optimal, Fair Cache Sharing" introduces a caching policy that provides isolation guarantees, prevents strategic behavior, and
Snapshots have been a key feature of primary storage infrastructures that IT professionals have relied on for years. But storage systems have traditionally been able to support only a limited number of active snapshots. And snapshots, being pointers and not actual data, are also susceptible to a primary storage system failure. As a result, most IT professionals use snapshots sparingly for protecting data. In this webinar Storage Switzerland and Nexenta show you how primary storage can be architected so that snapshots are able to meet almost all of the data protection requirements an organization has.
Compare Clustering Methods for MS SQL ServerAlexDepo
Clustering is very important technology for High Availability and it is important for DBA to understand benefits and pitfalls. With very few available techniques and a lot of gray areas right decision might help to avoid extra costs. Presentation is unveiling clustering basics, reviews and compares clustering technologies including Microsoft, XCOTO Gridscale, and HP PolyserveMatrix. This presentation can be helpful not only to beginners but to intermediate level DBAs and infrastructure managers.
New Business Applications Powered by In-Memory Technology @MIT Forum for Supp...Paul Hofmann
Talk about 4 leading edge projects:
1) Optimal pricing for energy management, online pricing, and truck scheduling @Princeton University
2) Infinite DRAM - RAMCloud @Stanford University
Applications: Extremely low latency and very high bandwidth
a) Facebook like problems with high read AND write rate,
b) advanced analytics, c) what-if scenarios for demand planning
3) Hybrid In-Memory Store @MIT CSAIL
4) Multithreading Real Time Event Platform @MIT Auto-ID Lab
500k events/s and millions of threads in-memory or distributed used for automatic meter reading, online billing, mobile billing and Smart Grid
Scalable Relational Databases with Amazon Aurora. Madrid Summit 2019javier ramirez
Amazon Aurora is a MySQL/PostgreSQL compatible managed database created by Amazon Web Services. Apart from being a managed service, it has better performance, better availability, and some nice extra features. Learn in this presentation some of the internals that make Aurora possible.
Research and technology explosion in scale-out storageJeff Spencer
A view of the directions storage is taking in science & technology from Ryan Sayre, technical strategist in the office of the CTO for EMC Isilon, using examples from recent work in life science genomics and other industries taking advantage of the combination of extreme computing (HPC) and big data. As presented at the Bull sponsored Science & Innovation 2013 conference Westminster.
This is the presentation I delivered on Hadoop User Group Ireland meetup in Dublin on Nov 28 2015. It covers at glance the architecture of GPDB and most important its features. Sorry for the colors - Slideshare is crappy with PDFs
Theresa Melvin, HP Enterprise - IOT/AI/ML at Hyperscale - how to go faster wi...Aerospike
The document discusses hyperscale AI and how to achieve faster performance with a smaller footprint. It outlines an end-to-end hyperscale design from HPE that spans from ingestion at the edge to analytics in the core and cloud. The design utilizes technologies like persistent memory, high-performance networking and hardware, and distributed databases optimized for AI workloads. It provides examples of how this design has been implemented and tested on real-world AI and analytics problems.
Must Know Postgres Extension for DBA and Developer during MigrationMydbops
Mydbops Opensource Database Meetup 16
Topic: Must-Know PostgreSQL Extensions for Developers and DBAs During Migration
Speaker: Deepak Mahto, Founder of DataCloudGaze Consulting
Date & Time: 8th June | 10 AM - 1 PM IST
Venue: Bangalore International Centre, Bangalore
Abstract: Discover how PostgreSQL extensions can be your secret weapon! This talk explores how key extensions enhance database capabilities and streamline the migration process for users moving from other relational databases like Oracle.
Key Takeaways:
* Learn about crucial extensions like oracle_fdw, pgtt, and pg_audit that ease migration complexities.
* Gain valuable strategies for implementing these extensions in PostgreSQL to achieve license freedom.
* Discover how these key extensions can empower both developers and DBAs during the migration process.
* Don't miss this chance to gain practical knowledge from an industry expert and stay updated on the latest open-source database trends.
Mydbops Managed Services specializes in taking the pain out of database management while optimizing performance. Since 2015, we have been providing top-notch support and assistance for the top three open-source databases: MySQL, MongoDB, and PostgreSQL.
Our team offers a wide range of services, including assistance, support, consulting, 24/7 operations, and expertise in all relevant technologies. We help organizations improve their database's performance, scalability, efficiency, and availability.
Contact us: info@mydbops.com
Visit: https://www.mydbops.com/
Follow us on LinkedIn: https://in.linkedin.com/company/mydbops
For more details and updates, please follow up the below links.
Meetup Page : https://www.meetup.com/mydbops-databa...
Twitter: https://twitter.com/mydbopsofficial
Blogs: https://www.mydbops.com/blog/
Facebook(Meta): https://www.facebook.com/mydbops/
The Microsoft 365 Migration Tutorial For Beginner.pptxoperationspcvita
This presentation will help you understand the power of Microsoft 365. However, we have mentioned every productivity app included in Office 365. Additionally, we have suggested the migration situation related to Office 365 and how we can help you.
You can also read: https://www.systoolsgroup.com/updates/office-365-tenant-to-tenant-migration-step-by-step-complete-guide/
AI in the Workplace Reskilling, Upskilling, and Future Work.pptxSunil Jagani
Discover how AI is transforming the workplace and learn strategies for reskilling and upskilling employees to stay ahead. This comprehensive guide covers the impact of AI on jobs, essential skills for the future, and successful case studies from industry leaders. Embrace AI-driven changes, foster continuous learning, and build a future-ready workforce.
Read More - https://bit.ly/3VKly70
Introducing BoxLang : A new JVM language for productivity and modularity!Ortus Solutions, Corp
Just like life, our code must adapt to the ever changing world we live in. From one day coding for the web, to the next for our tablets or APIs or for running serverless applications. Multi-runtime development is the future of coding, the future is to be dynamic. Let us introduce you to BoxLang.
Dynamic. Modular. Productive.
BoxLang redefines development with its dynamic nature, empowering developers to craft expressive and functional code effortlessly. Its modular architecture prioritizes flexibility, allowing for seamless integration into existing ecosystems.
Interoperability at its Core
With 100% interoperability with Java, BoxLang seamlessly bridges the gap between traditional and modern development paradigms, unlocking new possibilities for innovation and collaboration.
Multi-Runtime
From the tiny 2m operating system binary to running on our pure Java web server, CommandBox, Jakarta EE, AWS Lambda, Microsoft Functions, Web Assembly, Android and more. BoxLang has been designed to enhance and adapt according to it's runnable runtime.
The Fusion of Modernity and Tradition
Experience the fusion of modern features inspired by CFML, Node, Ruby, Kotlin, Java, and Clojure, combined with the familiarity of Java bytecode compilation, making BoxLang a language of choice for forward-thinking developers.
Empowering Transition with Transpiler Support
Transitioning from CFML to BoxLang is seamless with our JIT transpiler, facilitating smooth migration and preserving existing code investments.
Unlocking Creativity with IDE Tools
Unleash your creativity with powerful IDE tools tailored for BoxLang, providing an intuitive development experience and streamlining your workflow. Join us as we embark on a journey to redefine JVM development. Welcome to the era of BoxLang.
"NATO Hackathon Winner: AI-Powered Drug Search", Taras KlobaFwdays
This is a session that details how PostgreSQL's features and Azure AI Services can be effectively used to significantly enhance the search functionality in any application.
In this session, we'll share insights on how we used PostgreSQL to facilitate precise searches across multiple fields in our mobile application. The techniques include using LIKE and ILIKE operators and integrating a trigram-based search to handle potential misspellings, thereby increasing the search accuracy.
We'll also discuss how the azure_ai extension on PostgreSQL databases in Azure and Azure AI Services were utilized to create vectors from user input, a feature beneficial when users wish to find specific items based on text prompts. While our application's case study involves a drug search, the techniques and principles shared in this session can be adapted to improve search functionality in a wide range of applications. Join us to learn how PostgreSQL and Azure AI can be harnessed to enhance your application's search capability.
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
inQuba Webinar Mastering Customer Journey Management with Dr Graham HillLizaNolte
HERE IS YOUR WEBINAR CONTENT! 'Mastering Customer Journey Management with Dr. Graham Hill'. We hope you find the webinar recording both insightful and enjoyable.
In this webinar, we explored essential aspects of Customer Journey Management and personalization. Here’s a summary of the key insights and topics discussed:
Key Takeaways:
Understanding the Customer Journey: Dr. Hill emphasized the importance of mapping and understanding the complete customer journey to identify touchpoints and opportunities for improvement.
Personalization Strategies: We discussed how to leverage data and insights to create personalized experiences that resonate with customers.
Technology Integration: Insights were shared on how inQuba’s advanced technology can streamline customer interactions and drive operational efficiency.
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation F...AlexanderRichford
QR Secure: A Hybrid Approach Using Machine Learning and Security Validation Functions to Prevent Interaction with Malicious QR Codes.
Aim of the Study: The goal of this research was to develop a robust hybrid approach for identifying malicious and insecure URLs derived from QR codes, ensuring safe interactions.
This is achieved through:
Machine Learning Model: Predicts the likelihood of a URL being malicious.
Security Validation Functions: Ensures the derived URL has a valid certificate and proper URL format.
This innovative blend of technology aims to enhance cybersecurity measures and protect users from potential threats hidden within QR codes 🖥 🔒
This study was my first introduction to using ML which has shown me the immense potential of ML in creating more secure digital environments!
Lee Barnes - Path to Becoming an Effective Test Automation Engineer.pdfleebarnesutopia
So… you want to become a Test Automation Engineer (or hire and develop one)? While there’s quite a bit of information available about important technical and tool skills to master, there’s not enough discussion around the path to becoming an effective Test Automation Engineer that knows how to add VALUE. In my experience this had led to a proliferation of engineers who are proficient with tools and building frameworks but have skill and knowledge gaps, especially in software testing, that reduce the value they deliver with test automation.
In this talk, Lee will share his lessons learned from over 30 years of working with, and mentoring, hundreds of Test Automation Engineers. Whether you’re looking to get started in test automation or just want to improve your trade, this talk will give you a solid foundation and roadmap for ensuring your test automation efforts continuously add value. This talk is equally valuable for both aspiring Test Automation Engineers and those managing them! All attendees will take away a set of key foundational knowledge and a high-level learning path for leveling up test automation skills and ensuring they add value to their organizations.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
What is an RPA CoE? Session 1 – CoE VisionDianaGray10
In the first session, we will review the organization's vision and how this has an impact on the COE Structure.
Topics covered:
• The role of a steering committee
• How do the organization’s priorities determine CoE Structure?
Speaker:
Chris Bolin, Senior Intelligent Automation Architect Anika Systems
QA or the Highway - Component Testing: Bridging the gap between frontend appl...zjhamm304
These are the slides for the presentation, "Component Testing: Bridging the gap between frontend applications" that was presented at QA or the Highway 2024 in Columbus, OH by Zachary Hamm.
In the realm of cybersecurity, offensive security practices act as a critical shield. By simulating real-world attacks in a controlled environment, these techniques expose vulnerabilities before malicious actors can exploit them. This proactive approach allows manufacturers to identify and fix weaknesses, significantly enhancing system security.
This presentation delves into the development of a system designed to mimic Galileo's Open Service signal using software-defined radio (SDR) technology. We'll begin with a foundational overview of both Global Navigation Satellite Systems (GNSS) and the intricacies of digital signal processing.
The presentation culminates in a live demonstration. We'll showcase the manipulation of Galileo's Open Service pilot signal, simulating an attack on various software and hardware systems. This practical demonstration serves to highlight the potential consequences of unaddressed vulnerabilities, emphasizing the importance of offensive security practices in safeguarding critical infrastructure.
From Natural Language to Structured Solr Queries using LLMsSease
This talk draws on experimentation to enable AI applications with Solr. One important use case is to use AI for better accessibility and discoverability of the data: while User eXperience techniques, lexical search improvements, and data harmonization can take organizations to a good level of accessibility, a structural (or “cognitive” gap) remains between the data user needs and the data producer constraints.
That is where AI – and most importantly, Natural Language Processing and Large Language Model techniques – could make a difference. This natural language, conversational engine could facilitate access and usage of the data leveraging the semantics of any data source.
The objective of the presentation is to propose a technical approach and a way forward to achieve this goal.
The key concept is to enable users to express their search queries in natural language, which the LLM then enriches, interprets, and translates into structured queries based on the Solr index’s metadata.
This approach leverages the LLM’s ability to understand the nuances of natural language and the structure of documents within Apache Solr.
The LLM acts as an intermediary agent, offering a transparent experience to users automatically and potentially uncovering relevant documents that conventional search methods might overlook. The presentation will include the results of this experimental work, lessons learned, best practices, and the scope of future work that should improve the approach and make it production-ready.
Essentials of Automations: Exploring Attributes & Automation ParametersSafe Software
Building automations in FME Flow can save time, money, and help businesses scale by eliminating data silos and providing data to stakeholders in real-time. One essential component to orchestrating complex automations is the use of attributes & automation parameters (both formerly known as “keys”). In fact, it’s unlikely you’ll ever build an Automation without using these components, but what exactly are they?
Attributes & automation parameters enable the automation author to pass data values from one automation component to the next. During this webinar, our FME Flow Specialists will cover leveraging the three types of these output attributes & parameters in FME Flow: Event, Custom, and Automation. As a bonus, they’ll also be making use of the Split-Merge Block functionality.
You’ll leave this webinar with a better understanding of how to maximize the potential of automations by making use of attributes & automation parameters, with the ultimate goal of setting your enterprise integration workflows up on autopilot.
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...DanBrown980551
This LF Energy webinar took place June 20, 2024. It featured:
-Alex Thornton, LF Energy
-Hallie Cramer, Google
-Daniel Roesler, UtilityAPI
-Henry Richardson, WattTime
In response to the urgency and scale required to effectively address climate change, open source solutions offer significant potential for driving innovation and progress. Currently, there is a growing demand for standardization and interoperability in energy data and modeling. Open source standards and specifications within the energy sector can also alleviate challenges associated with data fragmentation, transparency, and accessibility. At the same time, it is crucial to consider privacy and security concerns throughout the development of open source platforms.
This webinar will delve into the motivations behind establishing LF Energy’s Carbon Data Specification Consortium. It will provide an overview of the draft specifications and the ongoing progress made by the respective working groups.
Three primary specifications will be discussed:
-Discovery and client registration, emphasizing transparent processes and secure and private access
-Customer data, centering around customer tariffs, bills, energy usage, and full consumption disclosure
-Power systems data, focusing on grid data, inclusive of transmission and distribution networks, generation, intergrid power flows, and market settlement data
LF Energy Webinar: Carbon Data Specifications: Mechanisms to Improve Data Acc...
Lecture v4
1. A Model for the Systems
Architecture of the Future
Prof. Paul A. Strassmann
George Mason University, December 5, 2005
1
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
2. Data-Centric Era; IBM Dominates
Hundred Sources
1950-1980
Months⇒Weeks
2
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
3. Workgroup-Centric Era; Microsoft, INTEL Dominate
Million Sources
Hundred Sources
1950-1980 1980-2010
Weeks⇒Days
Months⇒Weeks
3
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
4. Network-Centric Era; Google and Cisco?
+Multi-Media
+Text Billions Sources
Data
Million Sources
Hundred Sources
1950-1980 1980-2010 2010-
Days⇒Real-Time
Weeks⇒Days
Months⇒Weeks
4
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
5. Example of a Network-Centric System
5
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
7. Performance (2005)
• Infrastructure = > 50% of spending;
• Security = ?;
• Integration = > 50% of applications;
• Network downtime = > 1 hour/year;
• Innovation = > 1 year.
7
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
8. Conclusion
• Network-Centric systems cannot be
built on Workgroup-Centric architecture.
8
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
9. Network-Centric Principles (Google)
1. Build & operate protected information
network;
2. Offer universal connectivity for:
– Collection, processing and storing of
information;
– Provide secured communications.
3. Maintain shared data models;
4. Require continued upgrading & innovation.
9
11. Standard Google Clusters Operate Net
• 359 racks
• 31,654 machines
• 63,184 CPUs
• 126,368 Ghz of processing power
• 63,184 Gb of RAM
• 2,527 Tb of Hard Drive space
• Appx. 40 million searches/day
11
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
12. Clusters Have Identical Architecture
Index
Servers
Web Web
Switch Servers
Document
Servers
12
14. Google Infrastructure: Key to Growth (1)
• >200,000 custom-built commodity servers;
• Acting as one parallel supercomputer;
• Fault tolerant hardware.
• Storage capacity >5 petabytes; low response latency
(0.2 sec); >80GB per server, distributed;
• Indexed >8 billion web pages; Indexing is
computationally complex (>500M * > 2B matrix)
• Capital and operating costs at fraction of large scale
commercial servers; traffic growth 20-30%/month; data
centers (>12); in US, Europe and Asia.
14
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
15. Google Infrastructure: Key to Growth (2)
• >200,000 commodity Linux servers built to custom
specifications; distributed cluster architecture; acting as one
parallel supercomputer; scaleable;
• >50,000 requests/sec; fault tolerant (no single point of failure);
diverse hardware; stripped version of Red Hat;
• Storage capacity >5 petabytes;
• >80GB per server;
• Indexed >8 billion web pages;
• Indexing is complex (500M x 2B matrix)
• Capital and operating costs at fraction of large scale commercial
servers; traffic growth 20-30%/month; data centers (>12); in US,
Europe and Asia.
15
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
16. Google Infrastructure: Key to Growth (3)
• >200,000 commodity Linux servers built to custom
specifications; distributed cluster architecture; acting as one
parallel supercomputer; scaleable;
• >50,000 requests/sec; fault tolerant (no single point of failure);
diverse hardware; stripped version of Red Hat;
• Storage capacity >5 petabytes; low response latency (0.2 sec);
>80GB per server, distributed;
• Indexed >8 billion web pages; Indexing is computationally
complex (>500M * > 2B matrix)
• Capital and operating costs a fraction of commercial servers;
• Traffic growth 20-30%/month;
• Data centers (>20), in US, Europe and Asia.
16
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
17. Architecture for Reliability
• Replication (3x+) for redundancy;
• Replication for proximity and response;
• Reliability with software and architecture,
not with hardware.
17
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
18. Indexing for Response
• Dynamic indexing of 8B+ pages;
• Dynamic indexing of 1B+ images;
• Indexing of 1B+ messages;
• Index broken into “shards” and
distributed across data centers.
18
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
19. Query Serving Infrastructure
• Processing a single query may involve
1000+ servers;
• Index Servers access Index Shards;
• Document Servers access Doc Shards;
• Response times monitored to assure
<0.25 sec latency.
19
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
20. Google MapReduce System (1)
• Coordinates servers in real-time;
• Automates distribution of workload;
• Fault tolerance and service reconstitution;
• Systems-wide I/O cluster scheduling;
• Status and performance monitoring.
20
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
21. Google MapReduce System (2)
• Coordination of servers in real-time;
• Automates distribution of workload;
• Fault tolerance & service reconstitution;
• Systems-wide cluster scheduling;
• Status and performance monitoring.
21
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
23. Multi-Lingual Services
23
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
24. Search in Arabic Media
24
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
25. Video Searches
25
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
26. Google Base - Connecting Diverse Sources
Locate events within
45 miles of New York in
November, 2005
26
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
27. Semantic Parsing
• Tools parse millions of documents;
• Automated learning for related information.
– Query: “Bay Area Cooking Classes”
– Finds: “San Francisco College Classes”;
“The Magic of Thai Cuisine”
27
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
29. Data Engineering
• Standard file management: The Google
File System (GFS);
• Standard job scheduling: The Global Work
Queue (GWQ);
• Standard network management: The
Google MapReduce system.
29
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
30. Google File System (GFS)
• Replicated Masters manage MetaData
directories;
• Data transfers directly at the machine
level within 2,000+ clusters;
• File broken into 64 MB chunks for
2000+ MB/second read/write load;
• All file chunks at least triplicate for
safety.
30
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
31. Data Dictionary for Interoperability
31
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
32. Application Interfaces
32
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
43. New Internet: Billions of Browsers,
Secure Shared Applications+Data
Browsers
Application
Application
Application 43
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
44. Workgroup vs. Network Architectures (1)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
Operating system dependency Open source browsers
License Software Pay for Use
44
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
45. Workgroup vs. Network Architectures (2)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
Operating system dependency Open source browsers
License Software Pay for Use
45
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
46. Workgroup vs. Network Architectures (3)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
Operating system dependency Open source browsers
License Software Pay for Use
46
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
47. Workgroup vs. Network Architectures (4)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
Operating system dependency Open source browsers
License Software Pay for Use
47
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
48. Workgroup vs. Network Architectures (5)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
Operating system dependent Open source browsers
License Software Pay for Use
48
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
49. Workgroup vs. Network Architectures (6)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
License Software Pay for Use
49
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
50. Workgroup vs. Network Architectures (7)
Workgroup Centric Network Centric
Strategy: Capture Desktop Strategy: Occupy Internet
Customer’s labor and capital Labor and capital in network
User-specific infrastructures Infrastructure is universal
Systems controls by user Network controls in network
License Software Pay for Use
Data read from files Data assembled in context
50
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
51. The Future
Technology
• All electronic devices on Internet.
• Data, voice, video, sensor inputs accessible.
• Phone, TV and print media displaced.
Services
• Systems respond to questions.
• Information is displayed in context.
• Applications for decision-making.
51
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY
52. Relevance for National Security Systems
• Workgroups to Network-Centric services.
• Migrate through displacement.
• Invest savings in innovation.
52
Prof. Strassmann, GMU Lecture, 12/05/05 - REPRODUCED BY PERMISSION ONLY