The role of databases in modern application developmentMariaDB plc
The rise of serverless microservices, event-driven application architecture and full-stack development with JavaScript and the MEAN stack is changing what application developers need from databases – and how they interact with them. In this session, MariaDB's Thomas Boyd discusses recent advancements in application development and architecture and explains how MariaDB supports them.
In this day and age, data grows so fast it’s not uncommon for those of us using a relational database to reach the limits of its capacity. In this session, Kwangbock Lee explains how Samsung uses ClustrixDB to handle fast-growing data without manual database sharding. He highlights lessons learned, including a few hiccups along the way, and shares Samsung's experience migrating to ClustrixDB.
This presentation is dedicated to Microsoft Azure. It contains an overview of the main trends in the development of Microsoft Azure, and the solutions that Microsoft offers with this product. There are also insights on its further development, and an impact it is going to bring to the market.
This presentation by Andriy Gnennyy (Senior Consultant, GlobalLogic Kharkiv) was delivered at GlobalLogic Kharkiv MS TechTalk on June 13, 2017.
Initial presentation of openstack (for montreal user group)Marcos García
Introduction to Openstack: basic concepts, latest Havana project release, cloud terminology (including IaaS, PaaS and SaaS). This presentation was shown in the first Openstack Montreal user group in November 19 2013 (http://montrealopenstack.org/)
FOSSASIA 2016 - 7 Tips to design web centric high-performance applicationsAshnikbiz
Ashnik Database Solution Architect, Sameer Kumar, an Open Source evangelist shared some tips at FOSSASIA 2016 about how to design web-centric high-performance applications.
How to power microservices with MariaDBMariaDB plc
Adoption of microservices is continuing at a rapid pace, but many deployments struggle when it comes to the database topology and data modeling. This session covers the pros and cons of different approaches (e.g., giving every microservice its own database or its own schema on a shared database) and various strategies for providing a consolidated view of data when different data is managed by different microservices.
Getting started in the cloud for developersMariaDB plc
Looking to get up and running in the cloud, and start building applications with MariaDB as fast as possible? In this session, Thomas Boyd walks through the quick-start process of deploying MariaDB in the most popular public clouds. He then touches on some of the essential differences between cloud database services, helping you to create the cloud database strategy that best meets your needs.
WSO2Con ASIA 2016: WSO2 Analytics Platform: The One Stop Shop for All Your Da...WSO2
Today’s highly connected world is flooding businesses with big and fast-moving data. The ability to trawl this data ocean and identify actionable insights can deliver a competitive advantage to any organization. The WSO2 Analytics Platform enables businesses to do just that by providing batch, real-time, interactive and predictive analysis capabilities all in one place.
In this tutorial we will
Plug in the WSO2 Analytics Platform to some common business use cases
Showcase the numerous capabilities of the platform
Demonstrate how to collect data, analyze, predict and communicate effectively
The role of databases in modern application developmentMariaDB plc
The rise of serverless microservices, event-driven application architecture and full-stack development with JavaScript and the MEAN stack is changing what application developers need from databases – and how they interact with them. In this session, MariaDB's Thomas Boyd discusses recent advancements in application development and architecture and explains how MariaDB supports them.
In this day and age, data grows so fast it’s not uncommon for those of us using a relational database to reach the limits of its capacity. In this session, Kwangbock Lee explains how Samsung uses ClustrixDB to handle fast-growing data without manual database sharding. He highlights lessons learned, including a few hiccups along the way, and shares Samsung's experience migrating to ClustrixDB.
This presentation is dedicated to Microsoft Azure. It contains an overview of the main trends in the development of Microsoft Azure, and the solutions that Microsoft offers with this product. There are also insights on its further development, and an impact it is going to bring to the market.
This presentation by Andriy Gnennyy (Senior Consultant, GlobalLogic Kharkiv) was delivered at GlobalLogic Kharkiv MS TechTalk on June 13, 2017.
Initial presentation of openstack (for montreal user group)Marcos García
Introduction to Openstack: basic concepts, latest Havana project release, cloud terminology (including IaaS, PaaS and SaaS). This presentation was shown in the first Openstack Montreal user group in November 19 2013 (http://montrealopenstack.org/)
FOSSASIA 2016 - 7 Tips to design web centric high-performance applicationsAshnikbiz
Ashnik Database Solution Architect, Sameer Kumar, an Open Source evangelist shared some tips at FOSSASIA 2016 about how to design web-centric high-performance applications.
How to power microservices with MariaDBMariaDB plc
Adoption of microservices is continuing at a rapid pace, but many deployments struggle when it comes to the database topology and data modeling. This session covers the pros and cons of different approaches (e.g., giving every microservice its own database or its own schema on a shared database) and various strategies for providing a consolidated view of data when different data is managed by different microservices.
Getting started in the cloud for developersMariaDB plc
Looking to get up and running in the cloud, and start building applications with MariaDB as fast as possible? In this session, Thomas Boyd walks through the quick-start process of deploying MariaDB in the most popular public clouds. He then touches on some of the essential differences between cloud database services, helping you to create the cloud database strategy that best meets your needs.
WSO2Con ASIA 2016: WSO2 Analytics Platform: The One Stop Shop for All Your Da...WSO2
Today’s highly connected world is flooding businesses with big and fast-moving data. The ability to trawl this data ocean and identify actionable insights can deliver a competitive advantage to any organization. The WSO2 Analytics Platform enables businesses to do just that by providing batch, real-time, interactive and predictive analysis capabilities all in one place.
In this tutorial we will
Plug in the WSO2 Analytics Platform to some common business use cases
Showcase the numerous capabilities of the platform
Demonstrate how to collect data, analyze, predict and communicate effectively
Membase Intro from Membase Meetup San FranciscoMembase
Membase is a distributed database that is simple, fast, and elastic. It can be deployed across commodity servers with zero downtime. Membase uses the memcached protocol and supports many programming languages out of the box. It provides high performance through techniques like asynchronous operations and write deduplication. Membase also offers elastic scaling through dynamic rebalancing and cloning of nodes. Future plans for Membase include advanced features like indexing and connectors to other systems through a new NodeCode extension framework.
This document discusses using Ansible for infrastructure automation. It provides examples of how Ansible can be used for provisioning infrastructure, configuring servers, patching, backups, cluster deployment, and scaling. It also gives three use cases: creating a platform for a client, integrating Ansible with other tools like vRA and CyberArk, and automating a two year project involving RedHat and Windows systems. It concludes by discussing common problems providing "DevOps as a service" and introducing Crevise PowerOps to address these.
Infinispan, transactional key value data grid and nosql databaseAlexander Petrov
The document discusses key topics related to distributed caching including cache technologies, consistency models, performance considerations, and challenges in introducing distributed caching to existing systems. It provides examples of how reference data and transactional data differ in maximum reads and writes per second. The document also covers cache eviction policies, transactions, and mixing technology stacks.
A presentation on how Showyou uses the Riak datastore at Showyou.com, as well as work we've been doing on a custom Riak backend for search and analytics.
Infinispan - Galder Zamarreno - October 2010JUG Lausanne
Galder Zamarreno gave a presentation on Infinispan, an open source data grid platform designed for cloud computing. He discussed how traditional databases do not work well in cloud environments due to their stateful and failure-prone nature. Data grids are better suited as they are highly scalable, have no single point of failure, and work with ephemeral cloud nodes. Infinispan is a new data grid that improves on an earlier product, JBoss Cache, with a more scalable architecture and features like a simple map API, client/server support, and integration with Hibernate and Lucene. Future plans for Infinispan include enhanced replication, distributed execution capabilities, and support for cloud-based data
This document summarizes a presentation about using the ELK stack to process logs at scale. It discusses using Logstash for log ingestion and filtering, Elasticsearch for indexing and searching logs, and Kibana for visualizing logs. The document provides details on using Logstash forwarders to ship logs to Logstash from application containers, scaling Logstash and Elasticsearch horizontally, hardware recommendations for Elasticsearch, and configuration techniques for optimizing Elasticsearch performance and reliability.
This document provides an overview and demonstration of Membase, a distributed database. It discusses how Membase can be deployed in five minutes or less on a single node or cluster. It is simple to develop applications using Membase's key-value interface without requiring a schema. The document also summarizes several use cases for Membase in domains like social games, advertising, and search/gaming portals. It provides a high-level overview of Membase's architecture, including its clustering functionality and TAP interface. Finally, it briefly previews Membase's NodeCode feature for extending the database with custom modules.
This document provides an overview of Membase, including:
- Membase is a distributed database that allows applications and data to scale independently. It uses the Memcached protocol and architecture.
- Membase can be deployed in various ways, including using the built-in Memcached caching layer or standalone proxies. It also supports secure multitenant buckets.
- The document demonstrates Membase's use cases with examples from large companies and discusses its architecture, including clustering, data access protocols, and a future NodeCode capability.
Solr cloud the 'search first' nosql database extended deep divelucenerevolution
Presented by Mark Miller, Software Engineer, Cloudera
As the NoSQL ecosystem looks to integrate great search, great search is naturally beginning to expose many NoSQL features. Will these Goliath's collide? Or will they remain specialized while intermingling – two sides of the same coin.
Come learn about where SolrCloud fits into the NoSQL landscape. What can it do? What will it do? And how will the big data, NoSQL, Search ecosystem evolve. If you are interested in Big Data, NoSQL, distributed systems, CAP theorem and other hype filled terms, than this talk may be for you.
The document provides an overview of NoSQL databases and focuses on MongoDB and Riak. It discusses how these databases address the needs of web applications by providing flexibility, scalability, and high performance. Riak is highlighted as using a distributed architecture with no single point of failure and tunable consistency properties. Its ability to link documents and handle high availability through replication is also summarized.
LivePerson uses CouchBase for real-time analytics of visitor data to provide visibility to customers on their online visitors. Previously, visitor state was stored in memory on stateful web servers, limiting scalability. CouchBase was chosen for its performance, resilience, linear scalability, schema flexibility, and ability to handle LivePerson's high throughput of over 1 million concurrent visitors and 100k operations per second. It is used to store visitor documents containing events and is queried to return relevant visitors to agents. Cross data center replication is also used to improve resilience. LivePerson has found CouchBase easy to develop on and has expanded its use to additional cases like session state and caching.
Kafka and Kafka Streams in the Global Schibsted Data PlatformFredrik Vraalsen
This document discusses Schibsted's use of Kafka and Kafka Streams in its global data platform. It describes how Schibsted processes streaming data using Kafka Streams to build a new streaming pipeline called Yggdrasil that provides low latency and delivery guarantees. It also discusses how Kafka Connect is used to get data into and out of the platform and how third party analytics tools can connect. The document outlines some challenges experienced in scaling up the platform.
Billions of Messages in Real Time: Why Paypal & LinkedIn Trust an Engagement ...confluent
(Bruno Simic, Solutions Engineer, Couchbase)
Breakout during Confluent’s streaming event in Munich. This three-day hands-on course focused on how to build, manage, and monitor clusters using industry best-practices developed by the world’s foremost Apache Kafka™ experts. The sessions focused on how Kafka and the Confluent Platform work, how their main subsystems interact, and how to set up, manage, monitor, and tune your cluster.
The document provides an overview of MongoDB administration including its data model, replication for high availability, sharding for scalability, deployment architectures, operations, security features, and resources for operations teams. The key topics covered are the flexible document data model, replication using replica sets for high availability, scaling out through sharding of data across multiple servers, and different deployment architectures including single/multi data center configurations.
- The company was founded in 2003 and provides sophisticated visualization and interpretation of genetic data through targeted analysis workflows and actionable results.
- The document discusses the architecture for a genome browser application, including a revised architecture with a VPC across two availability zones with private subnets for security.
- It describes the web stack behind a load balancer with session info stored in Elasticache and autoscaling from 2-6 machines depending on load. The database is a statically scaled MongoDB cluster that cannot use RDS due to local deployment requirements.
This document discusses Netflix's use of near real-time recommendations using Spark streaming. It provides examples of use cases like video insights and billboard recommendations that require processing data with low latency. The infrastructure for handling terabytes of daily data across regions at Netflix's scale is also described, along with challenges of scaling streaming workloads and ensuring reliability.
This document discusses ideas and technologies for building scalable software systems and processing big data. It covers:
1. Bi-modal distribution of developers shapes architecture/design and the need for loosely/tightly coupled code.
2. Internet companies like Google and Facebook innovate at large scale using open source tools and REST architectures.
3. A REST architecture allows scalability, extensible development, and integration of tools/ideas from the internet for non-internet applications.
This document provides an overview of Google Cloud Platform (GCP) services. It begins by explaining why GCP is underpinned by Google's infrastructure and innovation. It then outlines GCP's compute, networking, storage, big data, and machine learning services. These include Compute Engine, Container Engine, App Engine, load balancing, Cloud DNS, Cloud Storage, Cloud Datastore, Cloud Bigtable, Cloud SQL, BigQuery, Dataflow, Pub/Sub, Dataproc, and Cloud Datalab. Machine learning services such as Translate API, Prediction API, Cloud Vision API, and Cloud Speech API are also introduced.
Membase Intro from Membase Meetup San FranciscoMembase
Membase is a distributed database that is simple, fast, and elastic. It can be deployed across commodity servers with zero downtime. Membase uses the memcached protocol and supports many programming languages out of the box. It provides high performance through techniques like asynchronous operations and write deduplication. Membase also offers elastic scaling through dynamic rebalancing and cloning of nodes. Future plans for Membase include advanced features like indexing and connectors to other systems through a new NodeCode extension framework.
This document discusses using Ansible for infrastructure automation. It provides examples of how Ansible can be used for provisioning infrastructure, configuring servers, patching, backups, cluster deployment, and scaling. It also gives three use cases: creating a platform for a client, integrating Ansible with other tools like vRA and CyberArk, and automating a two year project involving RedHat and Windows systems. It concludes by discussing common problems providing "DevOps as a service" and introducing Crevise PowerOps to address these.
Infinispan, transactional key value data grid and nosql databaseAlexander Petrov
The document discusses key topics related to distributed caching including cache technologies, consistency models, performance considerations, and challenges in introducing distributed caching to existing systems. It provides examples of how reference data and transactional data differ in maximum reads and writes per second. The document also covers cache eviction policies, transactions, and mixing technology stacks.
A presentation on how Showyou uses the Riak datastore at Showyou.com, as well as work we've been doing on a custom Riak backend for search and analytics.
Infinispan - Galder Zamarreno - October 2010JUG Lausanne
Galder Zamarreno gave a presentation on Infinispan, an open source data grid platform designed for cloud computing. He discussed how traditional databases do not work well in cloud environments due to their stateful and failure-prone nature. Data grids are better suited as they are highly scalable, have no single point of failure, and work with ephemeral cloud nodes. Infinispan is a new data grid that improves on an earlier product, JBoss Cache, with a more scalable architecture and features like a simple map API, client/server support, and integration with Hibernate and Lucene. Future plans for Infinispan include enhanced replication, distributed execution capabilities, and support for cloud-based data
This document summarizes a presentation about using the ELK stack to process logs at scale. It discusses using Logstash for log ingestion and filtering, Elasticsearch for indexing and searching logs, and Kibana for visualizing logs. The document provides details on using Logstash forwarders to ship logs to Logstash from application containers, scaling Logstash and Elasticsearch horizontally, hardware recommendations for Elasticsearch, and configuration techniques for optimizing Elasticsearch performance and reliability.
This document provides an overview and demonstration of Membase, a distributed database. It discusses how Membase can be deployed in five minutes or less on a single node or cluster. It is simple to develop applications using Membase's key-value interface without requiring a schema. The document also summarizes several use cases for Membase in domains like social games, advertising, and search/gaming portals. It provides a high-level overview of Membase's architecture, including its clustering functionality and TAP interface. Finally, it briefly previews Membase's NodeCode feature for extending the database with custom modules.
This document provides an overview of Membase, including:
- Membase is a distributed database that allows applications and data to scale independently. It uses the Memcached protocol and architecture.
- Membase can be deployed in various ways, including using the built-in Memcached caching layer or standalone proxies. It also supports secure multitenant buckets.
- The document demonstrates Membase's use cases with examples from large companies and discusses its architecture, including clustering, data access protocols, and a future NodeCode capability.
Solr cloud the 'search first' nosql database extended deep divelucenerevolution
Presented by Mark Miller, Software Engineer, Cloudera
As the NoSQL ecosystem looks to integrate great search, great search is naturally beginning to expose many NoSQL features. Will these Goliath's collide? Or will they remain specialized while intermingling – two sides of the same coin.
Come learn about where SolrCloud fits into the NoSQL landscape. What can it do? What will it do? And how will the big data, NoSQL, Search ecosystem evolve. If you are interested in Big Data, NoSQL, distributed systems, CAP theorem and other hype filled terms, than this talk may be for you.
The document provides an overview of NoSQL databases and focuses on MongoDB and Riak. It discusses how these databases address the needs of web applications by providing flexibility, scalability, and high performance. Riak is highlighted as using a distributed architecture with no single point of failure and tunable consistency properties. Its ability to link documents and handle high availability through replication is also summarized.
LivePerson uses CouchBase for real-time analytics of visitor data to provide visibility to customers on their online visitors. Previously, visitor state was stored in memory on stateful web servers, limiting scalability. CouchBase was chosen for its performance, resilience, linear scalability, schema flexibility, and ability to handle LivePerson's high throughput of over 1 million concurrent visitors and 100k operations per second. It is used to store visitor documents containing events and is queried to return relevant visitors to agents. Cross data center replication is also used to improve resilience. LivePerson has found CouchBase easy to develop on and has expanded its use to additional cases like session state and caching.
Kafka and Kafka Streams in the Global Schibsted Data PlatformFredrik Vraalsen
This document discusses Schibsted's use of Kafka and Kafka Streams in its global data platform. It describes how Schibsted processes streaming data using Kafka Streams to build a new streaming pipeline called Yggdrasil that provides low latency and delivery guarantees. It also discusses how Kafka Connect is used to get data into and out of the platform and how third party analytics tools can connect. The document outlines some challenges experienced in scaling up the platform.
Billions of Messages in Real Time: Why Paypal & LinkedIn Trust an Engagement ...confluent
(Bruno Simic, Solutions Engineer, Couchbase)
Breakout during Confluent’s streaming event in Munich. This three-day hands-on course focused on how to build, manage, and monitor clusters using industry best-practices developed by the world’s foremost Apache Kafka™ experts. The sessions focused on how Kafka and the Confluent Platform work, how their main subsystems interact, and how to set up, manage, monitor, and tune your cluster.
The document provides an overview of MongoDB administration including its data model, replication for high availability, sharding for scalability, deployment architectures, operations, security features, and resources for operations teams. The key topics covered are the flexible document data model, replication using replica sets for high availability, scaling out through sharding of data across multiple servers, and different deployment architectures including single/multi data center configurations.
- The company was founded in 2003 and provides sophisticated visualization and interpretation of genetic data through targeted analysis workflows and actionable results.
- The document discusses the architecture for a genome browser application, including a revised architecture with a VPC across two availability zones with private subnets for security.
- It describes the web stack behind a load balancer with session info stored in Elasticache and autoscaling from 2-6 machines depending on load. The database is a statically scaled MongoDB cluster that cannot use RDS due to local deployment requirements.
This document discusses Netflix's use of near real-time recommendations using Spark streaming. It provides examples of use cases like video insights and billboard recommendations that require processing data with low latency. The infrastructure for handling terabytes of daily data across regions at Netflix's scale is also described, along with challenges of scaling streaming workloads and ensuring reliability.
This document discusses ideas and technologies for building scalable software systems and processing big data. It covers:
1. Bi-modal distribution of developers shapes architecture/design and the need for loosely/tightly coupled code.
2. Internet companies like Google and Facebook innovate at large scale using open source tools and REST architectures.
3. A REST architecture allows scalability, extensible development, and integration of tools/ideas from the internet for non-internet applications.
This document provides an overview of Google Cloud Platform (GCP) services. It begins by explaining why GCP is underpinned by Google's infrastructure and innovation. It then outlines GCP's compute, networking, storage, big data, and machine learning services. These include Compute Engine, Container Engine, App Engine, load balancing, Cloud DNS, Cloud Storage, Cloud Datastore, Cloud Bigtable, Cloud SQL, BigQuery, Dataflow, Pub/Sub, Dataproc, and Cloud Datalab. Machine learning services such as Translate API, Prediction API, Cloud Vision API, and Cloud Speech API are also introduced.
The document discusses public cloud computing concepts including cloud infrastructure, services, and architectures. Some key points:
- Cloud provides on-demand access to computing resources like servers and storage over the internet. Major cloud providers include AWS, Azure, and Google Cloud.
- Cloud services include Infrastructure as a Service (IaaS), Platform as a Service (PaaS), Software as a Service (SaaS), and Function as a Service (FaaS).
- Microservices architecture breaks applications into small, independent services that communicate over the network. This allows independent scaling and improves resilience.
- Auto scaling helps automatically scale cloud resources like servers up and down based on demand to optimize costs and performance.
Pytheas is a web-based resource and UI framework for dashboards, web consoles, and exploring structured and unstructured data. It is based on open source frameworks like Guice, Jersey, FreeMarker, jQuery, and uses a modular design. Conformity Monkey helps keep cloud instances and clusters following best practices by using a mark and notify approach with customizable rules and rule sets. Zuul is Netflix's edge tier service that acts on HTTP requests using dynamic filters written in Groovy. Genie provides an abstraction of physical Hadoop clusters and a simple API to run jobs on them. Lipstick provides a visualization of Pig workflows. ICE is a tool for analyzing AWS usage data by tagging billing files and providing a
Big data conference europe real-time streaming in any and all clouds, hybri...Timothy Spann
Biography
Tim Spann is a Principal DataFlow Field Engineer at Cloudera where he works with Apache NiFi, MiniFi, Pulsar, Apache Flink, Apache MXNet, TensorFlow, Apache Spark, big data, the IoT, machine learning, and deep learning. Tim has over a decade of experience with the IoT, big data, distributed computing, streaming technologies, and Java programming. Previously, he was a senior solutions architect at AirisData and a senior field engineer at Pivotal. He blogs for DZone, where he is the Big Data Zone leader, and runs a popular meetup in Princeton on big data, the IoT, deep learning, streaming, NiFi, the blockchain, and Spark. Tim is a frequent speaker at conferences such as IoT Fusion, Strata, ApacheCon, Data Works Summit Berlin, DataWorks Summit Sydney, and Oracle Code NYC. He holds a BS and MS in computer science.
Talk
Real-Time Streaming in Any and All Clouds, Hybrid and Beyond
Today, data is being generated from devices and containers living at the edge of networks, clouds and data centers. We need to run business logic, analytics and deep learning at the scale and as events arrive.
Tools:
Apache Flink, Apache Pulsar, Apache NiFi, MiNiFi, DJL.ai Apache MXNet.
References:
https://www.datainmotion.dev/2019/11/introducing-mm-flank-apache-flink-stack.html
https://www.datainmotion.dev/2019/08/rapid-iot-development-with-cloudera.html
https://www.datainmotion.dev/2019/09/powering-edge-ai-for-sensor-reading.html
https://www.datainmotion.dev/2019/05/dataworks-summit-dc-2019-report.html
https://www.datainmotion.dev/2019/03/using-raspberry-pi-3b-with-apache-nifi.html
Source Code: https://github.com/tspannhw/MmFLaNK
FLiP Stack
StreamNative
"NoSQL on the move" by Glynn Bird
Mobile-first app web development is a solved problem, but how can you websites and apps the continue to work with little or internet connectivity? Discover how Offline-first development allows apps to present an "always on" experience for their user
HPC and cloud distributed computing, as a journeyPeter Clapham
Introducing an internal cloud brings new paradigms, tools and infrastructure management. When placed alongside traditional HPC the new opportunities are significant But getting to the new world with micro-services, autoscaling and autodialing is a journey that cannot be achieved in a single step.
USENIX LISA15: How TubeMogul Handles over One Trillion HTTP Requests a MonthNicolas Brousse
TubeMogul grew from few servers to over two thousands servers and handling over one trillion http requests a month, processed in less than 50ms each. To keep up with the fast growth, the SRE team had to implement an efficient Continuous Delivery infrastructure that allowed to do over 10,000 puppet deployment and 8,500 application deployment in 2014. In this presentation, we will cover the nuts and bolts of the TubeMogul operations engineering team and how they overcome challenges.
BISSA: Empowering Web gadget Communication with Tuple SpacesSrinath Perera
BISSA is a framework that enables communication between web gadgets using a tuple space model. It proposes a global, peer-to-peer based tuple space and an in-browser tuple space that are linked. The global tuple space is highly scalable and reliable, using a DHT for data distribution and indexing to support search queries. The in-browser space provides local and global APIs. Together this allows truly client-side web applications to communicate and store data without backend code. Performance tests on the global space showed good scalability and latency. Several use cases are proposed including coordinated dashboard gadgets, multiplayer games, and social applications.
Hello All,
Let's meet and discuss what are the new announcements from Build 2016 and how we can best leverage them in our business!
Here are some of the topics we will cover this time:
- Azure Functions
- Service Fabric
- Azure Storage
- Document DB
- Azure Container Services
- Power BI Embedded
- ASP.NET Core
- Virtual Machine Scale Sets
I will be happy to share my experience from the conference, especially the session I visited and also the conversations I had with various Microsoft representatives.
Azure is developing faster than ever and Microsoft is driving the platform in very interesting direction that require us to know and work with more and more new technologies!
Come and join us to learn more about Azure!
I am arranging the venue but my plan for the meetup is to be on April 25-th or April 27-th from 19:30. I will keep you updated on that!
Thank you!
Kanio
This presentation was first held at the OpenSQL Camp 2009, part of the FrOSCon conference in St. Augustin, Germany. It gives a nice overview over the project, technology and how it will progress. Find more information at http://www.blackray.org
Lessons Learned from Real-World Deployments of Java EE 7 at JavaOne 2014Arun Gupta
This document discusses lessons learned from real-world deployments of Java EE 7. Key points include increased developer productivity through features like batch processing, concurrency, simplified JMS, more annotated POJOs, and a cohesive integrated platform. Specific technologies used include JSON, WebSockets, Servlet 3.1 NIO, and REST. Real-world examples of implementations include an application for a UN agency to support refugees and a running social network application for runners.
This summary provides an overview of the key points from the document in 3 sentences:
The document outlines the agenda for Season 3 Episode 1 of the Netflix OSS podcast, which includes lightning talks on 8 new projects including Atlas, Prana, Raigad, Genie 2, Inviso, Dynomite, Nicobar, and MSL. Representatives from Netflix, IBM Watson, Nike Digital, and Pivotal then each provide a 3-5 minute presentation on their featured project. The presentations describe the motivation, features and benefits of each project for observability, integration with the Netflix ecosystem, automation of Elasticsearch deployments, job scheduling, dynamic scripting for Java, message security, and developing microservices
Microservices - opportunities, dilemmas and problemsŁukasz Sowa
Presentation from Warsjawa 2014 workshop "Microservices in Scala". Topics covered:
- What are microservices?
- What's the difference between them vs monolithic
architectures?
- What are the different flavours of microservices?
This document provides an overview of Postgresql, including its history, capabilities, advantages over other databases, best practices, and references for further learning. Postgresql is an open source relational database management system that has been in development for over 30 years. It offers rich SQL support, high performance, ACID transactions, and extensive extensibility through features like JSON, XML, and programming languages.
The document provides an overview and agenda for a presentation on the BlackRay database. It summarizes BlackRay's history and capabilities, positions it relative to other projects, and outlines the team and roadmap. The presentation covers BlackRay's architecture, APIs, management features, clustering support, and roadmap for future improvements.
Roko Kruze of vectorized.io describes real-time analytics using Redpanda event streams and ClickHouse data warehouse. 15 December 2021 SF Bay Area ClickHouse Meetup
Rami Sayar - Node microservices with DockerWeb à Québec
The document discusses converting a monolithic Node.js application into microservices and deploying them using Docker. It begins by defining microservices and their benefits. It then describes converting a sample pizza ordering application into independent microservices for handling messages, serving the frontend, and providing an API. Next, it covers patterns for networking microservices, including using an API gateway. It concludes by demonstrating how to deploy the microservices to Docker containers and use an orchestration tool like Kubernetes to manage them.
SpringPeople - Introduction to Cloud ComputingSpringPeople
Cloud computing is no longer a fad that is going around. It is for real and is perhaps the most talked about subject. Various players in the cloud eco-system have provided a definition that is closely aligned to their sweet spot –let it be infrastructure, platforms or applications.
This presentation will provide an exposure of a variety of cloud computing techniques, architecture, technology options to the participants and in general will familiarize cloud fundamentals in a holistic manner spanning all dimensions such as cost, operations, technology etc
Similar to Summer 2017 undergraduate research powerpoint (20)
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...shadow0702a
This document serves as a comprehensive step-by-step guide on how to effectively use PyCharm for remote debugging of the Windows Subsystem for Linux (WSL) on a local Windows machine. It meticulously outlines several critical steps in the process, starting with the crucial task of enabling permissions, followed by the installation and configuration of WSL.
The guide then proceeds to explain how to set up the SSH service within the WSL environment, an integral part of the process. Alongside this, it also provides detailed instructions on how to modify the inbound rules of the Windows firewall to facilitate the process, ensuring that there are no connectivity issues that could potentially hinder the debugging process.
The document further emphasizes on the importance of checking the connection between the Windows and WSL environments, providing instructions on how to ensure that the connection is optimal and ready for remote debugging.
It also offers an in-depth guide on how to configure the WSL interpreter and files within the PyCharm environment. This is essential for ensuring that the debugging process is set up correctly and that the program can be run effectively within the WSL terminal.
Additionally, the document provides guidance on how to set up breakpoints for debugging, a fundamental aspect of the debugging process which allows the developer to stop the execution of their code at certain points and inspect their program at those stages.
Finally, the document concludes by providing a link to a reference blog. This blog offers additional information and guidance on configuring the remote Python interpreter in PyCharm, providing the reader with a well-rounded understanding of the process.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELgerogepatton
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Understanding Inductive Bias in Machine LearningSUTEJAS
This presentation explores the concept of inductive bias in machine learning. It explains how algorithms come with built-in assumptions and preferences that guide the learning process. You'll learn about the different types of inductive bias and how they can impact the performance and generalizability of machine learning models.
The presentation also covers the positive and negative aspects of inductive bias, along with strategies for mitigating potential drawbacks. We'll explore examples of how bias manifests in algorithms like neural networks and decision trees.
By understanding inductive bias, you can gain valuable insights into how machine learning models work and make informed decisions when building and deploying them.
The CBC machine is a common diagnostic tool used by doctors to measure a patient's red blood cell count, white blood cell count and platelet count. The machine uses a small sample of the patient's blood, which is then placed into special tubes and analyzed. The results of the analysis are then displayed on a screen for the doctor to review. The CBC machine is an important tool for diagnosing various conditions, such as anemia, infection and leukemia. It can also help to monitor a patient's response to treatment.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
Redefining brain tumor segmentation: a cutting-edge convolutional neural netw...IJECEIAES
Medical image analysis has witnessed significant advancements with deep learning techniques. In the domain of brain tumor segmentation, the ability to
precisely delineate tumor boundaries from magnetic resonance imaging (MRI)
scans holds profound implications for diagnosis. This study presents an ensemble convolutional neural network (CNN) with transfer learning, integrating
the state-of-the-art Deeplabv3+ architecture with the ResNet18 backbone. The
model is rigorously trained and evaluated, exhibiting remarkable performance
metrics, including an impressive global accuracy of 99.286%, a high-class accuracy of 82.191%, a mean intersection over union (IoU) of 79.900%, a weighted
IoU of 98.620%, and a Boundary F1 (BF) score of 83.303%. Notably, a detailed comparative analysis with existing methods showcases the superiority of
our proposed model. These findings underscore the model’s competence in precise brain tumor localization, underscoring its potential to revolutionize medical
image analysis and enhance healthcare outcomes. This research paves the way
for future exploration and optimization of advanced CNN models in medical
imaging, emphasizing addressing false positives and resource efficiency.
UNLOCKING HEALTHCARE 4.0: NAVIGATING CRITICAL SUCCESS FACTORS FOR EFFECTIVE I...amsjournal
The Fourth Industrial Revolution is transforming industries, including healthcare, by integrating digital,
physical, and biological technologies. This study examines the integration of 4.0 technologies into
healthcare, identifying success factors and challenges through interviews with 70 stakeholders from 33
countries. Healthcare is evolving significantly, with varied objectives across nations aiming to improve
population health. The study explores stakeholders' perceptions on critical success factors, identifying
challenges such as insufficiently trained personnel, organizational silos, and structural barriers to data
exchange. Facilitators for integration include cost reduction initiatives and interoperability policies.
Technologies like IoT, Big Data, AI, Machine Learning, and robotics enhance diagnostics, treatment
precision, and real-time monitoring, reducing errors and optimizing resource utilization. Automation
improves employee satisfaction and patient care, while Blockchain and telemedicine drive cost reductions.
Successful integration requires skilled professionals and supportive policies, promising efficient resource
use, lower error rates, and accelerated processes, leading to optimized global healthcare outcomes.
Advanced control scheme of doubly fed induction generator for wind turbine us...IJECEIAES
This paper describes a speed control device for generating electrical energy on an electricity network based on the doubly fed induction generator (DFIG) used for wind power conversion systems. At first, a double-fed induction generator model was constructed. A control law is formulated to govern the flow of energy between the stator of a DFIG and the energy network using three types of controllers: proportional integral (PI), sliding mode controller (SMC) and second order sliding mode controller (SOSMC). Their different results in terms of power reference tracking, reaction to unexpected speed fluctuations, sensitivity to perturbations, and resilience against machine parameter alterations are compared. MATLAB/Simulink was used to conduct the simulations for the preceding study. Multiple simulations have shown very satisfying results, and the investigations demonstrate the efficacy and power-enhancing capabilities of the suggested control system.
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTjpsjournal1
The rivalry between prominent international actors for dominance over Central Asia's hydrocarbon
reserves and the ancient silk trade route, along with China's diplomatic endeavours in the area, has been
referred to as the "New Great Game." This research centres on the power struggle, considering
geopolitical, geostrategic, and geoeconomic variables. Topics including trade, political hegemony, oil
politics, and conventional and nontraditional security are all explored and explained by the researcher.
Using Mackinder's Heartland, Spykman Rimland, and Hegemonic Stability theories, examines China's role
in Central Asia. This study adheres to the empirical epistemological method and has taken care of
objectivity. This study analyze primary and secondary research documents critically to elaborate role of
china’s geo economic outreach in central Asian countries and its future prospect. China is thriving in trade,
pipeline politics, and winning states, according to this study, thanks to important instruments like the
Shanghai Cooperation Organisation and the Belt and Road Economic Initiative. According to this study,
China is seeing significant success in commerce, pipeline politics, and gaining influence on other
governments. This success may be attributed to the effective utilisation of key tools such as the Shanghai
Cooperation Organisation and the Belt and Road Economic Initiative.
3. Sensor Network
• Composed of
• Microcontroller units (MCU)
• Various sensors
• Temperature
• Humidity
• Photocells
• Web API
• Control MCUs
• Message Broker
• Route messages
• Manage communication channels
• Manage ques
4. The Ideal MCU
• Size
• Small form factor
• Power
• Low Power Consumption
• Battery Based
• Sleep Modes
• GPIO
• Digital
• Analog
• Communication
• Wifi 802.11 b/g/n
• Bluetooth
• Cellular 2G/3G
• Software
• Developed support
• Functionality Libraries
• Language Support
• C
• Javascript
5. Our Pick: ESP8266-12e
● Size
○ 24mm x16mm x 3mm
● Power
○ 3v3
○ Sleep mode
● Full IP Stack
○ WIFI 802.11 b/g/n
○ Can act as client or host
● Low Cost
○ $1.50
● Software
○ Several firmata available
○ Huge online community
https://acrobotic.com/media/wysiwyg/products/esp8266_esp12e_horizontal-01.png
6. Web API
• Web API
• Supervises MCUs
• MCUs initialization
• Connect to API Web Server
• Over Websockets
• Enter REPL
• Read Evaluate Print Loop (REPL)
• Inject Code directly into device
• Keep Status of Devices
• Modes
• Sleep
• On
• Off
7. Message Queue Telemetry Transport
(MQTT)
• MQTT: Lightweight Comm. Protocol
• Over TCP/IP
• Topic
• Data stream subject/identifier
• Publisher
• Transmits data on select topic
• Active connection
• Subscriber
• Listens for data on select topic
• Passive connection
• MQTT Broker
• Routes traffic based on subscriptions
8. Message Broker
• Intermediary
• Between sender and receiver
• Manages
• Delivery
• Routing
• Message Queue
• Protocol Conversion
• Message translation
Topic Subscriber Publisher
Temperature Server Temp. Sensor
Feedback Server Thermostat
mn … m4 m3 m2 m1 m0
m0 m1 m2 m3 m4 … mn
Temp. Sensor
Thermostat
Message Broker
Topic: Feedback
Topic: Temperature Server Side
Calculations
Message Queues
9. RabbitMQ
• Cross Language Support
• Java, Python, JavaScript, Ruby, & .NET
• Cross Protocol Support
• MQTT, AMQP, HTTP, & STOMP
• Asynchronous Messaging
• Many subs
• Many pubs
• QoS
• Data persists in queues until read by subscriber
• Low Latency
• Critical for real time apps.
10. Alternative Input Devices
• ROS Robots
• Drones
• Rovers
• Multi Input, Multi Output Systems
• Sensor + Actuator Pairs
12. What is Kura?
● OSGI based framework for IOT Gateways
● Runs in a JVM
● Built-in MQTT cloud services
● Browser based GUI
13. Key Features-Network and Cloud Services
● Extensive MQTT configuration options
● Helps implement more complex interaction flows beyond publish/subscribe
● Remote management of M2M applications
14. Key Features-Configurable Services
● 3 ways to add service packages:
○ Eclipse Marketplace
○ URL
○ Uploaded files.
● Can be configured during runtime through the web GUI
● Can access device hardware: GPIO, GPS, etc
15. Key Features-Kura Wires
● Add new processes to Kura in a block based visual representation
● Runs automatically once changes are applied; no need to compile
● Implementations for MQTT, building databases, data filtering, and more
● Additional assets can be added to Kura Wires by adding packages
17. Prototype Project with Temperature Sensor
Processing Tool:
● Handle streams of big data in real-time
● Low Latency
● Robust
● Simple / Flexible Implementation
23. SQL & NoSQL Comparison
SQL
● Vertically Scalable
● Predetermined data structure
● ACID (Atomicity, Consistency, Isolation
and Durability)
● Uses the Universal SQL (Structured
Query Language) language which
provides a powerful tool to manipulate
and define data
● Allows for complex queries
● Examples:
○ MySQL, Sqlite, and Postgres
NoSQL
● Horizontally Scalable
● Unstructured data storage
● CAP theorem ( Consistency, Availability
and Partition tolerance )
● Typically NoSQL relies on a collection of
documents and the syntax varies per the
database
● No standard interface for complex
queries
● Examples:
○ MongoDB, Redis, and Hbase
24. MySQL
● Open source relational database management system
(RDBMS)
● Provides a password system that is very flexible and
secure
● MySQL supports large databases, up to 50 million rows
or more in a table
● MySQL is also used in the industry as well (Facebook,
Twitter, Flickr and YouTube)
25. Where does
MySQL fit?
● Data Storage portion of the
architecture
● Apache Storm writes data
to MySQL
● Spark is then able to query
the database and provide a
visualization tool
27. Design Challenges (Kyle)
● Normalizing the database was challenging
because I never experienced high level data
modeling before. It took me a couple of office
hour sessions with my professor for me to
grasp the concept.
● Creating the connection between Apache
Storm was incredibly frustrating. It was a
gruelling debugging process that me and a
fellow engineer were stuck on for about a day
or two.
● The time spent to build up my knowledge of all
the technologies integrated into the framework.
31. Cloud Batch Processing
• Four components to Apache Spark:
• Spark SQL
• Introduces the concept of the “Resilient
Distributed Dataset”.
• Enables reading/writing and querying a
database using SQL.
• Spark MLib
• Machine Learning Algorithms
• Spark Streaming
• Real-time streaming using the process
of micro-batching.
• Spark GraphX
• Extends RDD’s for graphs and graph-
parallel computation.
33. Data Visualization
• Web-based book for data analysis
and visualization.
• Frames the question: “What is really
happening behind the hood?”.
• Versatile
• Interpreters for any language.
• Easy to share web-based notebook
• Able to set permissions per person.
• Integration with Apache Spark
Web Browser
Web
Server
Local
Interpreters
Zeppelin Daemon
Remote
Interpreters
Spark Master Node
Spark Worker
Node
Spark Worker
Node
34. Apache Spark Apache Zeppelin
✔ Python rather than Scala.
✔Plenty of Examples with a big community.
✔Troubleshooting was not that bad.
✔Basically Spark, but visualized.
✔Creating interpreters is useful.
✔Fun to explore and use.
🗙I used Python, but I’m new to Python
also.
🗙Understanding the program is one
thing, using it is another.
🗙New to Open-Source and Linux OS.
🗙 I learned Zeppelin, before Spark
🗙 Not as big of a community as Spark.
🗙 Troubleshooting was more difficult.
🗙More to Zeppelin than just Spark
interpreter..
36. Parking on campus
• Time-to-park is an issue on UTSA campus
• UTSA claims to have enough parking to meet typical demand
• Too many people in the same parking lot / Random distribution of
newly opened parking spots
• Solution: Accurate and reliable system to track (and predict)
parking patterns, providing real-time information to drivers to
assist in parking decisions