The new functionalities of KNOWAGE 8 allow to integrate advanced and increasingly personalized analyses in decision-making processes in a simple and rapid way.
Enforcing Bespoke Policies in KubernetesTorin Sandall
Kubernetes enables fully-automated, self-service management of large-scale, heterogenous deployments. These deployments are often managed by distributed engineering teams that have unique requirements for how the platform treats their workloads, but at the same time, they must conform to organization-wide constraints around cost, security, and performance. As Kubernetes matures, extensibility has become a critical feature that organizations can leverage to enforce their organization’s bespoke policies.
In this talk, Torin explains how to use extensibility features in Kubernetes (e.g., External Admission Control) to enforce custom policies over workloads. The talk shows how to build custom admission controllers using Initializers and Webhooks, and shows how the same features lay the groundwork for policy-based control through integration with third party policy engines like the Open Policy Agent project.
This document discusses using external scalers with KEDA (Kubernetes Event-driven Autoscaling) to allow KEDA to interact with systems outside the cluster to drive application scaling needs in a more customized way. It describes using gRPC to connect an external metrics adapter to KEDA's controller and provides an example demo of scaling applications based on messages in an ActiveMQ Artemis broker deployed outside the cluster. Relevant references are also included.
Platform Strategy to Deliver Digital Experiences on AzureWSO2
This slide deck introduces Choreo, a cloud native internal developer platform by Microsoft independent software vendor (ISV) Partner, WSO2. It enables your developers to create, deploy, and run new digital components like APIs, microservices, and integrations in serverless mode on any Kubernetes cluster with built-in DevSecOps.
Recording: https://wso2.com/choreo/resources/webinar/platform-strategy-to-deliver-digital-experiences-on-azure/
Github Copilot vs Amazon CodeWhisperer for Java developers at JCON 2023Vadym Kazulkin
The document compares GitHub Copilot, Amazon CodeWhisperer, and ChatGPT for Java developers. It provides an overview of each tool, compares their programming language support, IDE support, and pricing. It demonstrates their abilities for general tasks, simple functions, more complex algorithms, JUnit testing, and Spring Boot web development. It concludes that while the tools provide helpful suggestions, developers are still needed to ensure correctness and efficiency. GitHub Copilot and ChatGPT benefit from OpenAI, while Amazon CodeWhisperer needs quality improvements for Java but may leverage AWS services.
Case Study: Migration to GitLab (from Bitbucket) at AppsFlyerNoa Harel
AppsFlyer migrated from BitBucket to GitLab for their 150 users and 680 repositories. They wanted to leave the hosted BitBucket solution due to API call limits and latency. The migration process involved converting repositories from Mercurial to Git, setting up the GitLab architecture on Amazon Web Services with an EFS file system, educating teams, and creating custom tooling like a Python script to notify Slack. Lessons learned included issues restoring backups and increasing Unicorn workers. The full technical details are available at the provided URL.
As more and more organizations and individual users turn to Apache Flink for their streaming workloads, there is a bigger demand for additional functionality out-of-the-box. On one hand, there is demand for more low-level APIs that allow for more control, while on the other, users ask for more high-level additions that make the common cases easier to express. This talk will present the new concepts added to the Datastream API in Flink-1.2 and for the upcoming Flink-1.3 release that tried to consolidate the aforementioned goals. We will talk, among others, about the ProcessFunction, a new low level stream processing primitive that gives the user full control over how each event is processed and can register and react to timers, changes in the windowing logic that allow for more flexible windowing strategies, side outputs, and new features concerning the Flink connectors.
Integrating Apache Kafka Into Your Environmentconfluent
Watch this talk here: https://www.confluent.io/online-talks/integrating-apache-kafka-into-your-environment-on-demand
Integrating Apache Kafka with other systems in a reliable and scalable way is a key part of an event streaming platform. This session will show you how to get streams of data into and out of Kafka with Kafka Connect and REST Proxy, maintain data formats and ensure compatibility with Schema Registry and Avro, and build real-time stream processing applications with Confluent KSQL and Kafka Streams.
This session is part 4 of 4 in our Fundamentals for Apache Kafka series.
Change Data Streaming Patterns For Microservices With Debezium (Gunnar Morlin...confluent
Gunnar Morling presents on change data streaming patterns for microservices using Debezium. Debezium is an open source platform for change data capture that retrieves change events from transaction logs of different databases. It streams these events to Apache Kafka in a unified format. This allows microservices to stay synchronized by consuming the change events and keeping their local data stores in sync without direct database access. Various patterns are demonstrated including microservice data synchronization, leveraging single message transformations, and ensuring data quality.
Enforcing Bespoke Policies in KubernetesTorin Sandall
Kubernetes enables fully-automated, self-service management of large-scale, heterogenous deployments. These deployments are often managed by distributed engineering teams that have unique requirements for how the platform treats their workloads, but at the same time, they must conform to organization-wide constraints around cost, security, and performance. As Kubernetes matures, extensibility has become a critical feature that organizations can leverage to enforce their organization’s bespoke policies.
In this talk, Torin explains how to use extensibility features in Kubernetes (e.g., External Admission Control) to enforce custom policies over workloads. The talk shows how to build custom admission controllers using Initializers and Webhooks, and shows how the same features lay the groundwork for policy-based control through integration with third party policy engines like the Open Policy Agent project.
This document discusses using external scalers with KEDA (Kubernetes Event-driven Autoscaling) to allow KEDA to interact with systems outside the cluster to drive application scaling needs in a more customized way. It describes using gRPC to connect an external metrics adapter to KEDA's controller and provides an example demo of scaling applications based on messages in an ActiveMQ Artemis broker deployed outside the cluster. Relevant references are also included.
Platform Strategy to Deliver Digital Experiences on AzureWSO2
This slide deck introduces Choreo, a cloud native internal developer platform by Microsoft independent software vendor (ISV) Partner, WSO2. It enables your developers to create, deploy, and run new digital components like APIs, microservices, and integrations in serverless mode on any Kubernetes cluster with built-in DevSecOps.
Recording: https://wso2.com/choreo/resources/webinar/platform-strategy-to-deliver-digital-experiences-on-azure/
Github Copilot vs Amazon CodeWhisperer for Java developers at JCON 2023Vadym Kazulkin
The document compares GitHub Copilot, Amazon CodeWhisperer, and ChatGPT for Java developers. It provides an overview of each tool, compares their programming language support, IDE support, and pricing. It demonstrates their abilities for general tasks, simple functions, more complex algorithms, JUnit testing, and Spring Boot web development. It concludes that while the tools provide helpful suggestions, developers are still needed to ensure correctness and efficiency. GitHub Copilot and ChatGPT benefit from OpenAI, while Amazon CodeWhisperer needs quality improvements for Java but may leverage AWS services.
Case Study: Migration to GitLab (from Bitbucket) at AppsFlyerNoa Harel
AppsFlyer migrated from BitBucket to GitLab for their 150 users and 680 repositories. They wanted to leave the hosted BitBucket solution due to API call limits and latency. The migration process involved converting repositories from Mercurial to Git, setting up the GitLab architecture on Amazon Web Services with an EFS file system, educating teams, and creating custom tooling like a Python script to notify Slack. Lessons learned included issues restoring backups and increasing Unicorn workers. The full technical details are available at the provided URL.
As more and more organizations and individual users turn to Apache Flink for their streaming workloads, there is a bigger demand for additional functionality out-of-the-box. On one hand, there is demand for more low-level APIs that allow for more control, while on the other, users ask for more high-level additions that make the common cases easier to express. This talk will present the new concepts added to the Datastream API in Flink-1.2 and for the upcoming Flink-1.3 release that tried to consolidate the aforementioned goals. We will talk, among others, about the ProcessFunction, a new low level stream processing primitive that gives the user full control over how each event is processed and can register and react to timers, changes in the windowing logic that allow for more flexible windowing strategies, side outputs, and new features concerning the Flink connectors.
Integrating Apache Kafka Into Your Environmentconfluent
Watch this talk here: https://www.confluent.io/online-talks/integrating-apache-kafka-into-your-environment-on-demand
Integrating Apache Kafka with other systems in a reliable and scalable way is a key part of an event streaming platform. This session will show you how to get streams of data into and out of Kafka with Kafka Connect and REST Proxy, maintain data formats and ensure compatibility with Schema Registry and Avro, and build real-time stream processing applications with Confluent KSQL and Kafka Streams.
This session is part 4 of 4 in our Fundamentals for Apache Kafka series.
Change Data Streaming Patterns For Microservices With Debezium (Gunnar Morlin...confluent
Gunnar Morling presents on change data streaming patterns for microservices using Debezium. Debezium is an open source platform for change data capture that retrieves change events from transaction logs of different databases. It streams these events to Apache Kafka in a unified format. This allows microservices to stay synchronized by consuming the change events and keeping their local data stores in sync without direct database access. Various patterns are demonstrated including microservice data synchronization, leveraging single message transformations, and ensuring data quality.
This is a hands-on lab to introduce you to CCE and DevCloud on Huawei Cloud to build a DevSecOps Platform.
Jirayut Nimsaeng
Founder & CEO
Opsta (Thailand) Co., Ltd.
HUAWEI CONNECT 2022 Bangkok: Hands-on Lab
September 21, 2022
An overview of Azure API Management, common use cases, and how it helps organizations to govern, publish, secure, analyze, and manage APIs for internal and external consumption whether their running in the cloud or on-prem.
Cloud-Native PostgreSQL is a Kubernetes Operator for Postgres written by EDB entirely from scratch in the Go language and relying exclusively on the Kubernetes API.
This webinar covered:
- About DevOps & Cloud Native
- Overview of Cloud Native Postgres
- Storage for Postgres workloads in Kubernetes
- Start Using Cloud-Native Postgres
- Demo
In this presentation, Enterprise Knowledge Consultants Yanko Ivanov and Tatiana Baquero Cakici share their experience regarding how the role of a Business Analyst (BA) changes in an agile environment when a software development team is trying to move from a traditional software development methodology, such as waterfall, to an Agile approach. The presentation discusses some of the main challenges that BAs face when moving to Agile, compares the difference in BA interactions between the different project approaches, and the key areas where BAs are facing a mental shift in their transition to Agile.
Kubernetes GitOps featuring GitHub, Kustomize and ArgoCDSunnyvale
A brief dissertation about using GitOps paradigm to operate an application on multiple Kubernetes environments thanks to GitHub, ArgoCD and Kustomize. A talk about this matters has been taken at the event #CloudConf2020
The monolith to cloud-native, microservices evolution has driven a shift from monitoring to observability. OpenTelemetry, a merger of the OpenTracing and OpenCensus projects, is enabling Observability 2.0. This talk gives an overview of the OpenTelemetry project and then outlines some production-proven architectures for improving the observability of your applications and systems.
This document discusses Netflix's "Paved Road" concept, which formalizes expectations and commitments between centralized teams and engineering teams. The Paved Road provides integrated, supported tools and services to help engineers focus on delivering business value. It is not mandatory for teams to use. Investments are made in partnership with teams to understand value. The Paved Road aims to streamline processes like security and development while allowing flexibility for teams. It helps reduce costs and friction so engineers can focus on their work.
This document provides an agenda for getting started with git and GitHub. It discusses version control and how git allows tracking changes to files. It explains how to set up git locally and initialize a repository. Basic git commands like add, commit, status, and log are covered. The document also discusses branching in git and what GitHub is. It provides steps for creating a GitHub account and repository and collaborating through forking, cloning, and pull requests. Finally, it discusses the benefits of open source work.
Everyone wants observability into their system, but find themselves with too many vendors and tools, each with its own API, SDK, agent and collectors.
In this talk I will present OpenTelemetry, an ambitious open source project with the promise of a unified framework for collecting observability data. With OpenTelemetry you could instrument your application in a vendor-agnostic way, and then analyze the telemetry data in your backend tool of choice, whether Prometheus, Jaeger, Zipkin, or others.
I will cover the current state of the various projects of OpenTelemetry (across programming languages, exporters, receivers, protocols), some of which not even GA yet, and provide useful guidance on how to get started with it.
Salesforce enabling real time scenarios at scale using kafkaThomas Alex
Nishant Gupta from Salesforce talked about Ajna, a service for monitoring system health across global data centers in real time, and how Kafka is at the center of this system. The talk covers the scenario, key challenges, learnings and best practices.
Build Real-Time Applications with Databricks StreamingDatabricks
This document discusses using Databricks, Spark, and Power BI for real-time data streaming. It describes a use case of a fire department needing real-time reporting of equipment locations, personnel statuses, and active incidents. The solution involves ingesting event data using Azure Event Hubs, processing the stream using Databricks and Spark Structured Streaming, storing the results in Delta Lake, and visualizing the data in Power BI dashboards. It then demonstrates the architecture by walking through creating Delta tables, streaming from Event Hubs to Delta Lake, and running a sample event simulator.
Performance Engineering Masterclass: Introduction to Modern PerformanceScyllaDB
Leandro Melendez from Grafana k6 starts off by providing a grounding in current expectations of what performance engineering and load testing entail. This session defines the modern challenges developers face, including continuous performance principles, Service Level Objectives (SLOs), and Service Level Indicators (SLIs). It delineates best practices and provides hands-on examples using Grafana k6, an open source modern load testing tool.
Slide deck of the presentation done at the Hactoberfest 2020 Singapore event. The talk and demo showed GitHub Actions in practice with examples of Github Superlinter, SonarCloud integration and CI CD to Azure Kubernetes service.
The recording of the session is available on YouTube
https://youtu.be/sFvCj62wmWU?t=6732&WT.mc_id=AZ-MVP-5003170
This document discusses Spring Cloud Data Flow, which provides tools for creating complex data pipelines for streaming and batch processes. It describes how Spring Cloud Data Flow orchestrates Spring Cloud Stream for event streaming and Spring Cloud Task for short-lived microservices. The document also references challenges around integration, event streaming, and orchestrating short-lived tasks that Spring Cloud Data Flow addresses.
Understanding the GitOps Workflow and CICD Pipeline - What It Is, Why It Matt...Gibran Badrulzaman
Travelio Tech Talks 2022 presentation
The recommended workflow for implementing GitOps with Kubernetes manifests is known as trunk-based development. This method defines one branch as the "trunk" and carries out development on each environment in a different short-lived branch. When development is complete for that environment, the developer creates a pull request for the branch to the trunk. Developers can also create a fork to work on an environment, and then create a branch to merge the fork into the trunk.
Once the proper approvals are done, the pull request (or the branch from the fork) gets merged into the trunk. The branch for that feature is deleted, keeping your branches to a minimum. Trunk-based development trades branches for directories.
You can think of the trunk as a "main" or primary branch. production and prod are popular names for the trunk branch.
Trunk-based development came about to enable continuous integration and continuous delivery by supplying a development model focused on the fast delivery of changes to applications. But this model also works for GitOps repositories because it keeps things simple and more in tune with how Kustomize and Helm work. When you record deltas between environments, you can clearly see what changes will be merged into the trunk. You won’t have to cherry-pick nearly as often, and you’ll have the confidence that what is in your Git repository is what is actually going into your environment. This is what you want in a GitOps workflow.
The document provides an agenda for a presentation on the role of a business analyst on Agile projects. The agenda includes an overview of Agile, the role of a business analyst on traditional and Agile projects, why business analysts are important for project success, and a question and answer session. The presenter is Bill Gaiennie who has 17 years of software development experience and has trained over 500 teams on Agile.
The document discusses Pentaho's business intelligence (BI) platform for big data analytics. It describes Pentaho as providing a modern, unified platform for data integration and analytics that allows for native integration into the big data ecosystem. It highlights Pentaho's open source development model and that it has over 1,000 commercial customers and 10,000 production deployments. Several use cases are presented that demonstrate how Pentaho helps customers unlock value from big data stores.
Using Visualization to Succeed with Big Data Pactera_US
The document summarizes a webinar on big data visualization. It discusses drivers for the big data visualization market and new tools emerging. It then profiles several major vendors that offer big data visualization solutions, including Microsoft, QlikView, TIBCO, Tableau, Platfora, Datameer, Splunk, Jaspersoft, and Alpine Data. It concludes with an overview of how Pactera can help clients build advanced analytics solutions.
This is a hands-on lab to introduce you to CCE and DevCloud on Huawei Cloud to build a DevSecOps Platform.
Jirayut Nimsaeng
Founder & CEO
Opsta (Thailand) Co., Ltd.
HUAWEI CONNECT 2022 Bangkok: Hands-on Lab
September 21, 2022
An overview of Azure API Management, common use cases, and how it helps organizations to govern, publish, secure, analyze, and manage APIs for internal and external consumption whether their running in the cloud or on-prem.
Cloud-Native PostgreSQL is a Kubernetes Operator for Postgres written by EDB entirely from scratch in the Go language and relying exclusively on the Kubernetes API.
This webinar covered:
- About DevOps & Cloud Native
- Overview of Cloud Native Postgres
- Storage for Postgres workloads in Kubernetes
- Start Using Cloud-Native Postgres
- Demo
In this presentation, Enterprise Knowledge Consultants Yanko Ivanov and Tatiana Baquero Cakici share their experience regarding how the role of a Business Analyst (BA) changes in an agile environment when a software development team is trying to move from a traditional software development methodology, such as waterfall, to an Agile approach. The presentation discusses some of the main challenges that BAs face when moving to Agile, compares the difference in BA interactions between the different project approaches, and the key areas where BAs are facing a mental shift in their transition to Agile.
Kubernetes GitOps featuring GitHub, Kustomize and ArgoCDSunnyvale
A brief dissertation about using GitOps paradigm to operate an application on multiple Kubernetes environments thanks to GitHub, ArgoCD and Kustomize. A talk about this matters has been taken at the event #CloudConf2020
The monolith to cloud-native, microservices evolution has driven a shift from monitoring to observability. OpenTelemetry, a merger of the OpenTracing and OpenCensus projects, is enabling Observability 2.0. This talk gives an overview of the OpenTelemetry project and then outlines some production-proven architectures for improving the observability of your applications and systems.
This document discusses Netflix's "Paved Road" concept, which formalizes expectations and commitments between centralized teams and engineering teams. The Paved Road provides integrated, supported tools and services to help engineers focus on delivering business value. It is not mandatory for teams to use. Investments are made in partnership with teams to understand value. The Paved Road aims to streamline processes like security and development while allowing flexibility for teams. It helps reduce costs and friction so engineers can focus on their work.
This document provides an agenda for getting started with git and GitHub. It discusses version control and how git allows tracking changes to files. It explains how to set up git locally and initialize a repository. Basic git commands like add, commit, status, and log are covered. The document also discusses branching in git and what GitHub is. It provides steps for creating a GitHub account and repository and collaborating through forking, cloning, and pull requests. Finally, it discusses the benefits of open source work.
Everyone wants observability into their system, but find themselves with too many vendors and tools, each with its own API, SDK, agent and collectors.
In this talk I will present OpenTelemetry, an ambitious open source project with the promise of a unified framework for collecting observability data. With OpenTelemetry you could instrument your application in a vendor-agnostic way, and then analyze the telemetry data in your backend tool of choice, whether Prometheus, Jaeger, Zipkin, or others.
I will cover the current state of the various projects of OpenTelemetry (across programming languages, exporters, receivers, protocols), some of which not even GA yet, and provide useful guidance on how to get started with it.
Salesforce enabling real time scenarios at scale using kafkaThomas Alex
Nishant Gupta from Salesforce talked about Ajna, a service for monitoring system health across global data centers in real time, and how Kafka is at the center of this system. The talk covers the scenario, key challenges, learnings and best practices.
Build Real-Time Applications with Databricks StreamingDatabricks
This document discusses using Databricks, Spark, and Power BI for real-time data streaming. It describes a use case of a fire department needing real-time reporting of equipment locations, personnel statuses, and active incidents. The solution involves ingesting event data using Azure Event Hubs, processing the stream using Databricks and Spark Structured Streaming, storing the results in Delta Lake, and visualizing the data in Power BI dashboards. It then demonstrates the architecture by walking through creating Delta tables, streaming from Event Hubs to Delta Lake, and running a sample event simulator.
Performance Engineering Masterclass: Introduction to Modern PerformanceScyllaDB
Leandro Melendez from Grafana k6 starts off by providing a grounding in current expectations of what performance engineering and load testing entail. This session defines the modern challenges developers face, including continuous performance principles, Service Level Objectives (SLOs), and Service Level Indicators (SLIs). It delineates best practices and provides hands-on examples using Grafana k6, an open source modern load testing tool.
Slide deck of the presentation done at the Hactoberfest 2020 Singapore event. The talk and demo showed GitHub Actions in practice with examples of Github Superlinter, SonarCloud integration and CI CD to Azure Kubernetes service.
The recording of the session is available on YouTube
https://youtu.be/sFvCj62wmWU?t=6732&WT.mc_id=AZ-MVP-5003170
This document discusses Spring Cloud Data Flow, which provides tools for creating complex data pipelines for streaming and batch processes. It describes how Spring Cloud Data Flow orchestrates Spring Cloud Stream for event streaming and Spring Cloud Task for short-lived microservices. The document also references challenges around integration, event streaming, and orchestrating short-lived tasks that Spring Cloud Data Flow addresses.
Understanding the GitOps Workflow and CICD Pipeline - What It Is, Why It Matt...Gibran Badrulzaman
Travelio Tech Talks 2022 presentation
The recommended workflow for implementing GitOps with Kubernetes manifests is known as trunk-based development. This method defines one branch as the "trunk" and carries out development on each environment in a different short-lived branch. When development is complete for that environment, the developer creates a pull request for the branch to the trunk. Developers can also create a fork to work on an environment, and then create a branch to merge the fork into the trunk.
Once the proper approvals are done, the pull request (or the branch from the fork) gets merged into the trunk. The branch for that feature is deleted, keeping your branches to a minimum. Trunk-based development trades branches for directories.
You can think of the trunk as a "main" or primary branch. production and prod are popular names for the trunk branch.
Trunk-based development came about to enable continuous integration and continuous delivery by supplying a development model focused on the fast delivery of changes to applications. But this model also works for GitOps repositories because it keeps things simple and more in tune with how Kustomize and Helm work. When you record deltas between environments, you can clearly see what changes will be merged into the trunk. You won’t have to cherry-pick nearly as often, and you’ll have the confidence that what is in your Git repository is what is actually going into your environment. This is what you want in a GitOps workflow.
The document provides an agenda for a presentation on the role of a business analyst on Agile projects. The agenda includes an overview of Agile, the role of a business analyst on traditional and Agile projects, why business analysts are important for project success, and a question and answer session. The presenter is Bill Gaiennie who has 17 years of software development experience and has trained over 500 teams on Agile.
The document discusses Pentaho's business intelligence (BI) platform for big data analytics. It describes Pentaho as providing a modern, unified platform for data integration and analytics that allows for native integration into the big data ecosystem. It highlights Pentaho's open source development model and that it has over 1,000 commercial customers and 10,000 production deployments. Several use cases are presented that demonstrate how Pentaho helps customers unlock value from big data stores.
Using Visualization to Succeed with Big Data Pactera_US
The document summarizes a webinar on big data visualization. It discusses drivers for the big data visualization market and new tools emerging. It then profiles several major vendors that offer big data visualization solutions, including Microsoft, QlikView, TIBCO, Tableau, Platfora, Datameer, Splunk, Jaspersoft, and Alpine Data. It concludes with an overview of how Pactera can help clients build advanced analytics solutions.
Augmentation, Collaboration, Governance: Defining the Future of Self-Service BIDenodo
Watch full webinar here: https://bit.ly/3zVJRRf
According to Dresner Advisory’s 2020 Self-Service Business Intelligence Market Study, 62% of the responding organizations say self-service BI is critical for their business. If we look deeper into the need for today’s self-service BI, it’s beyond some Executives and Business Users being enabled by IT for self-service dashboarding or report generation. Predictive analytics, self-service data preparation, collaborative data exploration are all different facets of new generation self-service BI. While democratization of data for self-service BI holds many benefits, strict data governance becomes increasingly important alongside.
In this session we will discuss:
- The latest trends and scopes of self-service BI
- The role of logical data fabric in self-service BI
- How Denodo enables self-service BI for a wide range of users - Customer case study on self-service BI
Gridlogics is a leading provider of products and custom software solutions for patent research, management, data analysis and project management. Our products leverage the latest techniques in information retrieval, data mining and visualizations to help clients globally in deriving actionable intelligence from the masses of patent data.
This document provides an overview of new features and enhancements in SAP BusinessObjects BI 4.1. Key updates include simplified deployment tools, support for additional data sources and platforms, enhanced self-service capabilities, and improvements to visualization, collaboration, and mobile access. The document also summarizes new features for several SAP analytics products, including Lumira, Design Studio, Analysis for Office, and Mobile BI.
zData BI & Advanced Analytics Platform + 8 Week Pilot ProgramszData Inc.
This document describes zData's BI/Advanced Analytics Platform and Pilot Programs. The platform provides tools for storing, collaborating on, analyzing, and visualizing large amounts of data. It offers machine learning and predictive analytics. The platform can be deployed on-premise or in the cloud. zData also offers an 8-week pilot program that provides up to 1TB of data storage and full access to the platform's tools and services to test out the Big Data solution.
Logitech - LOGITECH ACCELERATES CLOUD ANALYTICS USING DATA VIRTUALIZATIONAvinash Deshpande
Logitech is using data virtualization and cloud analytics to accelerate insights from their growing data. They implemented Denodo on AWS to create a decentralized self-service analytics environment. This allows business users to perform descriptive, diagnostic, predictive, and prescriptive analytics. Logitech aims to provide real-time, natural language insights on desktops and phones to support business decisions. Data virtualization has helped Logitech reduce costs while improving data access, governance, and analytics capabilities.
This talk presents some use cases from both public sector and private companies where KNOWAGE, the open source analytics and business intelligence suite made in Italy, enabled a deeper understanding of ongoing phenomena and allowed reaching data-driven decisions. For each project, the speaker will highlight goals and initial requirements, the open source stack exploited, the values brought by the solution and the benefits gained through open source.
Alten calsoft labs analytics service offeringsSandeep Vyas
This document provides an overview of ALTEN Calsoft Labs' BI and analytics solutions. It discusses challenges companies face with disparate data sources, lengthy data preparation processes, and hardware dependencies. The company's unified data analysis platform addresses these challenges by processing data more efficiently and reducing application development time. The platform integrates structured, unstructured, and semi-structured data from various sources and formats it for analysis and visualization to provide business insights.
Exclusive Verizon Employee Webinar: Getting More From Your CDR DataPentaho
This document discusses a project between Pentaho and Verizon to leverage big data analytics. Verizon generates vast amounts of call detail record (CDR) data from mobile networks that is currently stored in a data warehouse for 2 years and then archived to tape. Pentaho's platform will help optimize the data warehouse by using Hadoop to store all CDR data history. This will free up data warehouse capacity for high value data and allow analysis of the full 10 years of CDR data. Pentaho tools will ingest raw CDR data into Hadoop, execute MapReduce jobs to enrich the data, load results into Hive, and enable analyzing the data to understand calling patterns by geography over time.
How to Empower Your Business Users with Oracle Data VisualizationPerficient, Inc.
With Oracle Data Visualization Cloud Service, your business users can perform self-service analytics, spot patterns, trends, correlations, and construct visual data stories for greater insight into how your product, service, or organization is performing.
In this webinar, we demonstrated how easily users can explore their data in new and different ways through stunning visualizations automatically, promoting self-service discovery.
Discussion included:
-In-depth review of Oracle Data Visualization Cloud Service
-Connecting different data sets like HCM, ERP, Sales Cloud and more
-Mobile and security
-Demo taking a real-world business use case from end to end
Jayaram_Parida- Big Data Architect and Technical Scrum MasterJayaram Parida
Jayaram Parida has over 19 years of experience in IT, including 3 years as a Big Data Technical Solution Architect. He has extensive skills in technologies like Hadoop, HDFS, HBase, Hive, MapReduce, Kafka, Storm, YARN, Pig, Python, and data analytics tools. He has experience architecting and developing big data solutions for clients in various industries. His roles have included designing Hadoop infrastructures, developing real-time analytics platforms, and creating visualizations and reports.
Data & Analytics with CIS & Microsoft PlatformsSonata Software
Sonata Software provides data and analytics services using Microsoft platforms and technologies. They help customers leverage data to drive intelligent actions and personalization at scale. Sonata has expertise in data warehousing, business analytics, AI, machine learning, and developing industry-specific analytics solutions and AI accelerators on the Microsoft stack. They assist customers with data strategy, analytics, visualization, and migrating to Azure-based platforms.
KNOWAGE evolution in 2022 mainly focuses on: new data preparation module and data federation in self-service process, augmented analytics to support every end-user touch point and provide automatic insights, usability and performance for a new effective UI, a core offering as SaaS ABI solution.
Keyrus is a data analytics consultancy that helps customers make data-driven decisions. It provides services including big data solutions, data management strategies, data integration, business intelligence dashboards, predictive analytics, and data science consulting. Keyrus has expertise in structured and unstructured data, data discovery visualization tools, and building end-to-end analytics solutions. Sample projects include building Hadoop environments for large telecom data and creating risk monitoring dashboards for investment banks.
Keyrus is a data analytics consultancy that helps customers make data-driven decisions. It provides services including big data solutions, data management strategies, data integration, machine learning, predictive analytics, and data visualization dashboards. Keyrus consultants have skills in databases, data modeling, programming, and business requirements. For example, for a bank, Keyrus built interactive dashboards from multiple databases to provide regulators with risk monitoring dashboards.
Watch full webinar here: https://bit.ly/3H4vrlD
Data as a strategic imperative for any company to compete, New common self-service data experience required for all things intelligent, Modern data platform focused on producing data products, Data platform, product, people, process key solution ingredients and Denodo is the future and time is now to get started.
From Business Idea to Successful Delivery by Serhiy Haziyev & Olha Hrytsay, S...SoftServe
The document discusses transitioning product development to an open innovation model using outsourcing partners. It provides an overview of SoftServe's capabilities including big data solutions, reference architectures, and case studies. The document also outlines SoftServe's involvement in collaborating with clients throughout the product development lifecycle from envisioning ideas to commercialization. Finally, it provides a checklist for successful delivery when partnering with outsourcing providers for big data and product development.
Analyse in detail all the different widgets available in KNOWAGE dashboards, learn how to set up your first dashboard and discover useful tips to improve data visualization.
KNOWAGE is the open source analytics and business intelligence suite.
KNOWAGE CUSTOM CHART WIDGET: a technical guideKNOWAGE
The Custom Chart widget allows you to create charts using native library directly. You are now able to use your favorite library to realize your own fully customized data visualization.
Knowage is the open source suite for any modern business analytics needing.
Knowage is an open source business intelligence suite that provides location intelligence and self-service reporting capabilities. It allows for analysis of big data from multiple sources and interactive dashboards on desktop and mobile. Key features include mapping widgets, security and profiling, unlimited users and data exploration.
Knowage features for advanced analytics: Data/text mining, Multidimensional what-if based on OLAP, Simulation over measures and dimensions, Python widget.
This document provides a technical guide for using the map widget within Knowage dashboards. It discusses how to plot business data on a map and connect map elements to other dashboard widgets. The guide walks through configuring the map widget by adding layers from datasets, specifying latitude and longitude columns, selecting measures, and customizing the map style. It also demonstrates designing a sample dashboard to monitor the number of car accidents in NYC using the map widget alongside other elements like charts and selectors.
The new R/Python widget allows to embed R and Python scripts directly within a cockpit filling a gap between datavisualization and datascience. With this widget, datascientists are able to integrate python facilities in a heterogeneous and multi-source environment and to present advanced analytics to a broader audience...
This training is intended to support Knowage beginners and guide them though the first essential steps required to take the most out of Knowage experience.
In the first session:
- Knowage installation tutorial
- Knowage overview
- Datasource connection and dataset creation.
In the second session:
- Cockpit creation step by step.
Knowage official presentation 2018. Knowage is the only open source and full capabilities suite for any modern business analysis.
Feel free to use this presentation to present and promote Knowage open source suite!
We are pleased to share with you the latest VCOSA statistical report on the cotton and yarn industry for the month of March 2024.
Starting from January 2024, the full weekly and monthly reports will only be available for free to VCOSA members. To access the complete weekly report with figures, charts, and detailed analysis of the cotton fiber market in the past week, interested parties are kindly requested to contact VCOSA to subscribe to the newsletter.
Codeless Generative AI Pipelines
(GenAI with Milvus)
https://ml.dssconf.pl/user.html#!/lecture/DSSML24-041a/rate
Discover the potential of real-time streaming in the context of GenAI as we delve into the intricacies of Apache NiFi and its capabilities. Learn how this tool can significantly simplify the data engineering workflow for GenAI applications, allowing you to focus on the creative aspects rather than the technical complexities. I will guide you through practical examples and use cases, showing the impact of automation on prompt building. From data ingestion to transformation and delivery, witness how Apache NiFi streamlines the entire pipeline, ensuring a smooth and hassle-free experience.
Timothy Spann
https://www.youtube.com/@FLaNK-Stack
https://medium.com/@tspann
https://www.datainmotion.dev/
milvus, unstructured data, vector database, zilliz, cloud, vectors, python, deep learning, generative ai, genai, nifi, kafka, flink, streaming, iot, edge
Generative Classifiers: Classifying with Bayesian decision theory, Bayes’ rule, Naïve Bayes classifier.
Discriminative Classifiers: Logistic Regression, Decision Trees: Training and Visualizing a Decision Tree, Making Predictions, Estimating Class Probabilities, The CART Training Algorithm, Attribute selection measures- Gini impurity; Entropy, Regularization Hyperparameters, Regression Trees, Linear Support vector machines.
2. Grazia Cazzin is the offering manager for Data & Analytics competency
center in the Research and Innovation Division of Engineering Group
S.p.A. She is the Knowage Labs Director (www.knowage-suite.com),
the founder of the SpagoBI project (http://www.spagobi.org) and its
historical project leader. Working in the IT field since 1992, she has been
involved in enterprise application development, data modelling, data
warehousing, dimensional analysis and business intelligence; she has
gained valuable expertise working in several market sectors (industry,
finance, public administration) and covering several thematic areas
(ERP, MRP, MPS, Enterprise Portals, CRM, DWH and BI, Big Data,
Analytics). She's currently involved in all the activities concerning Big
Data Analysis, as for SmartCity, Smart Agriculture and Industry 4.0
scenarios.
INTRODUCTION / ABOUT ME
4. OVERVIEW / THE ABI MARKET
# Capability Description KNOWAGE
1 Security
Capabilities that enable platform security, administering of users, auditing of platform access and
authentication
2 Manageability
Capabilities that track usage of the ABI platform and manage how information is shared (and by
whom).
3 Cloud-Enabled Analytics
The ability to support building, deployment and management of analytics in the cloud, based on data
stored both in the cloud and on-premises
4 Data Source Connectivity
Capabilities that enable users to connect to, query and ingest data, while optimizing for performance
5 Data Preparation
Support for drag-and-drop, user-driven combination of data from different sources, and the creation
of analytic models (such as user-defined measures, sets, groups and hierarchies).
6 Catalog
The ability to automatically generate and curate a searchable catalog of analytic content, thus making
it easier for analytic consumers to know what content is available.
7 Automated Insights
A core attribute of augmented analytics, this is the application of ML techniques to automatically
generate findings for end users (for example, by identifying the most important attributes in a
dataset).
8 Data Visualization
Support for highly interactive dashboards and exploration of data through manipulation of chart
images
9 Natural Language Query (NLQ)
This enables users to ask questions and query data and analytic content using terms that are either
typed into a search box or spoken
10 Data Storytelling
The ability to combine interactive data visualization with narrative techniques in order to package and
deliver analytic content in a compelling, easily understood form for presentation to decision makers
11
Natural Language Generation
(NLG)
The automatic creation of linguistically rich descriptions of answers, data and analytic content. Within
the analytics context, as the user interacts with data, the narrative changes dynamically to explain key
findings or the meaning of charts or dashboards
12 Reporting
The ability to create and distribute (or “burst”) pixel-perfect, grid-layout, multipage reports to users
on a scheduled basis.
5. OVERVIEW / KNOWAGE IN A NUTSHELL
• Analytics and Business Intelligence (ABI) criteria
• Advanced features for modern BI
• Self-service & ad-hoc reporting
• Mash-up and high-customization
• Responsive GUI, ready for mobile
• Multiple data sources (in the same document)
• Any data, many goals, open architecture
• OEM enabler
• Open source and open standard
• Community edition & enterprise edition
• On-premise/on-cloud
6. INTERACTIVE DASHBOARD
STATIC & SCHEDULED REPORTS
OVERVIEW / FULL ABI SUITE
SELF-SERVICE ANALYSIS
KPIs
OLAP & WHAT-IF
VISUAL DATA EXPLORATION
7. EMBEDDED CUSTOM CODE FULL ADMIN CAPABILITIES
MY WORKSPACE & OPEN DATA
HIGH LEVEL CUSTOMIZATION
OVERVIEW / FULL ABI SUITE
SPATIAL DATA DATA MASHUP
13. OPEN BI / CROSS DATA & MASH-UP
Spatial data, multilayer view Solr index, faceted view, full text search
14. HTML5/CSS3 INJECTION for powerful DATAVIZ (HTML WIDGET) Custom/external libraries for new visualizations (CUSTOM CHART)
R/PYTHON INJECTION for advanced DATAVIZ (R/Python widget)
OPEN BI / HIGH-CUSTOMIZATION
R/PYTHON INJECTION for embedding analytics (R/Python functions)
15. OPEN BI / HIGH-CUSTOMIZATION
GALLERY FOR CUSTOM WIDGETS
• Pre-built custom widgets (HTML, R/PYTHON, CHART)
• Design one, use everywhere
• Enrich and empower your dashboards
• Enable templating
CATALOGUE FOR CUSTOM FUNCTIONS
• Pre-built analytical functions
• Make data scientists part of the BI team
• Empower data set and widgets in your dashboards
• Spread analytics
16. Input data Explore Transform Use / Save
• Add data
• Explore
• Clean/update
• Split
• Merge
• Enrich / Classify
• Filter
• Grouping
• Swap/Pivoting
• Discretization
• Masking
ON-GOING / DATA PREPARATION
20. Enterprise edition (EE)
• Available for fee, under annual subscription
• Private repository
• Usable with AGPL exceptions (no constraints)
• OEM for commercial products
• Facilities for admin, advanced features for end-users
• Guaranteed third-part rights
• Maintenance and professional services
• Professional support with SLA
• Professional documentation
Community edition (CE)
• Available for free
• Open repository
• Usable under the AGPL license
• OEM result must be open source (AGPL)
• Full usability for end-users
• Not guaranteed third-part libraries rights
• Community-based support
• No professional services
• General documentation
OFFERING / CE vs EE
21. Subscription based on :
• Modules (SI,ER)
• Plugin (SD, LI, PM, CA)
• # Core
• Unlimited users
• Unlimited data
• Multi-environment (dev/prod)
OEM formula:
• % of the OEM product licence
(subscription or licence)
• ≈30% of the usual subscription
• Minimum commitment
• Progressive discounts (# clients)
OFFERING / PRICING MODEL
22. OFFERING / B2C business models
EE - ANNUAL
SUBSCRIPTION
EE - ANNUAL
LICENSE
COMMUNITY
EDITION
• 90% of capabilities
• AGPL license
• POC & testing (more than TRIAL)
• Download & begin
• No-critical projects
• Source code available
• Community support
• 100% of capabilities
• AGPL with exceptions
• Third-part licenses
• Easy to manage
• No-critical projects
• Full product at low cost
• Professional Services
• 100% of capabilities
• AGPL with exceptions
• Third-part licenses
• Easy to manage
• Critical projects
• Full-served product
• Maintenance with SLA
• Professional Services
KNOWAGE for you
KNOWAGE for your customer
23. OFFERING / B2B business models
REBRANDING
INTEGRATOR
PARTNERSHIP
OEM SaaS
RESELLER
• Marketing materials
• Discount (fee) on price list
• EE for study purposes
• # credits for support / TOJ
• Up-to-date webinar
• Marketing materials
• Discount (fee) on price list
Fee on KN licenses
Fee on KN services
ABI projects (your services)
• Empower your product
• Competitive advantage
• Perpetual licence
• Source code
• Keep your data
• Your brand
• Structure your knowledge
• Perpetual licence
• Source code
• Your service
• Your brand
Fee on KN licenses
Fee on KN services
High price list (new
capabilities)
New service line (data
analysis)
Ready-to-use knowledge
Tool & skills
• Service provider
• Perpetual licence
• Source code
• OEM/rebranding
• Your brand
Your service
24. Looking at services:
• Quick start
• Fast time-to-market
• End-to-end support
• Analytical tasks
• Templating
• Integrated Security & SSO
• Usable dataviz pattern
• Full customization
• Custom development
Looking at the product:
• Openess
• Flexibility
• High-customization
• Pay when use (in production)
• Economical scaling (spread adoption)
• Flexible model (b2b,b2c)
• Hybrid architecture
• Trusted open source
• Open roadmap
OFFERING / BENEFITS WORKING WITH US
25. Installer and binaries Source Code Docker Chef Script Amazon AWS Marketplace
https://www.knowage-suite.com
https://github.com/KnowageLabs
http://release.ow2.org/knowage/
OFFERING / HOW TO GET KNOWAGE