This webinar gives an overview of the Pentaho technology stack and then delves deep into its features like ETL, Reporting, Dashboards, Analytics and Big Data. The webinar also facilitates a cross industry perspective and how Pentaho can be leveraged effectively for decision making. In the end, it also highlights how apart from strong technological features, low TCO is central to Pentaho’s value proposition. For BI technology enthusiasts, this webinar presents easiest ways to learn an end to end analytics tool. For those who are interested in developing a BI / Analytics toolset for their organization, this webinar presents an interesting option of leveraging low cost technology. For big data enthusiasts, this webinar presents overview of how Pentaho has come out as a leader in data integration space for Big data.
Pentaho is one of the leading niche players in Business Intelligence and Big Data Analytics. It offers a comprehensive, end-to-end open source platform for Data Integration and Business Analytics. Pentaho’s leading product: Pentaho Business Analytics is a data integration, BI and analytics platform composed of ETL, OLAP, reporting, interactive dashboards, ad hoc analysis, data mining and predictive analytics.
Pentaho Data Integration. Preparing and blending data from any source for analytics. Thus, enabling data-driven decision making. Application for education, specially, academic and learning analytics.
Here is a case study that I developed to explain the different sets of functionality with the Pentaho Suite. I focused on the functionality, features, illustrative tools and key strengths. I've provided an understanding toward evaluating BI tools when selecting vendors. Enjoy!
Pentaho Data Integration: Extrayendo, integrando, normalizando y preparando m...Alex Rayón Jerez
Sesión de Pentaho Data Integration impartida en Noviembre de 2015 en el marco del Programa de Big Data y Business Intelligence de la Universidad de Deusto (detalle aquí http://bit.ly/1PhIVgJ).
Pentaho Data Integration. Preparing and blending data from any source for analytics. Thus, enabling data-driven decision making. Application for education, specially, academic and learning analytics.
Here is a case study that I developed to explain the different sets of functionality with the Pentaho Suite. I focused on the functionality, features, illustrative tools and key strengths. I've provided an understanding toward evaluating BI tools when selecting vendors. Enjoy!
Pentaho Data Integration: Extrayendo, integrando, normalizando y preparando m...Alex Rayón Jerez
Sesión de Pentaho Data Integration impartida en Noviembre de 2015 en el marco del Programa de Big Data y Business Intelligence de la Universidad de Deusto (detalle aquí http://bit.ly/1PhIVgJ).
A world's one of the first complete Online Web-based Development Frameworks to develop and deploy Decision Support Systems, Knowledge-based systems, Web-sites and Applications backed by Expert System, Case-Based Reasoning and Hybrid AI Technologies
User can run queries via MicroStrategy’s visual interface without the need to write unfamiliar HiveQL or MapReduce scripts. In essence, any user, without programming skill in Hadoop, can ask questions against vast volumes of structured and unstructured data to gain valuable business insights.
Watch the companion webinar at:
Join John Sterrett, Senior Advisor at Linchpin People and Scott Walz, Director of Software Consultants, to learn how execution plans get invalidated and why data skew could be the root cause to seeing different execution plans for the same query. We will look at options for forcing a query to use a particular execution plan. Finally, you will learn how this complex problem can be identified and resolved simply using a new feature in SQL Server 2016 called Query Store.
D365 Finance & Operations - Data & Analytics (see newer release of this docum...Gina Pabalan
This very comprehensive white paper provides a detailed and clear overview of Microsoft's D365 Finance & Operations solutions to support Data & Analytics.
There is a newer version of this available - search SlideShare for the new version of this deck.
A world's one of the first complete Online Web-based Development Frameworks to develop and deploy Decision Support Systems, Knowledge-based systems, Web-sites and Applications backed by Expert System, Case-Based Reasoning and Hybrid AI Technologies
User can run queries via MicroStrategy’s visual interface without the need to write unfamiliar HiveQL or MapReduce scripts. In essence, any user, without programming skill in Hadoop, can ask questions against vast volumes of structured and unstructured data to gain valuable business insights.
Watch the companion webinar at:
Join John Sterrett, Senior Advisor at Linchpin People and Scott Walz, Director of Software Consultants, to learn how execution plans get invalidated and why data skew could be the root cause to seeing different execution plans for the same query. We will look at options for forcing a query to use a particular execution plan. Finally, you will learn how this complex problem can be identified and resolved simply using a new feature in SQL Server 2016 called Query Store.
D365 Finance & Operations - Data & Analytics (see newer release of this docum...Gina Pabalan
This very comprehensive white paper provides a detailed and clear overview of Microsoft's D365 Finance & Operations solutions to support Data & Analytics.
There is a newer version of this available - search SlideShare for the new version of this deck.
An introduction to Docker native clustering: Swarm.
Deployment and configuration, integration with Consul, for a product-like cluster to serve web-application with multiple containers on multiple hosts. #dockerops
Scaling Jenkins with Docker and KubernetesCarlos Sanchez
Docker is revolutionizing the way people think about applications and deployments. It provides a simple way to run and distribute Linux containers for a variety of use cases, from lightweight virtual machines to complex distributed micro-services architectures. Kubernetes is an open source project to manage a cluster of Linux containers as a single system, managing and running Docker containers across multiple Docker hosts, offering co-location of containers, service discovery and replication control. It was started by Google and now it is supported by Microsoft, RedHat, IBM and Docker Inc amongst others. Jenkins Continuous Integration environment can be dynamically scaled by using the Kubernetes and Docker plugins, using containers to run slaves and jobs, and also isolate job execution.
NGINX Plus PLATFORM For Flawless Application DeliveryAshnikbiz
Flawless Application Delivery using Nginx Plus
By leveraging these latest features:
• Support for HTTP/2 standard
• Thread pools and socket sharding and how it can help improve performance
• NTLM support and new TCP security enhancements
• Advanced NGINX Plus monitoring, management and visibility of health & load checks
Catch this exclusive Google Hangout live!
November 4th, 2015 | 2.00-2.30PM IST | 4.30-5.00PM SGT
About the speaker: Sandeep Khuperkar, Director and CTO at Ashnik will be heading this session. He is an author, enthusiast and community moderator at opensource.com. He is also member of Open Source Initiative, Linux Foundation and Open Source Consortium Of India.
Building Data Integration and Transformations using PentahoAshnikbiz
This presentation will showcase the Data Integration capabilities of Pentaho which helps in building data transformations, through two demonstrations:
- How to build your first transformation to extract, transform and blend the data from various data sources
- How to add additional steps and filters to your transformation
Scaling Jenkins with Docker: Swarm, Kubernetes or Mesos?Carlos Sanchez
The Jenkins platform can be dynamically scaled by using several Docker cluster and orchestration platforms, using containers to run slaves and jobs and also isolating job execution. But which cluster technology should be used? Docker Swarm? Apache Mesos? Kubernetes? How do they compare? All of them can be used to dynamically run jobs inside containers. This talk will cover these main container clusters, outlining the pros and cons of each, the current state of the art of the technologies and Jenkins support.
EDW CENIPA is a opensource project designed to enable analysis of aeronautical incidentes that occured in the brazilian civil aviation. The project uses techniques and BI tools that explore innovative low-cost technologies. Historically, Business Intelligence platforms are expensive and impracticable for small projects. BI projects require specialized skills and high development costs. This work aims to break this barrier.
Load Balancing Apps in Docker Swarm with NGINXNGINX, Inc.
On-demand webinar recording: http://bit.ly/2mRjk2g
Docker and other container technologies continue to gain in popularity. We recently surveyed the broad community of NGINX and NGINX Plus users and found that two-thirds of organizations are either investigating containers, using them in development, or using them in production. Why? Because abstracting your applications from the underlying infrastructure makes developing, distributing, and running software simpler, faster, and more robust than ever before.
But when you move from running your app in a development environment to deploying containers in production, you face new challenges – such as how to effectively run and scale an application across multiple hosts with the performance and uptime that your customers demand.
The latest Docker release, 1.12, supports multihost container orchestration, which simplifies deployment and management of containers across a cluster of Docker hosts. In a complex environment like this, load balancing plays an essential part in delivering your container-based application with reliability and high performance.
Join us in this webinar to learn:
* The basic built-in load balancing options available in Docker Swarm Mode
* The pros and cons of moving to an advanced load balancer like NGINX
* How to integrate NGINX and NGINX Plus with Swarm Mode to provide an advanced load-balancing solution for a cluster with orchestration
* How to scale your Docker-based application with Swarm Mode and NGINX Plus
This session was held by Vladimir Brenner, Partner Account Manager, Disruptors & AI, Intel AI at the Dive into H2O: London training on June 17, 2019.
Please find the recording here: https://youtu.be/60o3eyG5OLM
Webinar: Open Source Business Intelligence IntroSpagoWorld
The presentation supported the webinar delivered by Stefano Scamuzzo, SpagoBI International Manager, on 22nd December 2010 within SpagoWorld Webinar Center. http://www.spagoworld.org/
Scalable ETL with Talend and Hadoop, Cédric Carbone, Talend.OW2
ETL is the process of extracting data from one location, transforming it, and loading it into a different location, often for the purposes of collection and analysis. As Hadoop becomes a common technology for sophisticated analysis and transformation of petabytes of structured and unstructured data, the task of moving data in and out efficiently becomes more important and writing transformation jobs becomes more complicated. Talend provides a way to build and automate complex ETL jobs for migration, synchronization, or warehousing tasks. Using Talend's Hadoop capabilities allows users to easily move data between Hadoop and a number of external data locations using over 450 connectors. Also, Talend can simplify the creation of MapReduce transformations by offering a graphical interface to Hive, Pig, and HDFS. In this talk, Cédric Carbone will discuss how to use Talend to move large amounts of data in and out of Hadoop and easily perform transformation tasks in a scalable way.
A modern, flexible approach to Hadoop implementation incorporating innovation...DataWorks Summit
A modern, flexible approach to Hadoop implementation incorporating innovations from HP Haven
Jeff Veis
Vice President
HP Software Big Data
Gilles Noisette
Master Solution Architect
HP EMEA Big Data CoE
Big data is an opportunity for communications service providers (CSPs) to create the intelligence for operating their infrastructures more efficiently, to analyze the success of their services, and to create a better personal experience for their customers.
CSP Top executives, Network and IT managers and Marketing, are eager to exploit the large amount of information to achieve better business decisions. They expect their Chief Technical Officer to provide end-to-end analytic solutions based on the data available in their IT and network infrastructure.
This presentation analyzes the complete value chain that can transform CSPs’ data to knowledge. It covers the sources of information, the data collection tools, the analytic platforms providing quick data access, and finally the business intelligence use cases with the presentation and visualization of the results and predictions.
Any data source becomes an SQL Query with all the power of
Apache Spark. Querona is a virtual database that seamlessly connects any data source with Power BI, TARGIT, Qlik, Tableau, Microsoft Excel or others. It lets you build your
own universal data model and share it among reporting tools.
Querona does not create another copy of your data, unless you want to accelerate your reports and use build-in execution engine created for purpose of Big Data analytics. Just write standard SQL query and let Querona consolidate data on the fly, use one of execution engines and accelerate processing no matter what kind and how many sources you have.
Powering Real-Time Big Data Analytics with a Next-Gen GPU DatabaseKinetica
Freed from the constraints of storage, network and memory, many big data analytics systems now are routinely revealing themselves to be compute bound. To compensate, big data analytic systems often result in wide horizontal sprawl (300-node Spark or NoSQL clusters are not unusual!)— to bring in enough compute for the task at hand. High system complexity and crushing operational costs often result. As the world shifts from physical to virtual assets and methods of engagement, there is an increasing need for systems of intelligence to live alongside the more traditional systems of record and systems of analysis. New approaches to data processing are required to support the real-time processing of data required to drive these systems of intelligence.
Join 451 Research and Kinetica to learn:
•An overview of the business and technical trends driving widespread interest in real-time analytics
•Why systems of analysis need to be transformed and augmented with systems of intelligence bringing new approaches to data processing
•How a new class of solution—a GPU-accelerated, scale out, in-memory database–can bring you orders of magnitude more compute power, significantly smaller hardware footprint, and unrivaled analytic capabilities.
•Hear how other companies in a variety of industries, such as financial services, entertainment, pharmaceutical, and oil and gas, benefit from augmenting their legacy systems with a modern analytics database.
A Comprehensive Approach to Building your Big Data - with Cisco, Hortonworks ...Hortonworks
Companies in every industry look for ways to explore new data types and large data sets that were previously too big to capture, store and process. They need to unlock insights from data such as clickstream, geo-location, sensor, server log, social, text and video data. However, becoming a data-first enterprise comes with many challenges.
Join this webinar organized by three leaders in their respective fields and learn from our experts how you can accelerate the implementation of a scalable, cost-efficient and robust Big Data solution. Cisco, Hortonworks and Red Hat will explore how new data sets can enrich existing analytic applications with new perspectives and insights and how they can help you drive the creation of innovative new apps that provide new value to your business.
Top 10 Data analytics tools to look for in 2021Mobcoder
This write-up has surrounded the top 10 tools used by data analysts, architects, scientists, and other professionals. Each tool has some specific feature that makes it an ideal fit for a specific task. So choose wisely depending on your business need, type of data, the volume of information, experience in analytical thinking.
Pentaho Big Data Analytics with Vertica and HadoopMark Kromer
Overview of the Pentaho Big Data Analytics Suite from the Pentaho + Vertica presentation at Big Data Techcon 2014 in Boston for the session called "The Ultimate Selfie | Picture Yourself with the Fastest Analytics on Hadoop with HP Vertica and Pentaho"
MoSync Cross Platform mobile app developmentUday Kothari
You can use MoSyc to develop cross platform mobile apps. Its free, very stable and easy to build mobile apps. Use MoSyc Reload to build prototype quickly and MoSync IDE to build native apps.
Cross platform mobile app development tools reviewUday Kothari
Presentation to discuss various Cross Platform Mobile Apps development tools. Rhodes, PhoneGap, Appcelerator, Sencha Touch, Xamarin, MoSync SDK and MoSync Reload.
Cross Compilation, VM, Mobile Web. Join our Meetup group for meeting great app developers, entrepreneurs, startups, students in Pune (Pune Mobile Apps Developers Meetup)
BI & Analytics in Action Using QlikViewUday Kothari
The QlikView Business Discovery platform delivers true self-service BI that empowers business users by driving innovative decision-making. It is one of the fastest growing BI products and has been recognized in the industry for its ease of use for business users, visualizing data with engaging, state-of-the-art graphics and ability to consolidate relevant data from multiple sources into a single application. Companies in India like Flipkart, Godrej, Canon, HDFC Life, Reliance life and shoppers stop are leveraging QlikView to empower their business users through QlikView.
This webinar will gives an overview of QlikView architecture, talks about why it is different and then, it will also take you through how to get started with using QlikView. For data analytics enthusiasts, this webinar presents simplest ways to learn QlikView. At the end of this webinar, participants will be equipped to install QlikView and create simple dashboards. In short this is a 'Fast track indtroduction to creating your first QlikView dashboard'.
The art technique of data visualizationUday Kothari
Decision making based on information has been the single most important objective of a data warehousing or big data pursuit. No matter how big, fast and varied data are generated and processed; decision makers are only concerned with the consumption of its end result – data visualization.
Data visualization simply means representing data in a visually appealing manner to enable understanding of the context in which we operate. Data visualization is a “moment of truth” that stems from a data management initiative. It is a very linear process of decision making; and hence, critical to its success. However, data visualizations also possess the potential to put an end to such initiatives; especially, when they are either heavily biased on just the design or contain information overload.
This webinar on the art and technique of data visualization focuses sharply on the one thing that matters most to qualify for effective data visualization: the truth that comes out from data. We have facilitated the discussion with the help of our 3D framework: Design, Discovery & Data.
After registering, you will receive a confirmation email containing information about joining the webinar.
Designing high performance datawarehouseUday Kothari
Just when the world of “Data 1.0” showed some signs of maturing; the “Outside In” driven demands seem to have already initiated some the disruptive changes to the data landscape. Parallel growth in volume, velocity and variety of data coupled with incessant war on finding newer insights and value from data has posed a Big Question: Is Your Data Warehouse Relevant?
In short, the surrounding changes happening real time is the new “Data 2.0”. It is characterized by feeding the ever hungry minds with sharper insights whether it is related to regulation, finance, corporate action, risk management or purely aimed at improving operational efficiencies. The source in this new “Data 2.0” has to be commensurate to the outside in demands from customers, regulators, stakeholders and business users; and hence, you would need a high relformance (relevance + performance) data warehouse which will be relevant to your business eco-system and will have the power to scale exponentially.
We starts this webinar by giving the audiences a sneak preview of what happened in the Data 1.0 world & which characteristics are shaping the new Data 2.0 world. It then delves deep on the challenges that growing data volumes have posed to the Data warehouse teams. It also presents the audiences some of the practical and proven methodologies to address these performance challenges. Finally, in the end it will highlight some of the thought provoking ways to turbo charge your data warehouse related initiatives by leveraging some of the newer technologies like Hadoop. Overall, the webinar will educate audiences with building high performance and relevant data warehouses which is capable of meeting the newer demands while significantly driving down the total cost of ownership.
Innovative Internet & Digital marketing Uday Kothari
This presentation was given on 24th Aug at MCCIA Pune event. It discusses how search engine works, SEO and SMO. How one can engage with web site visitors.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
PHP Frameworks: I want to break free (IPC Berlin 2024)
Business Intelligence and Big Data Analytics with Pentaho
1. Welcome to the webinar on
Business Intelligence and Big Data Analytics
with Pentaho
Presented by
&
www.compulinkacademy.com
www.ellicium.com
2. Contents
1
An Introduction to Pentaho
2
Overview of Pentaho technology stack
3
Pentaho ETL
4
Data Exploration using Pentaho
5
Big Data with Pentaho
6
Getting started with Pentaho
3. Welcome to Open source world
Open-source software is computer software with its source
code made available and licensed with a license in which
the copyright holder provides the rights to study, change
and distribute the software to anyone and for any purpose.
Open-source software is very often developed in a public,
collaborative manner.
Reporting
•
•
•
•
Analysis
•
•
•
•
Actuate BIRT
Jasper Reports
Pentaho
Open Reports
ETL Tools
•
•
•
•
JPivot
Mondrian/
Pentaho
PALO
You already use it!!!
•
•
• Jasper
• Pentaho
• SpagoBI
Napster
•
Amazon reviews,
•
YouTube
Data Mining /
Statistics
• Weka /
Pentaho
• R
BI Platforms
Clover ETL
Enhydra Octopus
Talend
Kettle / Pentaho
Linux
Databases
•
•
•
•
Derby
Ingres
MySQL
PostgreSQL
What it means for BI and analytics
A report by the Standish Group states that adoption of open-source software models has resulted
in savings of about $60 billion per year to consumers.
4. Welcome to Pentaho!!!!
•Commercial open source alternative for business intelligence (BI) Founded
in 2004 by five founders
•Management - proven BI and open source veterans from Business Objects,
Cognos, Hyperion, JBoss, Oracle, Red Hat, SAS
• Pioneer in Commercial open source BI Large reference able customer
base, wide range of BI/DW deployments !
•It offers a suite of open source Business Intelligence (BI) products called
Pentaho Business Analytics providing data integration, OLAP services,
reporting, dashboarding, data mining and ETL capabilities
6. What analysts are saying about Pentaho
Pentaho is the only open source company featured in Ovum's Ovum Decision Matrix
for Business Intelligence. "Pentaho is one of the few vendors that provide a direct
integration into Hadoop and NoSQL databases, allowing users to analyse and visualize
NoSQL data alongside traditional data sources"
Forrester recognized Pentaho as the sole "Strong Performer“. "Pentaho provides an
impressive Hadoop data integration tool." Pentaho was cited for its rich functionality
and extensive integration with Apache Hadoop, and for providing certified integration
with distributions from Cloudera, EMC Greenplum and Hortonworks.
Passionned's Business Intelligence Tools Survey highlighted the completeness of the
Pentaho product suite compared to other vendors, as well as Pentaho's significant
cost-saving by pricing products per deployment, not per-user. Pentaho earned
recommendation as a complete enterprise solution.
Pentaho was included in Gartner's Magic Quadrant for Business Intelligence Platforms.
The report, published, offers the analyst firm's insights on business intelligence
vendors who meet an inclusion threshold based on annual sales, capabilities, and
customer survey responses.
7. Pentaho Licensing
The current version of the Pentaho BI Platform will be distributed under
the terms of the GNU General Public License (GPL).
Under the GPL, if you intend to distribute GPL-licensed code to your
customers as part of other software you have created, you may, depending
on the software you have created, be required to GPL that code.
Companies that wish to distribute the Pentaho BI Platform have the option
of purchasing a commercial license from Pentaho Corporation. A
commercial license would exempt you from GPL obligations.
The GNU General Public License (GPL) is the most widely used free
software license, which guarantees end users the freedoms to use, study,
share and modify the software. Derived works can only be distributed
under the same license terms.
11. Delivering Value in Different Deployment Models
Coexistence with traditional proprietary BI
•Minimize risk/exposure with consolidated vendors
•Prove technology and services internally
•Explore the relationship benefits of a transparent model without
software lock-in
Co-deployment with traditional proprietary BI
•Leverage existing investments
•Pragmatically “use what works”
•Reduce overall TCO by incorporating commercial open source
Replacement of traditional proprietary BI
•Upgrade BI capabilities
•Reduce TCO
•Capitalize on the opportunity of a “disruption” (software upgrade,
license change, etc.) in your BI environment
13. Pentaho Kettle ETL
•Pentaho Data Integration (PDI, also called Kettle) is the component of Pentaho responsible
for the Extract, Transform and Load (ETL) processes. Though ETL tools are most frequently
used in data warehouses environments, PDI can also be used for other purposes:
•Migrating data between applications or databases
•Exporting data from databases to flat files
•Loading data massively into databases
•Data cleansing
•Integrating applications
15. Pentaho Kettle ETL
Spoon
• GUI that allows you to design transformations and jobs
•Transformations and Jobs can describe themselves using an XML file or
can be put in a Kettle database repository.
•Spoon is available as executable script and batch file to make use of tool
in heterogeneous environment.
Pan
•A program to execute transformations designed by Spoon in XML or database
repository.
•Transformations are scheduled in batch mode to be run automatically at regular
intervals
Carte
•Simple web server to execute transformations and jobs remotely.
•Accept an XML that contains transformation to execute and the execution
configuration.
•Allows to remotely monitor, start and stop the transformations and jobs
20. Pentaho Dashboards
What is CDE ?
* CDE is one of the plug-in for Pentaho BI Server, contributed and maintained by Pentaho Partner
webdetails.
* We create dashboards using this tool.
* Community Dashboard Editor (CDE) was born to simplify the creation, edition and rendering
processes of the Dashboards.
* CDE is a very powerful and complete tool, combining front end with data sources and custom
components in a seamless way.
CDE has 3 major components
They are.
* Layout
* Components
* Data Sources.CDE has developed based on MVC-2 architecture of Advanced Java
23. Main Big Data Technologies
Hadoop
NoSQL Databases
Analytic RDBMS
•
•
•
•
•
Low cost, reliable scaleout architecture
Distributed computing
Proven success in
Fortune 500 companies
Exploding interest
Hadoop
•
•
Huge horizontal scaling
and high availability
Highly optimized for
retrieval and appending
Types
•
•
•
Document stores
Key Value stores
Graph databases
NoSQL Databases
•
Optimized for bulk-load
and fast aggregate query
workloads
Types
•
•
•
Column-oriented
MPP
In-memory
Analytic Databases
24. What makes Pentaho different for big data
Ingestion / Manipulation
/ Integration
Scheduling
Modeling
Would you rather do this?
… OR THIS?
25. Pentaho Big Data Integration
Pentaho is integrated with Hadoop at many levels
•Traditional ETL - Graphical designer to visually build transformations that read and write data
in Hadoop from/to anywhere and transform the data on the way. No coding required
•HBase Read/Write
•Hive, Hive2 SQL Query and Write
•Impala SQL Query and Write
•Support for Avro file format and snappy compression
•Data Orchestration - Graphical designer to visually build and schedule jobs that orchestrate
processing, data movement and most aspects of operationalizing your data preparation.
•HDFS Copy files
•Map Reduce Job Execution
•Pig Script Execution
•Amazon EMR Job Execution
•Oozie integration
•Sqoop Import/Export
•Pentaho MapReduce Execution
26. Pentaho Big Data Integration
•Pentaho MapReduce - Graphical designer to visually build MapReduce jobs and run
them in cluster. With a simple, point-and-click alternative to writing Hadoop
MapReduce programs in Java or Pig, Pentaho exposes a familiar ETL-style user
interface.
•Traditional Reporting - All data sources supported above can be used directly or
blended with other data to drive our pixel perfect reporting engine. The reports can
be secured, parameterized and published to the web. The reports can be mashed up
with other pentaho visualizations to create dashboards.
•Web Based Interactive Reporting - Pentaho's Metadata layer leverages data stored in
Hive, Hive2 and Impala for WYSIWYG, interactive, self-service reporting.
•Pentaho Analyzer - Leverage your data stored Impala or Hive2 for interactive visual
analysis with drill through, lasso filtering, zooming, and attribute highlighting for
greater insight.
28. Getting started with Pentaho
•Download Pentaho from http://community.pentaho.com/
•Download MySQL from
http://dev.mysql.com/downloads/mysql/
• Download CDE from www.webdetails.pt/ctools/cde.html
Read installation instructions from following blogs:
•http://pentaho-bi-suite.blogspot.in/2013/04/installation-ofpentaho-bi-server.html
• We have a Pentaho installation guide available. Please request
for guide at: info@ellicium.com
29. Thank you !!!
Contact us for customized Pentaho
training on
info@compulinkacademy.com
info@ellicium.com
Or Call Sameer on +91-8793334411