This is an intro talk about data visualization, focused on showing few basic concepts on data visualization.
Presented during 1st Machine Learning Meetup - Porto Alegre - 1st June 2016
Presenter - Roberto Silveira
This presentation is part of LearnItFirst's SQL Server 2012: A Comprehensive Introduction course. The video that contains this presentation can be watched here:
https://www.youtube.com/watch?v=TJIFMS_lKys
There are many different report types that we can work with within Reporting Services, and this video will cover many of the common types. We will not cover every reporting type, but rather give you an overview of the visual effects that a particular element will add to a report, and when you should use them.
Highlights from this slideshow:
- Walking through the different parts of a dashboard report
- What are Sparklines and what are they good for?
- What are line charts and what do they offer that Sparklines can't?
- Bar charts are better for what?
- What are gauges and indicators mainly use for?
- What type of chart would you use to show proportion?
Open Data of the past
Open History Map (https://www.openhistorymap.org) is a non-profit based in Bologna, Italy, aiming at the creation of an open map of the past, reconciling and interconnecting data from Digital Humanities, Public History and Digital Archaeology sources. The presentation will cover the philosophy, the technology and the libraries used in this endeavor, explaining how users can reuse, and integrate the data from the platform into their applications or visualizations.
This presentation is part of LearnItFirst's SQL Server 2012: A Comprehensive Introduction course. The video that contains this presentation can be watched here:
https://www.youtube.com/watch?v=TJIFMS_lKys
There are many different report types that we can work with within Reporting Services, and this video will cover many of the common types. We will not cover every reporting type, but rather give you an overview of the visual effects that a particular element will add to a report, and when you should use them.
Highlights from this slideshow:
- Walking through the different parts of a dashboard report
- What are Sparklines and what are they good for?
- What are line charts and what do they offer that Sparklines can't?
- Bar charts are better for what?
- What are gauges and indicators mainly use for?
- What type of chart would you use to show proportion?
Open Data of the past
Open History Map (https://www.openhistorymap.org) is a non-profit based in Bologna, Italy, aiming at the creation of an open map of the past, reconciling and interconnecting data from Digital Humanities, Public History and Digital Archaeology sources. The presentation will cover the philosophy, the technology and the libraries used in this endeavor, explaining how users can reuse, and integrate the data from the platform into their applications or visualizations.
Euro30 2019 - Benchmarking tree approaches on street dataFabion Kauker
By examining the use of algorithms to solve the Prize Collecting Steiner Tree (PCST) problem we consider the facets which determine effectiveness. Specifically, by measuring a number of solution approaches and comparing them based on metrics. In order to understand the solution approach we must asses why it is useful. Our goal is to determine the effectiveness of Mixed Integer Programming (MIP) and heuristic methods. Utilizing freely available street and address data a base graph representation is created and then computed on. Such that a tree connects every address utilizing the minimum total length of edges from the street network. This is the basis of many approaches used to solve infrastructure problems including telecommunications network design and costing. The analysis is conducted on methods developed by Hegde et al. 2015, Ljubić et al. 2006, and Teitz et al. 1963. We present a data processing architecture, as well as a concise set of results and a framework for assessing the facets and trade-offs for a given approach. In this case the heuristic approaches are proven to have advantages in the simplistic case but fail when more complex requirements are added. This is where the MIP approach is able to capitalize, whilst detrimentally limiting the flexibility due to the strictness and specificity in modelling.
The underlying purpose of RasterFrames™ is to allow data scientists and software developers to process and analyze geospatial-temporal raster data with the same flexibility and ease as any other Spark Catalyst data type. At its core is a user-defined type (UDT) called TileUDT, which encodes a GeoTrellis Tile in a form the Spark Catalyst engine can process. Furthermore, we extend the definition of a DataFrame to encompass some additional invariants, allowing for geospatial operations within and between RasterFrames to occur, while still maintaining necessary geo-referencing constructs.
CEDAR & PRELIDA Preservation of Linked Socio-Historical DataPRELIDA Project
by Albert Meroño, presented at the 3rd PRELIDA Consolidation and Dissemination Workshop, Riva, Italy, October, 17, 2014. More information about the workshop at: prelida.eu
ThinkDo Project is he first cutting-edge app that automated schedules for better project management. Its revolutionary math model and chart are a step beyond Gantt, PERT and Kanban pipeline.
The concept behind ThinkDo you can find at this presentation as a report at the Innovation Section of the Project Management Institute's World Conference (PMICOS-2011).
Machine Learning in 5 Minutes— ClassificationBrian Lange
Slides from a lightning talk on classification methods, originally given at Open Source Open Mic Chicago 01/2016. Yes, I know I left things you. You try covering this in 5 minutes.
Biourbanism focuses on the urban organism, considering it as a hypercomplex system, according to its internal and external dynamics and their mutual interactions.
The urban body is composed of several interconnected layers of dynamic structure, all influencing each other in a non-linear manner. This interaction results in emergent properties, which are not predictable except through a dynamical analysis of the connected whole. This approach therefore links Biourbanism to the Life Sciences, and to Integrated Systems Sciences like Statistical Mechanics, Thermodynamics, Operations Research, and Ecology in an essential manner. The similarity of approaches lies not only in the common methodology, but also in the content of the results (hence the prefix “Bio”), because the city represents the living environment of the human species. Biourbanism recognizes “optimal forms” defined at different scales (from the purely physiological up to the ecological levels) which, through morphogenetic processes, guarantee an optimum of systemic efficiency and for the quality of life of the inhabitants. A design that does not follow these laws produces anti-natural, hostile environments, which do not fit into an individual’s evolution, and thus fail to enhance life in any way.
Biourbanism acts in the real world by applying a participative and helping methodology. It verifies results inter-subjectively (as people express their physical and emotional wellbeing through feedback) as well as objectively (via experimental measures of physiological, social, and economic reactions).
The aim of Biourbanism is to make a scientific contribution towards: (i) the development and implementation of the premises of Deep Ecology (Bateson) on social-environmental grounds; (ii) the identification and actualization of environmental enhancement according to the natural needs of human beings and the ecosystem in which they live; (iii) managing the transition of the fossil fuel economy towards a new organizational model of civilization; and (iv) deepening the organic interaction between cultural and physical factors in urban reality (as, for example, the geometry of social action, fluxes and networks study, etc.).
Interactive visualization and exploration of network data with gephiBernhard Rieder
Presentation for a workshop given at the Centre for Interdisciplinary Methodologies at Warwick University on May 9 2013. Focuses on conceptual and historical questions. Comments, references, and explanations are in the notes.
Euro30 2019 - Benchmarking tree approaches on street dataFabion Kauker
By examining the use of algorithms to solve the Prize Collecting Steiner Tree (PCST) problem we consider the facets which determine effectiveness. Specifically, by measuring a number of solution approaches and comparing them based on metrics. In order to understand the solution approach we must asses why it is useful. Our goal is to determine the effectiveness of Mixed Integer Programming (MIP) and heuristic methods. Utilizing freely available street and address data a base graph representation is created and then computed on. Such that a tree connects every address utilizing the minimum total length of edges from the street network. This is the basis of many approaches used to solve infrastructure problems including telecommunications network design and costing. The analysis is conducted on methods developed by Hegde et al. 2015, Ljubić et al. 2006, and Teitz et al. 1963. We present a data processing architecture, as well as a concise set of results and a framework for assessing the facets and trade-offs for a given approach. In this case the heuristic approaches are proven to have advantages in the simplistic case but fail when more complex requirements are added. This is where the MIP approach is able to capitalize, whilst detrimentally limiting the flexibility due to the strictness and specificity in modelling.
The underlying purpose of RasterFrames™ is to allow data scientists and software developers to process and analyze geospatial-temporal raster data with the same flexibility and ease as any other Spark Catalyst data type. At its core is a user-defined type (UDT) called TileUDT, which encodes a GeoTrellis Tile in a form the Spark Catalyst engine can process. Furthermore, we extend the definition of a DataFrame to encompass some additional invariants, allowing for geospatial operations within and between RasterFrames to occur, while still maintaining necessary geo-referencing constructs.
CEDAR & PRELIDA Preservation of Linked Socio-Historical DataPRELIDA Project
by Albert Meroño, presented at the 3rd PRELIDA Consolidation and Dissemination Workshop, Riva, Italy, October, 17, 2014. More information about the workshop at: prelida.eu
ThinkDo Project is he first cutting-edge app that automated schedules for better project management. Its revolutionary math model and chart are a step beyond Gantt, PERT and Kanban pipeline.
The concept behind ThinkDo you can find at this presentation as a report at the Innovation Section of the Project Management Institute's World Conference (PMICOS-2011).
Machine Learning in 5 Minutes— ClassificationBrian Lange
Slides from a lightning talk on classification methods, originally given at Open Source Open Mic Chicago 01/2016. Yes, I know I left things you. You try covering this in 5 minutes.
Biourbanism focuses on the urban organism, considering it as a hypercomplex system, according to its internal and external dynamics and their mutual interactions.
The urban body is composed of several interconnected layers of dynamic structure, all influencing each other in a non-linear manner. This interaction results in emergent properties, which are not predictable except through a dynamical analysis of the connected whole. This approach therefore links Biourbanism to the Life Sciences, and to Integrated Systems Sciences like Statistical Mechanics, Thermodynamics, Operations Research, and Ecology in an essential manner. The similarity of approaches lies not only in the common methodology, but also in the content of the results (hence the prefix “Bio”), because the city represents the living environment of the human species. Biourbanism recognizes “optimal forms” defined at different scales (from the purely physiological up to the ecological levels) which, through morphogenetic processes, guarantee an optimum of systemic efficiency and for the quality of life of the inhabitants. A design that does not follow these laws produces anti-natural, hostile environments, which do not fit into an individual’s evolution, and thus fail to enhance life in any way.
Biourbanism acts in the real world by applying a participative and helping methodology. It verifies results inter-subjectively (as people express their physical and emotional wellbeing through feedback) as well as objectively (via experimental measures of physiological, social, and economic reactions).
The aim of Biourbanism is to make a scientific contribution towards: (i) the development and implementation of the premises of Deep Ecology (Bateson) on social-environmental grounds; (ii) the identification and actualization of environmental enhancement according to the natural needs of human beings and the ecosystem in which they live; (iii) managing the transition of the fossil fuel economy towards a new organizational model of civilization; and (iv) deepening the organic interaction between cultural and physical factors in urban reality (as, for example, the geometry of social action, fluxes and networks study, etc.).
Interactive visualization and exploration of network data with gephiBernhard Rieder
Presentation for a workshop given at the Centre for Interdisciplinary Methodologies at Warwick University on May 9 2013. Focuses on conceptual and historical questions. Comments, references, and explanations are in the notes.
Data driven storytelling tips from an iron viz champion ryan sleeperRyan Sleeper
"Data-Driven Storytelling: Tip from an Iron Viz Champion" is a presentation by Ryan Sleeper, Manager of Data Visualization & Analysis at Evolytics. Ryan Sleeper is the 2013 Tableau Iron Viz Champion. This presentation was presented at the 2014 Tableau Conference in Seattle, Washington on 9/9/2014. The presentation offers fifteen actionable tips for improving your data visualization. For more information, visit Evolytics.com/15tips
MachineLearning for dummies with Python
Have you heard that Machine Learning is the next big thing?
Are you a dummy in terms of Machine Learning, and think that is a topic for mathematics with black-magic skills?
If your response to both questions is 'Yes', we are in the same position.
Still, thanks to the Web, Python and OpenSource libraries, we can overcome this situation and do some interesting stuff with Machine Learning.
http://imatge-upc.github.io/telecombcn-2016-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of big annotated data and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which had been addressed until now with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or text captioning.
This is the slide that Terry. T. Um gave a presentation at Kookmin University in 22 June, 2014. Feel free to share it and please let me know if there is some misconception or something.
(http://t-robotics.blogspot.com)
(http://terryum.io)
How to Become a Data Scientist
SF Data Science Meetup, June 30, 2014
Video of this talk is available here: https://www.youtube.com/watch?v=c52IOlnPw08
More information at: http://www.zipfianacademy.com
Zipfian Academy @ Crowdflower
List of top Machine Learning algorithms are making headway in the world of data science. Explained here are the top 10 of these machine learning algorithms - https://www.dezyre.com/article/top-10-machine-learning-algorithms/202
While the Rio 2016 Olympics are winding down and the final medals are being handed out, we thought we would share a bit of work that was done recently by Rik Van Bruggen to explore a really interesting dataset in Neo4j.
Based on an original public dataset by the UK newspaper The Guardian, Rik completed the medallist dataset to contain over 30,000 Olympians between 1896 and 2012. He created a graph model, loaded the data, and wrote a bunch of example queries that yielded some very interesting results. Join us for this 30 minute webinar where we’ll take you through this great Olympian graph and take the data for a spin yourself afterwards.
Every year the financial industry loses billions because of fraud while in the meantime fraudsters are coming up with more and more sophisticated patterns.
Financial institutions have to find the balance between fraud protection and negative customer experience. Fraudsters bury their patterns in lots of data, but the traditional technologies are not designed to detect fraud in real-time or to see patterns beyond the individual account.
Analyzing relations with graph databases helps uncover these larger complex patterns and speeds up suspicious behavior identification.
Furthermore, graph databases enable fast and effective real-time link queries and passing context to machine learning models.
The earlier fraud pattern or network is identified, the faster the activity is blocked. As a result, losses and fines are minimized.
A brief intro about me, my past experience, and a small history of data analytics and data science and the latest trends in data science such as deep learning, graph analysis, distributed computing with spark and streaming computing.
Let me say something obvious. Estimating is hard. So how come that we at Lunar Logic barely remember times when it was an issue?
The standard approach to estimation is based on expert guesses. This means that subconsciously we make ourselves vulnerable to cognitive biases which render our estimates too optimistic and unreliable. That neither is fun for development teams nor provides much value to clients.
We can employ a different strategy. Instead of relying on vulnerable human memory we use the actual historical data and run statistical simulations to produce a forecast. Not only is the outcome significantly better than what expert guess provides but also it requires less work from development teams.
We will share a story how our approach to estimation at Lunar Logic evolved from a common scenario to what we call statistical forecasting. We’ll share the theory behind, show how we do it and present some real data to back up the story. After all, once we understand why estimation is hard we can easily improve how we do it.
Map Reduce in Hazelcast - Hazelcast User Group London VersionChristoph Engelbert
A quick walktrough map-reduce and introduction into the Hazelcast implementation of Google's Whitepaper given at the Hazelcast User Group London on 11th June 2014
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.