Content reuse is one of the key reasons for converting to a single source. But how do you know what the reuse potential really is? Because converting content formats is a major investment, you need to base your business case on real data.
You can find the data with these steps:
• Observation
• Comparison
• Analysis
This session gives you a methodology to learn how to determine the true opportunity for content reuse and quantify it for your business case.
This presentation was given at Information Development World on October 1, 2015.
Module 1 introduction to machine learningSara Hooker
We believe in building technical capacity all over the world.
We are building and teaching an accessible introduction to machine learning for students passionate about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our work, visit www.deltanalytics.org
Module 1 introduction to machine learningSara Hooker
We believe in building technical capacity all over the world.
We are building and teaching an accessible introduction to machine learning for students passionate about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our work, visit www.deltanalytics.org
As a part of my Information Retrieval module in the University of East Anglia, we had to build classifier to detect deceptive review spam. Review spam was described by Nitin Jindal as follows: "It is now a common practice for e-commerce Web sites to enable their customers to write reviews of products that they have purchased. Such reviews provide valuable sources of information on these products .. Unfortunately, this importance of reviews also gives good incentive for spam, which contains false positive or malicious negative opinions". This is my poster presentation where I implemented 3 classification algorithms using Python, as well as feature selection and preprocessor modules.
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyMarina Santini
Definition of Machine Learning
Type of Machine Learning:
Classification
Regression
Supervised Learning
Unsupervised Learning
Reinforcement Learning
Supervised Learning:
Supervised Classification
Training set
Hypothesis class
Empirical error
Margin
Noise
Inductive bias
Generalization
Model assessment
Cross-Validation
Classification in NLP
Types of Classification
Module 9: Natural Language Processing Part 2Sara Hooker
Delta Analytics is a 501(c)3 non-profit in the Bay Area. We believe that data is powerful, and that anybody should be able to harness it for change. Our teaching fellows partner with schools and organizations worldwide to work with students excited about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our mission or provide feedback, take a look at www.deltanalytics.org. If you would like to use this material to further our mission of improving access to machine learning. Education please reach out to inquiry@deltanalytics.org .
Delta Analytics is a 501(c)3 non-profit in the Bay Area. We believe that data is powerful, and that anybody should be able to harness it for change. Our teaching fellows partner with schools and organizations worldwide to work with students excited about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our mission or provide feedback, take a look at www.deltanalytics.org.
Valencian Summer School 2015
Day 1
Lecture 3
Ensembles of Decision Trees
Gonzalo Martínez (UAM)
https://bigml.com/events/valencian-summer-school-in-machine-learning-2015
As a part of my Information Retrieval module in the University of East Anglia, we had to build classifier to detect deceptive review spam. Review spam was described by Nitin Jindal as follows: "It is now a common practice for e-commerce Web sites to enable their customers to write reviews of products that they have purchased. Such reviews provide valuable sources of information on these products .. Unfortunately, this importance of reviews also gives good incentive for spam, which contains false positive or malicious negative opinions". This is my poster presentation where I implemented 3 classification algorithms using Python, as well as feature selection and preprocessor modules.
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyMarina Santini
Definition of Machine Learning
Type of Machine Learning:
Classification
Regression
Supervised Learning
Unsupervised Learning
Reinforcement Learning
Supervised Learning:
Supervised Classification
Training set
Hypothesis class
Empirical error
Margin
Noise
Inductive bias
Generalization
Model assessment
Cross-Validation
Classification in NLP
Types of Classification
Module 9: Natural Language Processing Part 2Sara Hooker
Delta Analytics is a 501(c)3 non-profit in the Bay Area. We believe that data is powerful, and that anybody should be able to harness it for change. Our teaching fellows partner with schools and organizations worldwide to work with students excited about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our mission or provide feedback, take a look at www.deltanalytics.org. If you would like to use this material to further our mission of improving access to machine learning. Education please reach out to inquiry@deltanalytics.org .
Delta Analytics is a 501(c)3 non-profit in the Bay Area. We believe that data is powerful, and that anybody should be able to harness it for change. Our teaching fellows partner with schools and organizations worldwide to work with students excited about the power of data to do good.
Welcome to the course! These modules will teach you the fundamental building blocks and the theory necessary to be a responsible machine learning practitioner in your own community. Each module focuses on accessible examples designed to teach you about good practices and the powerful (yet surprisingly simple) algorithms we use to model data.
To learn more about our mission or provide feedback, take a look at www.deltanalytics.org.
Valencian Summer School 2015
Day 1
Lecture 3
Ensembles of Decision Trees
Gonzalo Martínez (UAM)
https://bigml.com/events/valencian-summer-school-in-machine-learning-2015
Example Mapping provides a structured approach to help teams tease out the essential business rules and examples that clarify a user story and improve shared understanding of story “doneness”
Recommending Scientific Papers: Investigating the User CurriculumJonathas Magalhães
In this paper, we propose a Personalized Paper Recommender System, a new user-paper based approach that takes into consideration the user academic curriculum vitae. To build the user profiles, we use a Brazilian academic platform called CV-Lattes. Furthermore, we examine some issues related to user profiling, such as (i) we define and compare different strategies to build and represent the user profiles, using terms and using concepts; (ii) we verify how much past information of a user is required to provide good recommendations; (iii) we compare our approaches with the state-of-art in paper recommendation using the CV-Lattes. To validate our strategies, we conduct a user study experiment involving 30 users in the Computer Science domain. Our results show that (i) our approaches outperform the state-of-art in CV-Lattes; (ii) concepts profiles are comparable with the terms profiles; (iii) analyzing the content of the past four years for terms profiles and five years for concepts profiles achieved the best results; and (iv) terms profiles provide better results but they are slower than concepts profiles, thus, if the system needs real time recommendations, concepts profiles are better.
We aim to create high-quality content. We really do. But, more-often-than-not, we fail. We understand that high-quality content must be clear, concise, and consistent in voice, tone, and terminology. We also know that it’s supposed to be easily findable, accessible, retrievable, and relevant those who need it—delivered when, where, and how they prefer it.
Crafting content that follows the rules (grammar, punctuation, linguistics) isn’t good enough. Our content also has to be helpful.
In this fast-paced talk, Scott Abel describes what it means to be helpful. You’ll discover how understanding the power of explanation
Presented November 27, 2018, at Quadrus Conference Center for Information Development World 2018.
Reimagining and reinventing customer support is expensive and hard. We hear that all the time. But it doesn’t have to be. In fact, if you do it “right” it can be fairly cheap and fun. In this session, we will look at using a Design Thinking approach to imagine new realities, create prototypes quickly and cheaply, and iterate on this to create a roadmap to your transformation.
Perhaps most important is that we will discuss some of the freely available tools that will help and guide you through the Design Thinking landscape. Unlike most speeches, we will give you specific, tangible baby steps to take once you get back to your own work lives.
Three Takeaways
1) Understand the power of Design Thinking
2) Imagine what Design Thinking can do for you
3) Know what tools are available and where to find them
Presented November 27, 2018, at Quadrus Conference Center for Information Development World 2018.
Part 1: Assessing the Current State: Needs Analysis and Information Gathering
Learn how to assess the current state of your technical support content by looking through the lens of content strategy and content engineering.
Traditionally, technical details about products and services were considered to be post-purchase content. Technical information — the stuff contained in owner’s manuals, user guides, and other instructional materials — was provided to consumers only after they purchased a product or service. However, that’s changing as companies recognize that prospects often search the web for technical content to make purchasing decisions.
Think of a technical resource center as an online, one-stop shop for information about your products and services. Over time, and done well, a technical resource center can help you grow your business by attracting prospects, while simultaneously working to support and build loyalty and trust with existing customers.
Presented November 27, 2018, at Quadrus Conference Center for Information Development World 2018.
Part 1: Assessing the Current State: Needs Analysis and Information Gathering
Learn how to assess the current state of your technical support content by looking through the lens of content strategy and content engineering.
Traditionally, technical details about products and services were considered to be post-purchase content. Technical information — the stuff contained in owner’s manuals, user guides, and other instructional materials — was provided to consumers only after they purchased a product or service. However, that’s changing as companies recognize that prospects often search the web for technical content to make purchasing decisions.
Think of a technical resource center as an online, one-stop shop for information about your products and services. Over time, and done well, a technical resource center can help you grow your business by attracting prospects, while simultaneously working to support and build loyalty and trust with existing customers.
Presented November 27, 2018, at Quadrus Conference Center for Information Development World 2018.
Moderated by Paul Perrotta with Panelists: Michael Rosinski of Astoria Software, Julie Newcome of Ultimate Software, Joe Gelb of Zoomin Software, Ray Gallon of The Transformation Society, Alex Masycheff of Intuillion, Ltd., and Anna N. Schlegel of Net App.
Budgets are tight. Times are lean. But you know you need to improve your Technical Resource Center. You could just hope it happens. Or, you could learn from the lessons of those who have gone before you. In this fast-paced panel discussion, Paul Perrotta will ask a panel of seasoned professionals for advice on how to pitch your ideas and secure funding. The panelists discuss the pitfalls to avoid, and they’ll share approaches, pro-tips, and advice to help you get what you need.
Ryan MacCarrigan’s keynote covers the growing role of Agile Development and Lean in the context of content development and delivery—where complex content is the “product” and the end goal is to shorten cycle times, eliminate waste, or improve measured business outcomes without sacrificing quality or accuracy.
The audience will learn:
How to structure strategic content development in a similar fashion to the Agile product development lifecycle
How the “Build-Measure-Learn” framework of Lean Startup fame can be applied to rapid content testing and delivery
How developing a Lean mindset can help content-driven organizations break down silos and “Fail Fast,” improving overall institutional knowledge.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
My colleague and I managed a team of 5 to 7 writers, using Agile processes to successfully overhaul a Help system for complex genetic sequencing software in just over six months. The approach uses 3 weekly sprints that gets each writer 1) analyzing and identifying gaps in existing content 2) writing and updating content, and then 3) peer editing and revising content. The sprints overlap so that every week each writer is actively writing, peer reviewing and editing content.
Facing deadlines for frequent quarterly releases, we used Excel spreadsheets and OneNote notebooks to record meeting notes, topic TOCs and assignments, rather than a more administrative intensive ticket-based system (such as JIRA). Writers, whose skill levels ranged from junior to senior, learned how to use the software through hour-long question-and-answer group sessions with SMEs.
Attend this session to learn how an agile writing process can help boost collaboration and increase comradeship amongst information developers; decrease the time spent with subject matter experts, and optimize content development.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
Designing the content experience revolves around the quality and the quantity of content. Answering questions like what kind of content, how much of it, and where should it be located are prime in a content professional’s mind. In her talk, Eeshita will discuss and share the pillars of content user experience — both quality and quantity. The attendees will learn techniques and processes to enable quality and monitor quantity of valuable content.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
Effective page design is often overlooked in the development of technical information. Studies have shown that the visual design of information has an immediate and lasting visceral impact on both credibility and usability. Good page design ensures that information is easy to find, read, understand, and remember. The science of human visual perception and attention provides a foundation for understanding traditional design elements and principles, and how they can be combined to ensure high-quality, effective information development.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
Before you’re ready to answer your customers’ questions, you need to ask some of your own: Where is the information leaking out of my content? How can I capture the human intelligence that went into writing the information in the first place? Where does my information development process have too much friction?
Find out how asking and answering these questions can help support your information developers as they create understandable information and actionable intelligence for both humans and bots. Identify your information leaks and learn how to stop them. Learn how to remove friction to stop wasting people’s time and to transform your information development and delivery process. Create the future by leveraging your intellectual property.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
Translation Commons is a nonprofit aiming to be an online one-stop community for all information relating to translation and localization. How do you organize content for an entire industry? How do you create a website structure that allows users to find the information they need, even when it’s a needle in a haystack?
Content planning or Information Architecture determines how information is displayed or accessed. For Translation Commons, planning took much longer than development and it was worth every second of it.
The audience will learn of various techniques and methodologies which will help them organize large sets of information.
Presented November 28, 2018, at Quadrus Conference Center for Information Development World 2018.
Workshop Part 2: Creating the Future State: Enterprise Content Creation, Structure and Distribution
Learn how to plan and implement a future state of enterprise content creation, structure, management, and delivery for a modern technical resource center.
Traditionally, technical details about products and services were considered to be post-purchase content. Technical information — the stuff contained in owner’s manuals, user guides, and other instructional materials — was provided to consumers only after they purchased a product or service. However, that’s changing as companies recognize that prospects often search the web for technical content to make purchasing decisions.
Think of a technical resource center as an online, one-stop shop for information about your products and services. Over time, and done well, a technical resource center can help you grow your business by attracting prospects, while simultaneously working to support and build loyalty and trust with existing customers.
Presented November 27-28, 2018, at Quadrus Conference Center for Information Development World 2018.
The NetApp content strategy team is driving a wide-reaching effort to simplify content creation, sharing, delivery, and content management systems reduction. Anna Schlegel will share how her team is leading an enterprise-wide effort to build a more connected content experience at NetApp with sponsorship at the CEO and SVP level across the entire organization.
In this presentation, attendees learn how to design a corporate content strategy, streamline the content ecosystem, obsolete unnecessary content, and formalize content governance. The key to this effort is selling the value proposition such as reduced cost, reduced complexity, and a better customer experience.
Anna will help you understand how to identify key players, navigate internal politics, and set the stage for content strategy success company-wide. You’ll leave knowing how to set the right goals, create teams, develop leaders, and utilize tracking methodologies.
Takeaways:
1) Setting the stage across the whole company for success
2) Identifying key players and navigating internal politics
3) Identifying the right goals, teams, leaders and tracking methodologies
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
More and more we’re seeing data that indicates we need increased focus on improving our clients’ experience with technical content. But, how do you know what to focus on and where to target first? Introducing a content analytics toolbox that we rolled out to our IBM Cloud content contributors. The toolbox includes a variety of tools that authors can use to identify what content to work on and how to measure content improvements.
This case study shows how we gained adoption of the use of the toolbox, as well as some concrete examples of the tooling and data.
Takeaways:
1) How do I know what to prioritize?
2) How can I prove my content is impacting the business?
3) What are others doing?
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Your customers demand reliable customer service and don’t have time to waste with poor self-serve support portals that contain less helpful content than they should. Many customer service agents suffer from a lack of good warrantied product information and spend a lot of time copying and pasting information from PDFs, emails, and websites. The technology they use seems to be in constant flux yet access to the information they need doesn’t seem to get that much better. There has got to be a better way.
What if there was a better medium for finding, using, and exchanging the highest value content in your organization? Microcontent is a basic building block of good product documentation. When it can be broken out of that content, it can be used in many ways to feed other documents, FAQs, emails, knowledge bases, and even chatbots. Microcontent is also an ideal level of granularity to contribute and curate new source information to be used across the enterprise. So what is it and how does it work to provide a better customer service experience? Attend this session to gain more insight into microcontent and how it can help.
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Mayo Clinic’s mobile app serves as both a resource for patients who depend on it for tasks like viewing lab results and making appointments and as a health engagement tool to keep the brand top-of-mind for anyone who might need Mayo Clinic services someday.
In this case study, find out how a Mayo Clinic team converted a huge library of health information to an engaging, mobile-friendly content experience. Learn how core content has been enhanced with hundreds of original visual and editorial pieces – built using a repeatable process geared for high-volume production. Explore how new features like mobile notifications and content search have addressed user needs while driving to new app downloads, now 1 million+ and counting.
Three Takeaways
1) How content can serve as an engagement tool while facilitating transactional tasks and resources
2) Simple curation and metadata strategies for delivering a seamless experience using multiple content sources
3) Tips on creating mobile-first content for short attention spans
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Do you see a problem that is so obvious that everyone should see it, but they don’t? Do you have great data about a pain point for your customers, but don’t know where to go with it? In this session, we’ll talk about project briefs — what they are, and how they can be an invaluable tool for building consensus and getting your stakeholders and teams on board.
In this session, you will learn:
1) How to pull together various data points into a cohesive project brief; 2) How to use a project brief to effectively present the problem/issue; 3) And, most importantly, why a project brief isn’t the right platform for solutioning.
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Building a conversational interface that people actually want to use can be tough a process. From understanding what users enter to providing logical responses, there’s a lot to create a successful experience. This presentation provides tips for designing conversational interfaces and the content that powers them. If you’re considering adding chatbots or voice-activated devices to your content delivery strategy, this session is right for you.
Takeaways
1) Tips for designing conversational interface
2) Tips for writing conversational content
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Most chatbots are rule-based. A rule defines that if certain keywords occur in the user’s question, a certain piece of content should be displayed. For example, “if the question includes words ‘replace’ and ‘battery,’ show the topic about replacing a battery.”
While this method is easy and relatively cheap to implement, it covers only simple use cases. It may work perfectly well if the amount of content is small and it’s not frequently updated. But what about a case when the procedure of replacing a battery is different for different product models? Or what if it’s different depending on the user’s role, and there are multiple possible roles and their combinations? You’d have to explicitly add rules for each variation and instruct the chatbot about the questions the user should be asked when information required for a precise and relevant answer is missing in the user’s question.
On top of that, every time you add new content, you have to manually add new rules. In the long run, rule-based chatbots are expensive and difficult to maintain, if the amount of content is significant and it’s frequently updated.
Another approach is to build a knowledge map of the subject domain which would automatically guide the chatbot about the questions the user should be asked, automatically identify semantic metadata of the content, and map the metadata to the knowledge map. This approach would make the chatbot smarter while reducing the maintenance efforts and costs.
In this session, Alex talks about both approaches and sees which approach works better in different use cases.
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
It can be difficult to onboard users to new and complex interfaces and workflows. New research shows that images and video enhance understanding and retention of complex information and tasks and can even increase productivity, but software often changes quickly and requires regular updates and localization.
How can we leverage the power of visual communication without having to constantly localize new visual content? Simplified User Interface (SUI) helps you create powerful and useful images to help your users better understand your content while extending its shelf life and often eliminating the need for localization.
Presented November 29, 2018, at Quadrus Conference Center for Information Development World 2018.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
GridMate - End to end testing is a critical piece to ensure quality and avoid...ThomasParaiso2
End to end testing is a critical piece to ensure quality and avoid regressions. In this session, we share our journey building an E2E testing pipeline for GridMate components (LWC and Aura) using Cypress, JSForce, FakerJS…
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...Neo4j
Leonard Jayamohan, Partner & Generative AI Lead, Deloitte
This keynote will reveal how Deloitte leverages Neo4j’s graph power for groundbreaking digital twin solutions, achieving a staggering 100x performance boost. Discover the essential role knowledge graphs play in successful generative AI implementations. Plus, get an exclusive look at an innovative Neo4j + Generative AI solution Deloitte is developing in-house.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
10. How we do it
This is the first example string
This is the fifth example string
• 3 characters would have to change to make them the same
• 32 characters in longest string
• Matching score = (39-3)/39 = 91% match
11. Example
• Text strings are then assigned to clusters
based on the matching score.
• The higher the matching score, the more
likely to obtain reuse, and the greater
business impact
16. Types of reuse
① Programmatic reuse
② Template reuse
③ Variable reuse
④ Content
repurposing
17. What do the numbers say?
• For you/your team
• Initial baseline
• Measurement tool
• Play what if? (different analytical views)
• Run it over and over again
Joan: The theme of this conference…. <make the connection with theme>
Amber: Customer, both internal and external, expect to find your content quickly regardless of the device they are using or the type of content. And that content had better be accurate and consistent.
Amber:
Joan: The opposite of the great customer experience is content confusion.
Once you have multiple instances the finding becomes harder.
Maintenance.
Where do I find it???? This is the hunt problem. How do I find every instance?
Update replace product names, dates, locations, specs—minor updates. Minor but time-consuming, difficult, error-prone, and not lasting.
Major content changes:
Content
Learning objectives
Cert questions
Manual = go and find it
Labs can change quickly—keep it aligned
Let’s say software changes—
Lab, where you use software, may need to change
Then cert questions may need to change
Joan:
You can address content confusion by:
Improving content consistency
Providing easy access to the content that each customer should be able to access
Making it easy to find the appropriate information at the right time
Other session at this conference will cover access and retrievability, but our session focuses on addressing inconsistency via reuse.
Amber
Based on our experience designing, developing, and delivering content, we know that that reuse doesn’t just happen: it takes planning and coordination.
While some teams think that they have lots of potential reuse, other teams think that they have very little opportunity to reuse content. Our recommendation is not to just think that you could reuse more content, but quantify the real reuse opportunity.
Our proposal is a 3-part strategy:
Observation
Comparison
Analysis
Amber:
Lots of folks start and end their reuse measurement with observation. This means that they usually compile a collection of content samples and then spot-check to see if there are redundant words/phrases in the different deliverables.
While this technique does give you a bit of knowledge about a small subset of text, it doesn’t really give a complete picture. It’s more like fishing for reuse…
The weaknesses of observation as the only measurement for potential reuse are:
It subjective: one person may search for one set text while another person searches for something completely different based upon what they think is important or useful
It’s not reproducible: the data depends upon who checks what phrases in which deliverables
Joan:
It’s incomplete: this manual method does not scale to give you a true count of the number of times a given word or phrase appears
It’s on exact matches only: searching for only exact matches doesn’t really show the opportunity for standardization; for many teams, the opportunity is in the near matches that could be standardized
If this is your only tool for estimating reuse opportunities, then you can never get real numbers for a business case to actually show the potential value of reuse
Instead of being the only step, we propose that observation is the first step and that instead of focusing on what words or phrases were repeated, we propose that the outcome of observation is identifying the content set for comparison and analysis
<<fishing—subjective—spot checking in thousands of deliverable—are you getting the right ones>>
Joan
By focusing on identifying the content for comparison and analysis, you’re not longer just fishing for information a huge ocean of text; instead you are defining the scope of your comparison.
We recommend that you use this exercise to select content set against a pattern or criteria:
Multiple deliverables of the same type across products
Multiple deliverables of different types that pertain to the same product
Multiple deliverables from different groups that apply to product families
???
Once you’ve identified the content set, you’re ready to do some thing with it.
Amber
The first thing you do is compare the words and phrases that comprise the content set. Not only are you looking for the items that are the exactly the same, but you want to know which of these things is not like the others…then you can analyze why it is different and whether it should be different.
The only way to do this type of comparison is with a tool or script. We’re going to show you the methodology behind a tool that we’ve been testing, but this is not a tool sales pitch.
Joan
First, get text in a format that it can be analyzed. Separate the text from the presentation layer. <<is this a transform? What can’t you work with images?>>
Then identify the content you want to compare—within a product line, across product lines, across functions?
Run the application to identify not only where reuse occurs, but also give the reuse a score: 100%, 90%
Benefit is you can run over and over—within product line across functions.
Joan
After you have identified the matches then you assign them into clusters based on their matching score. Once you have run the comparison and identified matches you can analyze the text to see where the reuse occurs and if text that is potential can be changed to be 100% matches.
Amber
After the tool compares the content clusters, it generates a spreadsheet with a lot of information.
Column A: lists the unique ID to each cluster.
Column B: lists the text in the cluster
Column C: lists the word count for the cluster
Column D: provides the file from which the cluster was extracted
Column E: lists a randomly assigned cluster number for 100% matches
Column F: identifies if this cluster is the root/base for the comparison #311 has the root cluster
Column G: identifies if this cluster is a compared cluster
Column H: identifies the number of words that were duplicates in the comparison between base cluster and the compared cluster
Column I-K: provides the same information for 90% matches
Column L-P: provides the same information for 80% matches
Column Q-T: provides the same information for the 70% matches
This means that cluster # 1050 was compared to and exactly matched all the other clusters that have the value of #304.
Amber
This type of comparison is incredibly detailed and generates thousands of comparisons. This data set included 10 fairly small online help systems and generated over 8K comparisons.
So, what do you do with all this data? Just having lots numbers doesn’t really help you quantify the opportunities for actual reuse. You need to analyze the data to understand what it really means.
Joan – can you speak to this one?
Do nothing to the slide.
The
Analyze the data from the comparison
Provide context by identifying the content types
<<making sense of the data>>
<<reverse engineering the data—where did it come from>>
pie chart on the left shows the percentage of paragraphs that match at different levels.Paragraphs that have a match of 60 percent or greater are considered to have attainable reuseand can be easily repurposed in the short term with low to medium effort. The pie chart on theright shows the breakdown of the 31 percent attainable reuse by functional group.
Joan?
Even at 70% will only see 26% reuse.
Applied report; graphs from Harmonic analysis
Analysis helps to quantify the opportunity—what’s the real opportunity
Walk through a blend of the comparison/analysis for Applied and Harmonic—comes from Andy too
<<screen cap of subset of data maybe with callouts>>
Break into 2 Below the line Total reuse %
Comparison: here’s what the comparison gives us—but wait there’s more
You’re not done yet—what does it mean—swimming in numbers, data doesn’t give you answers
This is more Andy
Amber
Now that we have a real understanding of the content that we can standardize, we can consider the types of reuse that may be possible.
100% match:
Programmatic reuse: all those clusters that are content labels and other standardized content should be automatically generated based upon the pattern and deliverableExample:
Template reuse: the content clusters that are actually consistent patterns, such as table headersExample: Alarm description content, the alarms are listed alphabetically and each “letter page” has a table with the same three column heads: alarm, explanation, recovery tips
Variable reuse: usually single words or small phrases that should be standardizedExample: products names, such as NMX
Content repurposing: reusing the same word or phrase in multiple contextsExample: “Start the Consolidated Alarms Application to view the alarms at the remote site.” this phrase appears 5 times; perhaps it would make sense to create a “library” of standard recovery options Example: “The ECM/ECMs in one or more SCG messages are missing and are not provisioned from NMX.” appears in two help systems. Maybe this is an opportunity to standardize OR NOT?
90% match:
You might find some items, such as Click OK versus Click OK button that are standardization opportunities.
Joan:
Joan?
Break into 2 Below the line Total reuse %
Walk through a blend of the comparison/analysis for Applied and Harmonic—comes from Andy too
<<screen cap of subset of data maybe with callouts>>
Amber
Many teams use the prospect of reuse as a driver for moving to DITA. However, before you ask for money to fund this initiative, quantify that real reuse opportunity.
Can you standardize your content? The comparison provides the data, but the analysis indicates what content can really be reused. In the example we showed today, many of the 90% matches can’t standardized and reuse via content references or other DITA mechanisms because the difference is between input and output. Further analysis by content type provides greater clarity
What processes would have to change to reuse content? If you’re trying to share content across teams, then you’ll to evaluate the readiness and feasibility of sharing/reusing content. Can you become more standard simply with some better processes?
What tools would help you reuse content? For many teams, reuse is limited until they are able to manage their content in a CCMS
Joan:
Give an example