This document discusses recipes for GDPR-compliant data science. It covers topics like data privacy, risks, ethics, compliance, and governance. On data privacy, it explains information privacy and regulations like GDPR and CCPA. On risks, it discusses risks in data like improper analytics and low data quality. On ethics, it discusses issues around automated decision-making, non-discrimination, and the right to explanation. On compliance, it advocates for monitoring and automated reporting. On governance, it notes challenges of constraints and advocates a bottom-up approach through monitoring data activities.
Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019...IDC4EU
This is the slide-deck of the community event held on November 14, 2019 in Brussels, titled "Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019". It includes the presentations given by the speakers.
"Towards Value-Centric Big Data" e-SIDES Workshop - "A win-win initiative for...e-SIDES.eu
The following presentation was given by Marina da Bormida, R&I Lawyer and Ethics Expert, during the e-SIDES workshop "Towards Value-Centric Big Data" held on April 2, 2019 in Brussels.
Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019...IDC4EU
This is the slide-deck of the community event held on November 14, 2019 in Brussels, titled "Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019". It includes the presentations given by the speakers.
"Towards Value-Centric Big Data" e-SIDES Workshop - "A win-win initiative for...e-SIDES.eu
The following presentation was given by Marina da Bormida, R&I Lawyer and Ethics Expert, during the e-SIDES workshop "Towards Value-Centric Big Data" held on April 2, 2019 in Brussels.
What is an Information Society
Why are Information Policies needed
What is an Information Policy
Elements of Information Policy
Who has Information Policies
E-Inclusion
Life Long Learning
E-Business strategies
Infrasture – physical (broadband/e-fibre)
Infrastructure – political / Legal and regulatory
Copyright, Intellectual Property, Data Protection, Freedom of Information
Regulation of Domain Name Spaces ( .ie)
E-government
Information Policy in Ireland
Ensuring Effective Information Security Management Information Classification...ijtsrd
This study is based on information security management in financial institutions from the perspective of information classification and access control. As objectives, the study set out to assess information classification practices in microfinance institutions and their effect on overall information security management, and to examine access control in microfinance institutions and how it impacts information security management. The study made use of the Information Security Theory by Horne, Ahmad and Maynard, and a sequential exploratory mixed method survey research design. As data collection instruments, a questionnaire and an interview guide were used, with validity and reliability guaranteed by subject experts, ISO IEC checklists, and Kuder Richardson formula 20 which realised a score of 0.81. Of the 30 managers and information security officers who participated in the study, a response rate of 100 was registered. To analyse data, descriptive statistics and thematic analysis were used. The findings portray loopholes in information classification and access control and thus in the information security management programme of participating institutions. Some recommendations put forth are the need to adopt information classification schedules with distinguished levels of sensitivity, drafting of access control policies, signing of non disclosure agreements and introduction of information security officers to ensure implementation and follow up. Rosemary M. Shafack | Awiye Sharon Serkwem "Ensuring Effective Information Security Management: Information Classification and Access Control Practices" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-1 , December 2020, URL: https://www.ijtsrd.com/papers/ijtsrd38122.pdf Paper URL : https://www.ijtsrd.com/management/other/38122/ensuring-effective-information-security-management-information-classification-and-access-control-practices/rosemary-m-shafack
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
Trivadis TechEvent 2016 Big Data Privacy and Security Fundamentals by Florian...Trivadis
In Big Data we focus on the 4 V's: Volume, Velocity, Varity and Veracity. But another important topic is often not in the focus: Privacy and Security. Yet as important and if not considered from the beginning it might put your Big Data project at risk. Learn about most important Privacy and Security fundamentals in Big Data, you should take into account in your next Big Data project.
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
A Dynamic Intelligent Policies Analysis Mechanism for Personal Data Processin...Konstantinos Demertzis
The evolution of the Internet of Things is significantly a
ected by legal restrictions imposed for personal data handling, such as the European General Data Protection Regulation (GDPR).
The main purpose of this regulation is to provide people in the digital age greater control over their personal data, with their freely given, specific, informed and unambiguous consent to collect and process the data concerning them. ADVOCATE is an advanced framework that fully complies with the requirements of GDPR, which, with the extensive use of blockchain and artificial intelligence technologies, aims to provide an environment that will support users in maintaining control of their personal data in the IoT ecosystem. This paper proposes and presents the Intelligent Policies Analysis Mechanism (IPAM) of the ADVOCATE framework, which, in an intelligent and fully automated manner, can identify conflicting rules or consents of the user, which may lead to the collection of personal data that can be used for profiling. In order to clearly identify and implement IPAM, the problem of recording user data from smart entertainment devices using Fuzzy Cognitive Maps (FCMs) was simulated. FCMs are an intelligent decision-making system that simulates the processes of a complex system, modeling the correlation base, knowing the behavioral and balance specialists of the system. Respectively, identifying conflicting rules that can lead to a profile, training is done using Extreme Learning Machines (ELMs), which are highly ecient neural systems of small and flexible architecture that can work optimally in complex environments.
Copy of OSTP RFI on Big Data and PrivacyMicah Altman
This document was originally published by OSTP here:
http://www.ofr.gov/(S(rfkilxaktjiadgtykwxaljqm))/OFRUpload/OFRData/2014-04660_PI.pdf
The original link is now broken, so this copy is provided for the transparency and commentary.
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
Artificial Intelligence - intersection with compliance. How AI principles work with compliance principles around data protection. AI and Compliance. AI - SYSC 13.7 - FCA Compliance. AI and regulation. AI and FCA regulation. AI and ICO regulation.
Extended discourse on the importance of data science governance for production ML and how GDPR can become the catalyst but also generate value for organizations!
What is an Information Society
Why are Information Policies needed
What is an Information Policy
Elements of Information Policy
Who has Information Policies
E-Inclusion
Life Long Learning
E-Business strategies
Infrasture – physical (broadband/e-fibre)
Infrastructure – political / Legal and regulatory
Copyright, Intellectual Property, Data Protection, Freedom of Information
Regulation of Domain Name Spaces ( .ie)
E-government
Information Policy in Ireland
Ensuring Effective Information Security Management Information Classification...ijtsrd
This study is based on information security management in financial institutions from the perspective of information classification and access control. As objectives, the study set out to assess information classification practices in microfinance institutions and their effect on overall information security management, and to examine access control in microfinance institutions and how it impacts information security management. The study made use of the Information Security Theory by Horne, Ahmad and Maynard, and a sequential exploratory mixed method survey research design. As data collection instruments, a questionnaire and an interview guide were used, with validity and reliability guaranteed by subject experts, ISO IEC checklists, and Kuder Richardson formula 20 which realised a score of 0.81. Of the 30 managers and information security officers who participated in the study, a response rate of 100 was registered. To analyse data, descriptive statistics and thematic analysis were used. The findings portray loopholes in information classification and access control and thus in the information security management programme of participating institutions. Some recommendations put forth are the need to adopt information classification schedules with distinguished levels of sensitivity, drafting of access control policies, signing of non disclosure agreements and introduction of information security officers to ensure implementation and follow up. Rosemary M. Shafack | Awiye Sharon Serkwem "Ensuring Effective Information Security Management: Information Classification and Access Control Practices" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-1 , December 2020, URL: https://www.ijtsrd.com/papers/ijtsrd38122.pdf Paper URL : https://www.ijtsrd.com/management/other/38122/ensuring-effective-information-security-management-information-classification-and-access-control-practices/rosemary-m-shafack
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
Trivadis TechEvent 2016 Big Data Privacy and Security Fundamentals by Florian...Trivadis
In Big Data we focus on the 4 V's: Volume, Velocity, Varity and Veracity. But another important topic is often not in the focus: Privacy and Security. Yet as important and if not considered from the beginning it might put your Big Data project at risk. Learn about most important Privacy and Security fundamentals in Big Data, you should take into account in your next Big Data project.
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
A Dynamic Intelligent Policies Analysis Mechanism for Personal Data Processin...Konstantinos Demertzis
The evolution of the Internet of Things is significantly a
ected by legal restrictions imposed for personal data handling, such as the European General Data Protection Regulation (GDPR).
The main purpose of this regulation is to provide people in the digital age greater control over their personal data, with their freely given, specific, informed and unambiguous consent to collect and process the data concerning them. ADVOCATE is an advanced framework that fully complies with the requirements of GDPR, which, with the extensive use of blockchain and artificial intelligence technologies, aims to provide an environment that will support users in maintaining control of their personal data in the IoT ecosystem. This paper proposes and presents the Intelligent Policies Analysis Mechanism (IPAM) of the ADVOCATE framework, which, in an intelligent and fully automated manner, can identify conflicting rules or consents of the user, which may lead to the collection of personal data that can be used for profiling. In order to clearly identify and implement IPAM, the problem of recording user data from smart entertainment devices using Fuzzy Cognitive Maps (FCMs) was simulated. FCMs are an intelligent decision-making system that simulates the processes of a complex system, modeling the correlation base, knowing the behavioral and balance specialists of the system. Respectively, identifying conflicting rules that can lead to a profile, training is done using Extreme Learning Machines (ELMs), which are highly ecient neural systems of small and flexible architecture that can work optimally in complex environments.
Copy of OSTP RFI on Big Data and PrivacyMicah Altman
This document was originally published by OSTP here:
http://www.ofr.gov/(S(rfkilxaktjiadgtykwxaljqm))/OFRUpload/OFRData/2014-04660_PI.pdf
The original link is now broken, so this copy is provided for the transparency and commentary.
The objective of this module is to gain an overview of the ethics surrounding big data and the legislation that governs it.
Upon completion of this module you will:
- Gain knowledge on how to recognize the necessity of regulating big data
- Obtain an understanding of the difference between privacy and data protection
- Understand the need to implement data protection actions into your own business
Artificial Intelligence - intersection with compliance. How AI principles work with compliance principles around data protection. AI and Compliance. AI - SYSC 13.7 - FCA Compliance. AI and regulation. AI and FCA regulation. AI and ICO regulation.
Extended discourse on the importance of data science governance for production ML and how GDPR can become the catalyst but also generate value for organizations!
Article started one year ago, obtains far more relevancy these days. Its meaning stays the same however: "Without laws and regulations would be chaos affecting our freedom and human nature."
Artificial Intelligence (AI)
Privacy
Data Collection
Surveliance
Biometric
Data
Facial Recognition
Preserving AI Techniques
Consent and Transparency
Privacy Regulation and Compliance
Education and Awareness
Anonos FTC Comment Letter Big Data: A Tool for Inclusion or ExclusionTed Myerson
FTC Comment Letter Big Data: A Tool for Inclusion or Exclusion. Filed on August 21, 2014.
Anonos has been working for over two years on technology that transforms data at the data element level enabling de-identification and functional obscurity that preserves the value of underlying data. Specifically, Anonos de-identification and functional obscurity risk management tools help to enable data subjects to share information in a controlled manner, enabling them to receive information and offerings truly personalized for them, while protecting misuse of their data; and to facilitate improved healthcare, medical research and personalized medicine by enabling aggregation of patient level data without revealing the identity of patients.
Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019...e-SIDES.eu
This is the slide-deck of the community event held on November 14, 2019 in Brussels, titled "Beyond Privacy: Learning Data Ethics - European Big Data Community Forum 2019". It includes the presentations given by the speakers.
An Investigation on Scalable and Efficient Privacy Preserving Challenges for ...IJERDJOURNAL
ABSTRACT:- Big data is a relative term describing a situation where the volume, velocity and variety of data exceed an organization’s storage or compute capacity for accurate and timely decision making. Big data refers to huge amount of digital information collected from multiple and different sources. With the development of application of Internet/Mobile Internet, social networks, Internet of Things, big data has become the hot topic of research across the world, at the same time; big data faces security risks and privacy protection during collecting, storing, analyzing and utilizing. Since a key point of big data is to access data from multiple and different domains security and privacy will play an important role in big data research and technology. Traditional security mechanisms, which are used to secure small scale static data, are inadequate. So the question is which security and privacy technology is adequate for efficient access to big data. This paper introduces the functions of big data, and the security threat faced by big data, then proposes the technology to solve the security threat, finally, discusses the applications of big data in information security. Main expectation from the focused challenges is that it will bring a novel focus on the big data infrastructure.
A practical data privacy and security approach to ffiec, gdpr and ccpaUlf Mattsson
With sensitive data residing everywhere, organizations becoming more mobile, and the breach epidemic growing, the need for advanced data privacy and security solutions has become even more critical. French regulators cited GDPR in fining Google $57 million and the U.K.'s Information Commissioner's Office is seeking a $230 million fine against British Airways and seeking $124 million from Marriott. Facebook is setting aside $3 billion to cover the costs of a privacy investigation launched by US regulators.
This session will take a practical approach to address guidance and standards from the Federal Financial Institutions Examination Council (FFIEC), EU GDPR, California CCPA, NIST Risk Management Framework, COBIT and the ISO 31000 Risk management Principles and Guidelines.
Learn how new data privacy and security techniques can help with compliance and data breaches, on-premises, and in public and private clouds.
Privacy through Anonymisation in Large-scale Socio-technical Systems: The BIS...Andrea Omicini
Large-scale socio-technical systems (STS) inextricably inter-connect individual – e.g., the right to privacy –, social – e.g., the effectiveness of organisational processes –, and technology issues —e.g., the software engineering process. As a result, the design of the complex software infrastructure involves also non-technological aspects such as the legal ones—so that, e.g., law-abidingness can be ensured since the early stages of the software engineering process. By focussing on contact centres (CC) as relevant examples of knowledge-intensive STS, we elaborate on the articulate aspects of anonymisation: there, individual and organisational needs clash, so that only an accurate balancing between legal and technical aspects could possibly ensure the system efficiency while preserving the individual right to privacy. We discuss first the overall legal framework, then the general theme of anonymisation in CC. Finally we overview the technical process developed in the context of the BISON project.
Project presentation @ DMI, Università di Catania, Italy, 25 July 2016
Non-technical talk for managers and Data Protection Officers about how the reasons behind the automation of creating a global data mapping for GDPR (at least), the challenges and possible methodologies using a new concept of Process Mining based on Data Activities
Scala: the unpredicted lingua franca for data scienceAndy Petrella
Talk given at Strata London with Dean Wampler (Lightbend) about Scala as the future of Data Science. First part is an approach of how scala became important, the remaining part of the talk is in notebooks using the Spark Notebook (http://spark-notebook.io/).
The notebooks are available on GitHub: https://github.com/data-fellas/scala-for-data-science.
Agile data science: Distributed, Interactive, Integrated, Semantic, Micro Ser...Andy Petrella
Distributed Data Science…
* A genomics use case
* Spark Notebook
* Interactive Distributed Data Science
Distributed Data Science… Pipeline
* Pipeline: productizing Data Science
* Demo of Distributed Pipeline (ADAM, Akka, Cassandra, Parquet, Spark)
* Why Micro Services?
* Painful points:
* Data science is Discontiguous
* Context Lost in Translation
* Solution: Data Fellas’ Agile Data Science Toolkit
What is a distributed data science pipeline. how with apache spark and friends.Andy Petrella
What was a data product before the world changed and got so complex.
Why distributed computing/data science is the solution.
What problems does that add?
How to solve most of them using the right technologies like spark notebook, spark, scala, mesos and so on in a accompanied framework
Towards a rebirth of data science (by Data Fellas)Andy Petrella
Nowadays, Data Science is buzzing all over the place.
But what is a, so-called, Data Scientist?
Some will argue that a Data Scientist is a person able to report and present insights in a data set. Others will say that a Data Scientist can handle a high throughput of values and expose them in services. Yet another definition includes the capacity to create meaningful visualizations on the data.
However, we enter an age where velocity is a key. Not only the velocity of your data is high, but the time to market is shortened. Hence, the time separating the moment you receive a set of data and the time you’ll be able to deliver added value is crucial.
In this talk, we’ll review the legacy Data Science methodologies, what it meant in terms of delivered work and results.
Afterwards, we’ll slightly move towards different concepts, techniques and tools that Data Scientists will have to learn and appropriate in order to accomplish their tasks in the age of Big Data.
The dissertation is closed by exposing the Data Fellas view on a solution to the challenges, specially thanks to the Spark Notebook and the Shar3 product we develop.
Distributed machine learning 101 using apache spark from a browser devoxx.b...Andy Petrella
A 3 hours session introducing the concept of Machine Learning and Distributed Computing.
It includes many examples running in notebooks of experience run on data exploring models like LM, RF, K-Means, Deep Learning.
Spark Summit Europe: Share and analyse genomic data at scaleAndy Petrella
Share and analyse genomic data
at scale with Spark, Adam, Tachyon & the Spark Notebook
Sharp intro to Genomics data
What are the Challenges
Distributed Machine Learning to the rescue
Projects: Distributed teams
Research: Long process
Towards Maximum Share for efficiency
Leveraging mesos as the ultimate distributed data science platformAndy Petrella
Keynote at the first @MesosCon #Europe on what was Data Science, what are the new challenge and needs and how we target them in Data Fellas with the Spark Notebook and Shar3
Data Enthusiasts London: Scalable and Interoperable data services. Applied to...Andy Petrella
Data science requires so many skills, people and time before the results can be accessed. Moreover, these results cannot be static anymore. And finally, the Big Data comes to the plate and the whole tool chain needs to change.
In this talk Data Fellas introduces Shar3, a tool kit aiming to bridged the gaps to build a interactive distributed data processing pipeline, or loop!
Then the talk covers genomics nowadays problems including data types, processing, discovery by introducing the GA4GH initiative and its implementation using Shar3.
Spark meetup london share and analyse genomic data at scale with spark, adam...Andy Petrella
Genomics and Health data is nowadays one of the hot topics requiring lots of computations and specially machine learning. This helps science with a very relevant societal impact to get even better outcome. That is why Apache Spark and its ADAM library is a must have.
This talk will be twofold.
First, we'll show how Apache Spark, MLlib and ADAM can be plugged all together to extract information from even huge and wide genomics dataset. Everything will be packed into examples from the Spark Notebook, showing how bio-scientists can work interactively with such a system.
Second, we'll explain how these methodologies and even the datasets themselves can be shared at very large scale between remote entities like hospitals or laboratories using micro services leveraging Apache Spark, ADAM, Play Framework 2, Avro and Tachyon.
Distributed machine learning 101 using apache spark from the browserAndy Petrella
Talk given by Xavier Tordoir and myself at Scala Days Amsterdam 2015.
Contains intro to ML, focusing on what is it and models selection via the Bias Variation constraint.
Then switches a gear to show how genomics can be learned using LDA, KMeans and Random Forest.
Finishes with some insight on what we'll change in the future regarding machine learning and modeling.
In this talk, I fly over the different concepts and advantages of Open Source, Open Data, Crowd Sourcing and Coworking in the context of Startups.
Yet, I put the focus on Data science related entrepreneurship, the domain I live in.
BioBankCloud: Machine Learning on Genomics + GA4GH @ Med at ScaleAndy Petrella
A talk given at the BioBankCloud conference in Feb 2015 about distributed computing in the contexts of genomics and health.
In this one, we exposed what results we obtained exploring the 1000genomes data using ADAM, followed by an introduction to our scalable GA4GH server implementation built using ADAM, Apache Spark and Play Framework 2.
What is Distributed Computing, Why we use Apache SparkAndy Petrella
In this talk we introduce the notion of distributed computing then we tackle the Spark advantages.
The Spark core content is very tiny because the whole explanation has been done live using a Spark Notebook (https://github.com/andypetrella/spark-notebook/blob/geek/conf/notebooks/Geek.snb).
This talk has been given together by @xtordoir and myself at the University of Liège, Belgium.
Lightning fast genomics with Spark, Adam and ScalaAndy Petrella
We are at a time where biotech allow us to get personal genomes for $1000. Tremendous progress since the 70s in DNA sequencing have been done, e.g. more samples in an experiment, more genomic coverages at higher speeds. Genomic analysis standards that have been developed over the years weren't designed with scalability and adaptability in mind. In this talk, we’ll present a game changing technology in this area, ADAM, initiated by the AMPLab at Berkeley. ADAM is framework based on Apache Spark and the Parquet storage. We’ll see how it can speed up a sequence reconstruction to a factor 150.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
State of ICS and IoT Cyber Threat Landscape Report 2024 previewPrayukth K V
The IoT and OT threat landscape report has been prepared by the Threat Research Team at Sectrio using data from Sectrio, cyber threat intelligence farming facilities spread across over 85 cities around the world. In addition, Sectrio also runs AI-based advanced threat and payload engagement facilities that serve as sinks to attract and engage sophisticated threat actors, and newer malware including new variants and latent threats that are at an earlier stage of development.
The latest edition of the OT/ICS and IoT security Threat Landscape Report 2024 also covers:
State of global ICS asset and network exposure
Sectoral targets and attacks as well as the cost of ransom
Global APT activity, AI usage, actor and tactic profiles, and implications
Rise in volumes of AI-powered cyberattacks
Major cyber events in 2024
Malware and malicious payload trends
Cyberattack types and targets
Vulnerability exploit attempts on CVEs
Attacks on counties – USA
Expansion of bot farms – how, where, and why
In-depth analysis of the cyber threat landscape across North America, South America, Europe, APAC, and the Middle East
Why are attacks on smart factories rising?
Cyber risk predictions
Axis of attacks – Europe
Systemic attacks in the Middle East
Download the full report from here:
https://sectrio.com/resources/ot-threat-landscape-reports/sectrio-releases-ot-ics-and-iot-security-threat-landscape-report-2024/
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
4. www.kensu.io
A. DATA PRIVACY
Information privacy, also known as data privacy or data protection, is the
relationship between the collection and dissemination of
a. data,
b. technology,
c. the public expectation of privacy,
d. legal
and political issues surrounding them.[1]
Privacy concerns exist wherever personally identifiable information or
other sensitive information is collected, stored, used, and finally
destroyed or deleted – in digital form or otherwise.
Improper or non-existent disclosure control can be the root cause for
privacy issues.
https://en.wikipedia.org/wiki/Information_privacy
5. www.kensu.io
Each controller/processor shall maintain a record of
processing activities under its responsibility (cf. Art. 30).
That record shall contain many information including:
• The purposes of the processing
• A description of the categories of data subjects and of
the categories of personal data
etc.
A. DATA PRIVACY
GDPR
6. www.kensu.io
A. DATA PRIVACY
Prior to collecting Californian’s personal data, businesses
must disclose in their privacy policy:
“the categories of personal information to be collected and
the purposes for which the categories of personal
information shall be used”
with any additional uses requiring notice to the
consumer
CaCPA: California Consumer Privacy Act of 2018
7. www.kensu.io
B. RISKS
Risks are present wherever data is used:
- Managing business risks with data
- Building new data business
https://www.eiuperspectives.economist.com/sites/default/files/RetailBanksandBigData.pdf
8. www.kensu.io
B. RISKS
- Retail worry about credit risk:
imbalance between the sizes of classes (defaulters <<< non-defaulters)
generates overly optimistic scores…
- Commercial focus on market risk:
VaR and variations requires important backtesting
- Investment are concerned about operational risk:
Just think about BCBS… govern, monitor, control!
Business’ risks… risks
12. www.kensu.io
C. ETHIC
Data Ethics refers to systemising, defending, and recommending
concepts of right and wrong conduct in relation to data, in particular
personal data.
Data ethics is different from information ethics because the focus of
information ethics is more concerned with issues of intellectual property.
https://en.wikipedia.org/wiki/Big_data_ethics
While data ethics is more concerned with collectors and
disseminators of structured or unstructured data such as
data brokers — governments — large corporations.
13. www.kensu.io
C. ETHIC
WAT?
http://rsta.royalsocietypublishing.org/content/roypta/374/2083/20160360.full.pdf
Data ethics can be defined as the branch of ethics that studies and evaluates moral
problems related to
data
- generation
- recording
- processing
- dissemination
- sharing and use
algorithms
- artificial intelligence
- artificial agents
- machine learning
- robots (well…)
practices
- responsible innovation
- programming
- hacking
- professional codes
in order to formulate and support morally good solutions
14. www.kensu.io
C. ETHIC
WAT?
http://rsta.royalsocietypublishing.org/content/roypta/374/2083/20160360.full.pdf
Data ethics can be defined as the branch of ethics that studies and evaluates moral
problems related to
data
- generation
- recording
- processing
- dissemination
- sharing and use
algorithms
- artificial intelligence
- artificial agents
- machine learning
- robots (well…)
practices
- responsible innovation
- programming
- hacking
- professional codes
in order to formulate and support morally good solutions
E
T
H
I C
?
E
T
H
I C
?
E
T
H
I C
?
15. www.kensu.io
C. ETHIC
WAT?
The ethics of data focuses on ethical problems posed by
the collection and analysis of large datasets and on
issues ranging from the use of big data in
- biomedical research and social sciences
- profilings
- advertising
- data philanthropy
- open data
16. www.kensu.io
C. ETHIC
WAT?
The ethics of algorithms addresses issues posed by the
increasing complexity and autonomy of algorithms
broadly understood, especially in the case of machine
learning applications.
Crucial challenges include moral responsibility and
accountability of both designers and data scientists with
respect to unforeseen and undesired consequences as
well as missed opportunities.
17. www.kensu.io
C. ETHIC
WAT?
The ethics of practices addresses the pressing questions
concerning the responsibilities and liabilities of people
and organizations in charge of data processes, strategies
and policies, including data scientists’ work to ensure ethical
practices fostering the protection of the data subject
rights.
23. www.kensu.io
C. ETHIC
Automated decision-making
https://arxiv.org/pdf/1606.08813.pdf
Right to explanation
Profiling is inherently discriminatory
Data subjects are grouped in categories and decisions
are made on this basis
Plus, as said, machine learning can reify existing patterns
of discrimination
Consequences: Biased decisions are presented as the
outcome of an “objective” algorithm.
25. www.kensu.io
C. ETHIC
Automated decision-making
https://arxiv.org/pdf/1606.08813.pdf
Right to explanation
For Burrell in How the machine “thinks”: Understanding opacity in
machine learning algorithms, there are three barriers to transparency
1. Intentional hiding of the decision procedures by corporations
2. Code sources are overly complex
3. Machine learning can reason at very high dimensions, humans’
brains don’t
27. www.kensu.io
X. HOW TO GUARANTEE COMPLIANCE
In (data) engineering, processes have improved to
satisfy the need for stability, quality and compliance
by introducing:
1. logging
2. testing
3. continuous deployment
Monitoring
28. www.kensu.io
X. HOW TO GUARANTEE COMPLIANCE
Data science projects are slightly different in nature than
pure engineering projects.
In that, most issues may come from the dynamicity of the
experimentations and the volatility of the data.
Such that, monitoring becomes key to AUTOMATED
compliance!
Monitoring
29. www.kensu.io
X. HOW TO GUARANTEE COMPLIANCE
For data project, monitoring is about:
- what/how data are used (e.g. data lineage, products, …)
- what/how models are build (e.g. methods, metrics, …)
- where/how data products are used (e.g. marketing, fraud, …)
Monitoring
30. www.kensu.io
X. HOW TO GUARANTEE COMPLIANCE
Pursuing the parallel with engineering:
CI/CD and Q/A are similar to our current compliance needs!
Automated Reporting
The automation of compliance can be approached with a
set of rules to estimate the level of risks and to limit the
efforts to only actionable events.
Reporting is mandatory for compliance.
Reports can be generated from the conjunction of
monitored activities and established rules dictated by
regulations.
31. www.kensu.io
X. HOW TO GUARANTEE COMPLIANCE
The Kensu way: Data Activity Manager
Monitor
Automated Registry Report
33. www.kensu.io
A. DATA IN THE WILD
Working on data is perceived as the Wild West.
• Experimentations in highly dynamic environments (e.g. notebooks)
• Local copy or duplication of datasets
• Creation of intermediate dumps (models, prepared datasets)
34. www.kensu.io
B. EFFECTS OF CONTRAINTS
Adding constraints (policies) to govern is a classic…
So, we would have the following examples:
- predefine the set of needed data
- list methods to be used
- create documents… maintain them
Rules, laws, …
35. www.kensu.io
B. EFFECTS OF CONTRAINTS
The consequences of such constraints are:
• Lack of freedom
• Anonymization
• what about marketing use case
• what is the reliability of the process
• anonymisation is actually itself a process to be listed!
• poor/slow reactivity to market changes (performance drop)
… might not be best
36. www.kensu.io
X. HOW TO GOVERN
For compliance reasons, we have to introduce monitoring.
Monitoring data opens new governance doors:
- Govern data activities with a bottom-up approach
- Control vs Constrain
In other terms, data governance in a data-driven fashion