Presentation given by Rommel Carvalho at the 5th Uncertainty Reasoning for the Semantic Web Workshop at the 8th International Semantic Web Conference in 2009.
Paper: Probabilistic Ontology and Knowledge Fusion for Procurement Fraud Detection in Brazil
Abstract: To cope with society’s demand for transparency and corruption prevention, the Brazilian Federal General Comptroller (CGU) has carried out a number of actions, including: awareness campaigns aimed at the private sector; campaigns to educate the public; research initiatives; and regular inspections and audits of municipalities and states. Although CGU has collected information from hundreds of different sources - Revenue Agency, Federal Police, and others - the process of fusing all this data has not been efficient enough to meet the needs of CGU’s decision makers. Therefore, it is natural to change the focus from data fusion to knowledge fusion. As a consequence, traditional syntactic methods must be augmented with techniques that represent and reason with the semantics of databases. However, commonly used approaches fail to deal with uncertainty, a dominant characteristic in corruption prevention. This paper presents the use of Probabilistic OWL (PR-OWL) to design and test a model that performs information fusion to detect possible frauds in procurements involving Federal money. To design this model, a recently developed tool for creating PR-OWL ontologies was used with support from PR-OWL specialists and careful guidance from a fraud detection specialist from CGU.
The document outlines the thesis defense of Violeta Damjanovic for her PhD in ambient intelligence and adaptive online experiments. The thesis addresses integrating probabilistic knowledge from pervasive semantic web environments into ontological models to enable adaptive and intelligent experimental environments. The proposed solution involves mechanisms for transforming probabilistic asynchronous process knowledge into ontologies and collecting ambient process knowledge to develop an adaptive semantic ambient system called AmIART. The thesis is expected to contribute to integrating the pervasive semantic web into online experimenting systems and adaptive systems considering uncertain knowledge.
The axiomatic power of Kolmogorov complexity lbienven
1. The document discusses random axioms and probabilistic proofs in Peano arithmetic. It describes a proof strategy where one could randomly select an integer n that satisfies some formula φ and add it as a new axiom.
2. While this intuition of probabilistic proofs makes sense, it is not really useful since any statement provable with sufficiently high probability is already provable in PA. However, probabilistic proofs can be exponentially more concise than deterministic proofs.
3. The document also discusses Kolmogorov complexity and how statements about it relate to the provability of PA. It can be shown that if C(x) is less than some value, PA will prove it, but PA will never prove a
UnBBayes is a probabilistic network framework written in Java. It has both a GUI and an API with inference, sampling, learning and evaluation. It supports BN, ID, MSBN, OOBN, HBN, MEBN/PR-OWL, structure, parameter and incremental learning.
The overview is presented through a slides potpourri from different presentations the Artificial Intelligence Group (GIA) from University of Brasilia (UnB) has given since 1999. It covers BN, ID, MSBN, UnBBayes Server, and MEBN.
This presentation was given by Rommel Carvalho when he started his PhD at George Mason University on the Friday seminar called Krypton (http://krypton.c4i.gmu.edu/).
Probabilistic Ontology: Representation and Modeling MethodologyRommel Carvalho
Oral Defense of Doctoral Dissertation
Volgenau School of Engineering, George Mason University
Rommel Novaes Carvalho
Bachelor of Science, University of Brasília, Brazil, 2003
Master of Science, University of Brasília, Brazil, 2008
Probabilistic Ontology: Representation and Modeling Methodology
Tuesday, June 28, 2011, 2:00pm -- 4:00pm
Nguyen Engineering Building, Room 4705
Committee
Kathryn Laskey, Chair
Paulo Costa
Kuo-Chu Chang
David Schum
Larry Kerschberg
Fabio Cozman
Abstract
The past few years have witnessed an increasingly mature body of research on the Semantic Web (SW), with new standards being developed and more complex problems being addressed. As complexity increases in SW applications, so does the need for principled means to cope with uncertainty in SW applications. Several approaches addressing uncertainty representation and reasoning in the SW have emerged. Among these is Probabilistic Web Ontology Language (PR-OWL), which provides Web Ontology Language (OWL) constructs for representing Multi-Entity Bayesian Network (MEBN) theories. However, there are several important ways in which the initial version PR-OWL 1.0 fails to achieve full compatibility with OWL. Furthermore, although there is an emerging literature on ontology engineering, little guidance is available on the construction of probabilistic ontologies.
This research proposes a new syntax and semantics, defined as PR-OWL 2.0, which improves compatibility between PR-OWL and OWL in two important respects. First, PR-OWL 2.0 follows the approach suggested by Poole et al. to formalizing the association between random variables from probabilistic theories with the individuals, classes and properties from ontological languages such as OWL. Second, PR-OWL 2.0 allows values of random variables to range over OWL datatypes.
To address the lack of support for probabilistic ontology engineering, this research describes a new methodology for modeling probabilistic ontologies called Uncertainty Modeling Process for Semantic Technologies (UMP-ST). To better explain the methodology and to verify that it can be applied to different scenarios, this dissertation presents step-by-step constructions of two different probabilistic ontologies. One is used for identifying frauds in public procurements in Brazil and the other is used for identifying terrorist threats in the maritime domain. Both use cases demonstrate the advantages of PR-OWL 2.0 over its predecessor.
1. The document describes Probabilistic Soft Logic (PSL), a probabilistic modeling language based on logics.
2. PSL uses rules to capture dependencies and constraints between continuous random variables represented as atoms. A PSL program consists of rules, sets, constraints, and atoms.
3. PSL provides a mathematical foundation based on constrained continuous Markov random fields and a logical foundation based on generalized annotated logic programs. It allows for collective probabilistic inference and learning over relational domains.
The document discusses probabilistic reasoning and probabilistic models. It introduces key concepts like representing knowledge with certainty factors rather than simple logic, defining sample spaces and probability distributions, calculating marginal and conditional probabilities, and using important probabilistic inference rules like the product rule and Bayes' rule. It provides examples of modeling problems with random variables and probabilities, like determining the probability of a disease given a positive test result.
Probabilistic Abductive Logic Programming using Possible WorldsFulvio Rotella
Reasoning in very complex contexts often requires purely deductive reasoning to be supported by a variety of techniques that can cope with incomplete data. Abductive inference allows to guess information that has not been explicitly observed. Since there are many explanations for such guesses, there is the need for assigning a probability to each one. This work exploits logical abduction to produce multiple explanations consistent with a given background knowledge and defines a strategy to prioritize them using their chance of being true. Another novelty is the introduction of probabilistic integrity constraints rather than hard ones. Then we propose a strategy that learns model and parameters from data and exploits our Probabilistic Abductive Proof Procedure to classify never-seen instances. This approach has been tested on some standard datasets showing that it improves accuracy in presence of corruptions and missing data.
The document outlines the thesis defense of Violeta Damjanovic for her PhD in ambient intelligence and adaptive online experiments. The thesis addresses integrating probabilistic knowledge from pervasive semantic web environments into ontological models to enable adaptive and intelligent experimental environments. The proposed solution involves mechanisms for transforming probabilistic asynchronous process knowledge into ontologies and collecting ambient process knowledge to develop an adaptive semantic ambient system called AmIART. The thesis is expected to contribute to integrating the pervasive semantic web into online experimenting systems and adaptive systems considering uncertain knowledge.
The axiomatic power of Kolmogorov complexity lbienven
1. The document discusses random axioms and probabilistic proofs in Peano arithmetic. It describes a proof strategy where one could randomly select an integer n that satisfies some formula φ and add it as a new axiom.
2. While this intuition of probabilistic proofs makes sense, it is not really useful since any statement provable with sufficiently high probability is already provable in PA. However, probabilistic proofs can be exponentially more concise than deterministic proofs.
3. The document also discusses Kolmogorov complexity and how statements about it relate to the provability of PA. It can be shown that if C(x) is less than some value, PA will prove it, but PA will never prove a
UnBBayes is a probabilistic network framework written in Java. It has both a GUI and an API with inference, sampling, learning and evaluation. It supports BN, ID, MSBN, OOBN, HBN, MEBN/PR-OWL, structure, parameter and incremental learning.
The overview is presented through a slides potpourri from different presentations the Artificial Intelligence Group (GIA) from University of Brasilia (UnB) has given since 1999. It covers BN, ID, MSBN, UnBBayes Server, and MEBN.
This presentation was given by Rommel Carvalho when he started his PhD at George Mason University on the Friday seminar called Krypton (http://krypton.c4i.gmu.edu/).
Probabilistic Ontology: Representation and Modeling MethodologyRommel Carvalho
Oral Defense of Doctoral Dissertation
Volgenau School of Engineering, George Mason University
Rommel Novaes Carvalho
Bachelor of Science, University of Brasília, Brazil, 2003
Master of Science, University of Brasília, Brazil, 2008
Probabilistic Ontology: Representation and Modeling Methodology
Tuesday, June 28, 2011, 2:00pm -- 4:00pm
Nguyen Engineering Building, Room 4705
Committee
Kathryn Laskey, Chair
Paulo Costa
Kuo-Chu Chang
David Schum
Larry Kerschberg
Fabio Cozman
Abstract
The past few years have witnessed an increasingly mature body of research on the Semantic Web (SW), with new standards being developed and more complex problems being addressed. As complexity increases in SW applications, so does the need for principled means to cope with uncertainty in SW applications. Several approaches addressing uncertainty representation and reasoning in the SW have emerged. Among these is Probabilistic Web Ontology Language (PR-OWL), which provides Web Ontology Language (OWL) constructs for representing Multi-Entity Bayesian Network (MEBN) theories. However, there are several important ways in which the initial version PR-OWL 1.0 fails to achieve full compatibility with OWL. Furthermore, although there is an emerging literature on ontology engineering, little guidance is available on the construction of probabilistic ontologies.
This research proposes a new syntax and semantics, defined as PR-OWL 2.0, which improves compatibility between PR-OWL and OWL in two important respects. First, PR-OWL 2.0 follows the approach suggested by Poole et al. to formalizing the association between random variables from probabilistic theories with the individuals, classes and properties from ontological languages such as OWL. Second, PR-OWL 2.0 allows values of random variables to range over OWL datatypes.
To address the lack of support for probabilistic ontology engineering, this research describes a new methodology for modeling probabilistic ontologies called Uncertainty Modeling Process for Semantic Technologies (UMP-ST). To better explain the methodology and to verify that it can be applied to different scenarios, this dissertation presents step-by-step constructions of two different probabilistic ontologies. One is used for identifying frauds in public procurements in Brazil and the other is used for identifying terrorist threats in the maritime domain. Both use cases demonstrate the advantages of PR-OWL 2.0 over its predecessor.
1. The document describes Probabilistic Soft Logic (PSL), a probabilistic modeling language based on logics.
2. PSL uses rules to capture dependencies and constraints between continuous random variables represented as atoms. A PSL program consists of rules, sets, constraints, and atoms.
3. PSL provides a mathematical foundation based on constrained continuous Markov random fields and a logical foundation based on generalized annotated logic programs. It allows for collective probabilistic inference and learning over relational domains.
The document discusses probabilistic reasoning and probabilistic models. It introduces key concepts like representing knowledge with certainty factors rather than simple logic, defining sample spaces and probability distributions, calculating marginal and conditional probabilities, and using important probabilistic inference rules like the product rule and Bayes' rule. It provides examples of modeling problems with random variables and probabilities, like determining the probability of a disease given a positive test result.
Probabilistic Abductive Logic Programming using Possible WorldsFulvio Rotella
Reasoning in very complex contexts often requires purely deductive reasoning to be supported by a variety of techniques that can cope with incomplete data. Abductive inference allows to guess information that has not been explicitly observed. Since there are many explanations for such guesses, there is the need for assigning a probability to each one. This work exploits logical abduction to produce multiple explanations consistent with a given background knowledge and defines a strategy to prioritize them using their chance of being true. Another novelty is the introduction of probabilistic integrity constraints rather than hard ones. Then we propose a strategy that learns model and parameters from data and exploits our Probabilistic Abductive Proof Procedure to classify never-seen instances. This approach has been tested on some standard datasets showing that it improves accuracy in presence of corruptions and missing data.
Discovering knowledge using web structure miningAtul Khanna
This document discusses web mining and algorithms for analyzing link structure on the web. It defines web mining as the process of discovering useful information from web data. There are three categories of web mining: web content mining, web structure mining, and web usage mining. Two important algorithms for analyzing hyperlink structure are HITS and PageRank. HITS identifies authoritative and hub pages, while PageRank calculates the importance of pages based on the number and quality of inbound links. The document provides details on how these algorithms work and potential applications.
This document introduces soft computing and provides an agenda for the lecture. Soft computing is defined as a fusion of fuzzy logic, neural networks, evolutionary computing, and probabilistic computing to deal with uncertainty and imprecision. Hybrid systems combine different soft computing techniques for improved performance. The lecture will cover an introduction to soft computing, fuzzy computing, neural networks, evolutionary computing, and hybrid systems. References are also provided.
The document summarizes Noah Goodman's talk on using mathematical principles to understand thought. It discusses how thought is productive through compositional representations and probabilistic inference. Thought combines basic elements in infinite combinations, like words into sentences. Thinking involves probabilistic inference over mental representations to explain observations and plan actions. The talk suggests thought may operate via a probabilistic language of thought based on probabilistic lambda calculus.
Engineering Ambient Intelligence Systems using Agent TechnologyNikolaos Spanoudakis
This presentation was given at the nectar session of the 9th Hellenic Conference on Artificial Intelligence (SETN 2016) that took place on May18th- 20th in Thessaloniki.
It is about applying an agent-oriented software engineering (AOSE) methodology, i.e. the Agent Systems Engineering Methodology (ASEME) for building intelligent systems. We present it along with a case study in the Ambient Intelligence (AmI) Application Domain. We discuss the challenges, the ASEME Methodology, the System Architecture and our results.
Bayesian Network Modeling using Python and RPyData
This document discusses Bayesian network modeling using Python and R. It begins with an introduction to Bayesian networks and their applications. It then outlines the main Bayesian network packages available in Python like scikit-learn, BayesPy, Bayes Blocks, and PyMC, and in R like bnlearn and RStan. It covers the basics of Bayes' theorem and how Bayesian networks represent probabilistic relationships between variables as a directed acyclic graph. The talk concludes with discussing algorithms for learning Bayesian networks from data and evaluating model performance.
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...PyData
This document discusses using Bayesian networks to model relationships in data. It introduces Bayesian networks as directed acyclic graphs that represent conditional dependencies between random variables. The document describes approaches for finding the optimal Bayesian network structure given data, including scoring functions and dealing with issues like cycles. It also introduces BNFinder, an open-source Python library for learning Bayesian networks from data that can handle both discrete and continuous variables efficiently in parallel. Examples are given demonstrating BNFinder's ability to learn predictive models from genomic and gene expression data.
Anthropological Research and TechniquesPaulVMcDowell
The document discusses two main approaches to anthropological theory - the scientific approach which seeks universal principles, and the experiential approach which focuses on personal experiences in other cultures. It also covers fundamental principles of anthropology like holism, cross-cultural comparison, and cultural relativism. The key methods discussed are observation, participant observation, interviews, and developing hypotheses through induction and testing them through deduction.
Transepistemic Abduction and its Application to Reflective Writing AnalyticsAndrew Gibson
The document discusses reflective writing analytics and proposes a conceptual model. It introduces two epistemic domains - the psychosocial and computational domains - that provide different perspectives on analyzing reflective writing. The model represents these domains and aspects within them, including interpretation, reflection, expression, construction, symbolization, processing, translation, and explanation. It also proposes a specialized mode of reasoning called "transepistemic abduction" to facilitate reconciliation between apparently irreconcilable explanations from the two domains. The goal is to enable progress in understanding reflective writing while maintaining integrity within the psychosocial and computational perspectives.
Biopsy is the removal of tissue from the living body for diagnostic purposes. It has a long history dating back to the 16th century. There are various biopsy techniques depending on the location and size of the lesion. The goal is to provide a representative tissue sample while minimizing patient discomfort. Common techniques include incisional, excisional, punch and needle biopsies. Indications are to confirm clinical impressions, determine treatment plans, and assess malignancy. Contraindications include certain vascular or pigmented lesions.
How to Become a Thought Leader in Your NicheLeslie Samuel
Are bloggers thought leaders? Here are some tips on how you can become one. Provide great value, put awesome content out there on a regular basis, and help others.
Ouvidoria de Balcão vs Ouvidoria Digital: Desafios na Era Big DataRommel Carvalho
Apresentação realizada no dia 14/03/2017 por Rommel N. Carvalho na Semana de Ouvidoria e Acesso à Informação de 2017, organizada pela CGU.
YouTube: https://youtu.be/vNMtULu5X1c?t=3h20m24s
Como transformar servidores em cientistas de dados e diminuir a distância ent...Rommel Carvalho
Palestra ministrada pelo Dr. Rommel Novaes Carvalho, Coordenador-Geral do Observatório da Despesa Pública e Professor do Mestrado Profissional em Computação Aplicada da UnB.
Evento: Brasil 100% Digital: Integração e transparência a serviço da sociedade
Website: http://www.brasildigital.gov.br/
Data: 10/11/2016
Vídeo: https://www.youtube.com/watch?v=3WYQlPR-RLw&feature=youtu.be&t=2h4m44s
Proposta de Modelo de Classificação de Riscos de Contratos PúblicosRommel Carvalho
O documento propõe três modelos para avaliar o risco de contratos públicos: 1) um modelo de aprendizagem supervisionada para classificar o risco de fornecedores com base em variáveis como doações políticas e histórico de punições; 2) um segundo modelo para classificar o risco de contratos com base em aspectos como competitividade e complexidade; 3) um modelo multicritério para selecionar casos de auditoria com base no risco do contrato, risco da empresa, e questões logísticas.
Categorização de achados em auditorias de TI com modelos supervisionados e nã...Rommel Carvalho
Palestra ministrada pela Patrícia Maia no 2o Seminário sobre Análise de Dados na Administração Pública @ http://www.brasildigital.gov.br/
Resumo: O trabalho consistiu na aplicação de técnicas de mineração de textos para identificação dos principais assuntos abordados nas auditorias dos últimos cinco anos. Foram utilizadas duas abordagens: a abordagem supervisionada aplicando classificação de textos com o algoritmo Random Forest e a abordagem não supervisionada através da técnica de modelagem de tópicos Latent Dirichlet Allocation (LDA). O projeto piloto foi validado com as constatações de TI e está agora sendo estendido a constatações relacionadas a outros temas. O objetivo é permitir catalogar o histórico de constatações emitidas e categorizar automaticamente novos registros. Com isso, os servidores poderão recuperar situações semelhantes para aplicação em novos trabalhos ou, ainda, tratar problemas recorrentes de forma estruturante. Além disso a mesma lógica pode ser usada para gerar conhecimento a partir de outros tipos de texto: pedidos com base na Lei de Acesso à Informações, manifestações do e-OUV, processos analisados pela CRG, notícias de interesse do órgão, etc.
Palestrante: Patrícia Maia - Ministério da Transparência, Fiscalização e Controle
Currículo: Possui mestrado em Computação Aplicada pela Universidade de Brasília (UNB), especialização em Modelagem de Processos e Engenharia de Requisitos pela Universidade Federal do Rio Grande do SUL (UFRGS) e graduação em Tecnologia da Informação. Tem experiência profissional nas áreas de mineração de textos, ETL, banco de dados e controle governamental. Trabalha atualmente no Ministério da Transparência, Fiscalização e Controle (MTFC), exercendo suas atividades na Diretoria de Pesquisas e Informações Estratégicas.
Mapeamento de risco de corrupção na administração pública federalRommel Carvalho
O documento descreve um projeto do governo brasileiro para mapear o risco de corrupção na administração pública federal através da análise e mineração de dados sobre servidores públicos e unidades governamentais. O projeto usa técnicas avançadas de aprendizado de máquina e análise estatística de grandes conjuntos de dados para gerar indicadores confiáveis de risco de corrupção. O objetivo final é fornecer uma ferramenta estratégica para prevenir e combater a corrupção de forma proativa.
1) O Observatório da Despesa Pública utiliza técnicas de ciência de dados para identificar riscos de fraude e irregularidades nos gastos públicos e apoiar a tomada de decisão dos gestores públicos.
2) Projetos como o Mapa de Risco de Fornecedores, a Análise Preventiva de Contratações e a Triagem Automática de Denúncias usam análises preditivas para prevenir situações de risco.
3) O Banco de Preços da APF permite pesquisas de mercado e identificação de sobrepreços nos contratos
Aplicação de técnicas de mineração de textos para classificação automática de...Rommel Carvalho
O uso de classificação automática de textos tem se tornado cada vez mais comum nos últimos anos. Contudo, ao se trabalhar com classificação em larga escala, a complexidade aumenta consideravelmente. Foi realizado um estudo de caso, aplicado à triagem de denúncias na Controladoria Geral da União, utilizando uma grande quantidade de categorias a serem classificadas. A solução proposta empregou aprendizagem de máquina e classificação multilabel. Essas técnicas tiveram como objetivo a construção de um modelo capaz de solucionar adversidades inerentes a este contexto, apresentando ganhos significativos
Patrícia Helena Maia Alves de Andrade - Controladoria-Geral da União
Analista de Finanças e Controle da CGU, atuando na área de mineração de textos e análise de dados, na Diretoria de Pesquisa e Informações Estratégicas. Atualmente está finalizando o Mestrado Profissional em Computação Aplicada na Universidade de Brasília
Filiação partidária e risco de corrupção de servidores públicos federaisRommel Carvalho
O documento discute o uso de aprendizado de máquina para analisar a relação entre filiação partidária e risco de corrupção entre servidores públicos federais brasileiros. Os dados mostraram uma correlação positiva entre filiação partidária e casos de corrupção. Um modelo de floresta aleatória obteve os melhores resultados, identificando variáveis-chave como tempo de filiação e motivo de cancelamento.
Uso de mineração de dados e textos para cálculo de preços de referência em co...Rommel Carvalho
Uma das grandes responsabilidades da CGU é identificar as compras do governo com valores diferentes dos praticados pelo mercado. Dessa forma, é possível mensurar o grau de eficiência das compras realizadas pelos órgãos governamentais. Essa informação é útil tanto para o auditor, que é responsável por fiscalizar o uso dos recursos públicos, como para o gestor, que pode melhorar seus processos observando as melhores práticas de outras unidades do governo. Dada a enorme quantidade e a diversidade das compras realizadas pelo Governo, essa análise se torna praticamente inviável sem a ajuda de algum mecanismo automatizado. No entanto, para que essa análise automatizada seja possível, é preciso ter antes de tudo, uma base de dados com os preços médios, ou de referência, para cada produto que se deseja analisar. Apesar de todas as compras do Governo Federal serem inseridas em um sistema único e centralizado, as informações armazenadas não são detalhadas e estruturadas o suficiente para se calcular esses preços de referência.
Essa palestra apresenta a metodologia desenvolvida na CGU, baseada em técnicas de mineração de dados, para extrair as informações necessárias desse sistema centralizado de forma a possibilitar o cálculo de preços de referência para produtos comprados pelo Governo Federal. Além disso, são apresentadas também algumas análises feitas com base no banco de preços criado a partir dessa metodologia de forma a enfatizar sua importância para a melhoria da gestão dos recursos públicos.
Rommel Novaes Carvalho - Controladoria-Geral da União
Coordenador-Geral do Observatório da Despesa Pública da CGU (http://www.cgu.gov.br/assuntos/informacoes-estrategicas/observatorio-da-despesa-publica), realizou seu PhD e Pós-Doc na George Mason University, EUA, na área de Inteligência Artificial, Web Semântica e Mineração de Dados e também é professor do Mestrado Profissional em Computação Aplicada da UnB
Detecção preventiva de fracionamento de comprasRommel Carvalho
O documento descreve um estudo sobre a detecção preventiva de fracionamento de compras no Brasil usando redes bayesianas. O estudo utilizou dados de compras do governo para criar um modelo capaz de identificar possíveis fracionamentos. Após a preparação dos dados, diferentes algoritmos de modelagem foram testados e avaliados, resultando em um modelo com alta acurácia e capacidade de classificação. O modelo foi implantado para alertar sobre possíveis fracionamentos em novas compras governamentais.
Identificação automática de tipos de pedidos mais frequentes da LAIRommel Carvalho
O documento descreve um método para identificar automaticamente os tipos de pedidos mais frequentes na Lei de Acesso à Informação (LAI) brasileira através da análise de tópicos em mais de 300 mil pedidos usando o modelo Latent Dirichlet Allocation (LDA). O método identificou vários tópicos comuns, incluindo pedidos sobre o Banco Central do Brasil (BACEN) e sobre concursos públicos. O processo levou cerca de 10 horas para analisar os 300 mil pedidos.
Discovering knowledge using web structure miningAtul Khanna
This document discusses web mining and algorithms for analyzing link structure on the web. It defines web mining as the process of discovering useful information from web data. There are three categories of web mining: web content mining, web structure mining, and web usage mining. Two important algorithms for analyzing hyperlink structure are HITS and PageRank. HITS identifies authoritative and hub pages, while PageRank calculates the importance of pages based on the number and quality of inbound links. The document provides details on how these algorithms work and potential applications.
This document introduces soft computing and provides an agenda for the lecture. Soft computing is defined as a fusion of fuzzy logic, neural networks, evolutionary computing, and probabilistic computing to deal with uncertainty and imprecision. Hybrid systems combine different soft computing techniques for improved performance. The lecture will cover an introduction to soft computing, fuzzy computing, neural networks, evolutionary computing, and hybrid systems. References are also provided.
The document summarizes Noah Goodman's talk on using mathematical principles to understand thought. It discusses how thought is productive through compositional representations and probabilistic inference. Thought combines basic elements in infinite combinations, like words into sentences. Thinking involves probabilistic inference over mental representations to explain observations and plan actions. The talk suggests thought may operate via a probabilistic language of thought based on probabilistic lambda calculus.
Engineering Ambient Intelligence Systems using Agent TechnologyNikolaos Spanoudakis
This presentation was given at the nectar session of the 9th Hellenic Conference on Artificial Intelligence (SETN 2016) that took place on May18th- 20th in Thessaloniki.
It is about applying an agent-oriented software engineering (AOSE) methodology, i.e. the Agent Systems Engineering Methodology (ASEME) for building intelligent systems. We present it along with a case study in the Ambient Intelligence (AmI) Application Domain. We discuss the challenges, the ASEME Methodology, the System Architecture and our results.
Bayesian Network Modeling using Python and RPyData
This document discusses Bayesian network modeling using Python and R. It begins with an introduction to Bayesian networks and their applications. It then outlines the main Bayesian network packages available in Python like scikit-learn, BayesPy, Bayes Blocks, and PyMC, and in R like bnlearn and RStan. It covers the basics of Bayes' theorem and how Bayesian networks represent probabilistic relationships between variables as a directed acyclic graph. The talk concludes with discussing algorithms for learning Bayesian networks from data and evaluating model performance.
Understanding your data with Bayesian networks (in Python) by Bartek Wilczyns...PyData
This document discusses using Bayesian networks to model relationships in data. It introduces Bayesian networks as directed acyclic graphs that represent conditional dependencies between random variables. The document describes approaches for finding the optimal Bayesian network structure given data, including scoring functions and dealing with issues like cycles. It also introduces BNFinder, an open-source Python library for learning Bayesian networks from data that can handle both discrete and continuous variables efficiently in parallel. Examples are given demonstrating BNFinder's ability to learn predictive models from genomic and gene expression data.
Anthropological Research and TechniquesPaulVMcDowell
The document discusses two main approaches to anthropological theory - the scientific approach which seeks universal principles, and the experiential approach which focuses on personal experiences in other cultures. It also covers fundamental principles of anthropology like holism, cross-cultural comparison, and cultural relativism. The key methods discussed are observation, participant observation, interviews, and developing hypotheses through induction and testing them through deduction.
Transepistemic Abduction and its Application to Reflective Writing AnalyticsAndrew Gibson
The document discusses reflective writing analytics and proposes a conceptual model. It introduces two epistemic domains - the psychosocial and computational domains - that provide different perspectives on analyzing reflective writing. The model represents these domains and aspects within them, including interpretation, reflection, expression, construction, symbolization, processing, translation, and explanation. It also proposes a specialized mode of reasoning called "transepistemic abduction" to facilitate reconciliation between apparently irreconcilable explanations from the two domains. The goal is to enable progress in understanding reflective writing while maintaining integrity within the psychosocial and computational perspectives.
Biopsy is the removal of tissue from the living body for diagnostic purposes. It has a long history dating back to the 16th century. There are various biopsy techniques depending on the location and size of the lesion. The goal is to provide a representative tissue sample while minimizing patient discomfort. Common techniques include incisional, excisional, punch and needle biopsies. Indications are to confirm clinical impressions, determine treatment plans, and assess malignancy. Contraindications include certain vascular or pigmented lesions.
How to Become a Thought Leader in Your NicheLeslie Samuel
Are bloggers thought leaders? Here are some tips on how you can become one. Provide great value, put awesome content out there on a regular basis, and help others.
Ouvidoria de Balcão vs Ouvidoria Digital: Desafios na Era Big DataRommel Carvalho
Apresentação realizada no dia 14/03/2017 por Rommel N. Carvalho na Semana de Ouvidoria e Acesso à Informação de 2017, organizada pela CGU.
YouTube: https://youtu.be/vNMtULu5X1c?t=3h20m24s
Como transformar servidores em cientistas de dados e diminuir a distância ent...Rommel Carvalho
Palestra ministrada pelo Dr. Rommel Novaes Carvalho, Coordenador-Geral do Observatório da Despesa Pública e Professor do Mestrado Profissional em Computação Aplicada da UnB.
Evento: Brasil 100% Digital: Integração e transparência a serviço da sociedade
Website: http://www.brasildigital.gov.br/
Data: 10/11/2016
Vídeo: https://www.youtube.com/watch?v=3WYQlPR-RLw&feature=youtu.be&t=2h4m44s
Proposta de Modelo de Classificação de Riscos de Contratos PúblicosRommel Carvalho
O documento propõe três modelos para avaliar o risco de contratos públicos: 1) um modelo de aprendizagem supervisionada para classificar o risco de fornecedores com base em variáveis como doações políticas e histórico de punições; 2) um segundo modelo para classificar o risco de contratos com base em aspectos como competitividade e complexidade; 3) um modelo multicritério para selecionar casos de auditoria com base no risco do contrato, risco da empresa, e questões logísticas.
Categorização de achados em auditorias de TI com modelos supervisionados e nã...Rommel Carvalho
Palestra ministrada pela Patrícia Maia no 2o Seminário sobre Análise de Dados na Administração Pública @ http://www.brasildigital.gov.br/
Resumo: O trabalho consistiu na aplicação de técnicas de mineração de textos para identificação dos principais assuntos abordados nas auditorias dos últimos cinco anos. Foram utilizadas duas abordagens: a abordagem supervisionada aplicando classificação de textos com o algoritmo Random Forest e a abordagem não supervisionada através da técnica de modelagem de tópicos Latent Dirichlet Allocation (LDA). O projeto piloto foi validado com as constatações de TI e está agora sendo estendido a constatações relacionadas a outros temas. O objetivo é permitir catalogar o histórico de constatações emitidas e categorizar automaticamente novos registros. Com isso, os servidores poderão recuperar situações semelhantes para aplicação em novos trabalhos ou, ainda, tratar problemas recorrentes de forma estruturante. Além disso a mesma lógica pode ser usada para gerar conhecimento a partir de outros tipos de texto: pedidos com base na Lei de Acesso à Informações, manifestações do e-OUV, processos analisados pela CRG, notícias de interesse do órgão, etc.
Palestrante: Patrícia Maia - Ministério da Transparência, Fiscalização e Controle
Currículo: Possui mestrado em Computação Aplicada pela Universidade de Brasília (UNB), especialização em Modelagem de Processos e Engenharia de Requisitos pela Universidade Federal do Rio Grande do SUL (UFRGS) e graduação em Tecnologia da Informação. Tem experiência profissional nas áreas de mineração de textos, ETL, banco de dados e controle governamental. Trabalha atualmente no Ministério da Transparência, Fiscalização e Controle (MTFC), exercendo suas atividades na Diretoria de Pesquisas e Informações Estratégicas.
Mapeamento de risco de corrupção na administração pública federalRommel Carvalho
O documento descreve um projeto do governo brasileiro para mapear o risco de corrupção na administração pública federal através da análise e mineração de dados sobre servidores públicos e unidades governamentais. O projeto usa técnicas avançadas de aprendizado de máquina e análise estatística de grandes conjuntos de dados para gerar indicadores confiáveis de risco de corrupção. O objetivo final é fornecer uma ferramenta estratégica para prevenir e combater a corrupção de forma proativa.
1) O Observatório da Despesa Pública utiliza técnicas de ciência de dados para identificar riscos de fraude e irregularidades nos gastos públicos e apoiar a tomada de decisão dos gestores públicos.
2) Projetos como o Mapa de Risco de Fornecedores, a Análise Preventiva de Contratações e a Triagem Automática de Denúncias usam análises preditivas para prevenir situações de risco.
3) O Banco de Preços da APF permite pesquisas de mercado e identificação de sobrepreços nos contratos
Aplicação de técnicas de mineração de textos para classificação automática de...Rommel Carvalho
O uso de classificação automática de textos tem se tornado cada vez mais comum nos últimos anos. Contudo, ao se trabalhar com classificação em larga escala, a complexidade aumenta consideravelmente. Foi realizado um estudo de caso, aplicado à triagem de denúncias na Controladoria Geral da União, utilizando uma grande quantidade de categorias a serem classificadas. A solução proposta empregou aprendizagem de máquina e classificação multilabel. Essas técnicas tiveram como objetivo a construção de um modelo capaz de solucionar adversidades inerentes a este contexto, apresentando ganhos significativos
Patrícia Helena Maia Alves de Andrade - Controladoria-Geral da União
Analista de Finanças e Controle da CGU, atuando na área de mineração de textos e análise de dados, na Diretoria de Pesquisa e Informações Estratégicas. Atualmente está finalizando o Mestrado Profissional em Computação Aplicada na Universidade de Brasília
Filiação partidária e risco de corrupção de servidores públicos federaisRommel Carvalho
O documento discute o uso de aprendizado de máquina para analisar a relação entre filiação partidária e risco de corrupção entre servidores públicos federais brasileiros. Os dados mostraram uma correlação positiva entre filiação partidária e casos de corrupção. Um modelo de floresta aleatória obteve os melhores resultados, identificando variáveis-chave como tempo de filiação e motivo de cancelamento.
Uso de mineração de dados e textos para cálculo de preços de referência em co...Rommel Carvalho
Uma das grandes responsabilidades da CGU é identificar as compras do governo com valores diferentes dos praticados pelo mercado. Dessa forma, é possível mensurar o grau de eficiência das compras realizadas pelos órgãos governamentais. Essa informação é útil tanto para o auditor, que é responsável por fiscalizar o uso dos recursos públicos, como para o gestor, que pode melhorar seus processos observando as melhores práticas de outras unidades do governo. Dada a enorme quantidade e a diversidade das compras realizadas pelo Governo, essa análise se torna praticamente inviável sem a ajuda de algum mecanismo automatizado. No entanto, para que essa análise automatizada seja possível, é preciso ter antes de tudo, uma base de dados com os preços médios, ou de referência, para cada produto que se deseja analisar. Apesar de todas as compras do Governo Federal serem inseridas em um sistema único e centralizado, as informações armazenadas não são detalhadas e estruturadas o suficiente para se calcular esses preços de referência.
Essa palestra apresenta a metodologia desenvolvida na CGU, baseada em técnicas de mineração de dados, para extrair as informações necessárias desse sistema centralizado de forma a possibilitar o cálculo de preços de referência para produtos comprados pelo Governo Federal. Além disso, são apresentadas também algumas análises feitas com base no banco de preços criado a partir dessa metodologia de forma a enfatizar sua importância para a melhoria da gestão dos recursos públicos.
Rommel Novaes Carvalho - Controladoria-Geral da União
Coordenador-Geral do Observatório da Despesa Pública da CGU (http://www.cgu.gov.br/assuntos/informacoes-estrategicas/observatorio-da-despesa-publica), realizou seu PhD e Pós-Doc na George Mason University, EUA, na área de Inteligência Artificial, Web Semântica e Mineração de Dados e também é professor do Mestrado Profissional em Computação Aplicada da UnB
Detecção preventiva de fracionamento de comprasRommel Carvalho
O documento descreve um estudo sobre a detecção preventiva de fracionamento de compras no Brasil usando redes bayesianas. O estudo utilizou dados de compras do governo para criar um modelo capaz de identificar possíveis fracionamentos. Após a preparação dos dados, diferentes algoritmos de modelagem foram testados e avaliados, resultando em um modelo com alta acurácia e capacidade de classificação. O modelo foi implantado para alertar sobre possíveis fracionamentos em novas compras governamentais.
Identificação automática de tipos de pedidos mais frequentes da LAIRommel Carvalho
O documento descreve um método para identificar automaticamente os tipos de pedidos mais frequentes na Lei de Acesso à Informação (LAI) brasileira através da análise de tópicos em mais de 300 mil pedidos usando o modelo Latent Dirichlet Allocation (LDA). O método identificou vários tópicos comuns, incluindo pedidos sobre o Banco Central do Brasil (BACEN) e sobre concursos públicos. O processo levou cerca de 10 horas para analisar os 300 mil pedidos.
BMAW 2014 - Using Bayesian Networks to Identify and Prevent Split Purchases i...Rommel Carvalho
Presentation given by Rommel N. Carvalho at the 11th Bayesian Modeling Applications Workshop (BMAW 2014) at the 30th Conference on Uncertainty in Artificial Intelligence (UAI 2014) in July 27, 2014, Quebec City, Quebec, Canada. This was a joint work between the Research and Strategic Information Directorate from Brazil's Office of the Comptroller General and the Department of Computer Science from the University of Brasília.
Talk: https://www.youtube.com/watch?v=UVOsztdSQ3A
Paper: http://seor.gmu.edu/~klaskey/BMAW2014/BMAW2014_papers/bmaw2014_paper_6.pdf
Title: Using Bayesian Networks to Identify and Prevent Split Purchases in Brazil.
Abstract: To cope with society's demand for transparency and corruption prevention, the Brazilian Office of the Comptroller General (CGU) has carried out a number of actions, including: awareness campaigns aimed at the private sector; campaigns to educate the public; research initiatives; and regular inspections and audits of municipalities and states. Although CGU has collected information from various different sources - Revenue Agency, Federal Police, and others -, going through all the data in order to find suspicious transactions has proven to be really challenging. In this paper, we present a Data Mining study applied on real data - government purchases - for finding transactions that might become irregular before they are considered as such in order to act proactively. Moreover, we compare the performance of various Bayesian Network (BN) learning algorithms with different parameters in order to fine tune the learned models and improve their performance. The best result was obtained using the Tree Augmented Network (TAN) algorithm and oversampling the minority class in order to balance the data set. Using a 10-fold cross-validation, the model correctly classified all split purchases, it obtained a ROC area of .999, and its accuracy was 99.197%.
Presentation given by Rommel N. Carvalho at the 9th International Workshop on Uncertainty Reasoning for the Semantic Web at the 12th International Semantic Web Conference in October 21, 2013, Sydney, Australia. This was a joint work between the Research and Strategic Information Directorate from Brazil's Office of the Comptroller General and the Department of Computer Science from the University of Brasília.
Title: A GUI for MLN.
Abstract: This paper focuses on the incorporation of the Markov Logic Network (MLN) formalism as a plug-in for UnBBayes, a Java framework for probabilistic reasoning based on graphical models. MLN is a formalism for probabilistic reasoning which combines the capacity of dealing with uncertainty tolerating imperfections and contradictory knowledge based a Markov Network (MN) with the expressiveness of First Order Logic. A MLN provides a compact language for specifying very large MNs and the ability to incorporate, in modular form, large domain of knowledge (expressed in First Order Logic sentences) inside itself. A Graphical User Interface for the software Tuffy was implemented into UnBBayes to facilitate the creation, and inference of MLN models. Tuffy is a Java open source MLN engine.
Presentation given by Rommel N. Carvalho at the 9th International Workshop on Uncertainty Reasoning for the Semantic Web at the 12th International Semantic Web Conference in October 21, 2013, Sydney, Australia. This was a joint work between the Research and Strategic Information Directorate from Brazil's Office of the Comptroller General and the Department of Computer Science from the University of Brasília.
Title: UMP-ST plug-in: a tool for documenting, maintaining, and evolving probabilistic ontologies.
Abstract: Although several languages have been proposed for dealing with uncertainty in the Semantic Web (SW), almost no support has been given to ontological engineers on how to create such probabilistic ontologies (PO). This task of modeling POs has proven to be extremely difficult and hard to replicate. This paper presents the first tool in the world to implement a process which guides users in modeling POs, the Uncertainty Modeling Process for Semantic Technologies (UMP-ST). The tool solves three main problems: the complexity in creating POs; the difficulty in maintaining and evolving existing POs; and the lack of a centralized tool for documenting POs. Besides presenting the tool, which is implemented as a plug-in for UnBBayes, this papers also presents how the UMP-ST plug-in could have been used to build the Probabilistic Ontology for Procurement Fraud Detection and Prevention in Brazil, a proof-of-concept use case created as part of a research project at the Brazilian Office of the Comptroller General (CGU).
Integração do Portal da Copa @ Comissão CMA do Senado FederalRommel Carvalho
Apresentação preparada por Rommel N. Carvalho e apresentada pela Diretora de Sistemas e Informações da Controladoria-Geral da União (CGU), Tatiana Z. Panisset, na reunião da Comissão de Meio Ambiente, Defesa do Consumidor e Fiscalização e Controle (CMA) do Senado Federal (SF). A reunião teve como foco o debate da unificação da entrada de dados dos Portais de Transparência da Copa de 2014 do SF (www.copatransparente.gov.br) e da CGU (http://transparencia.gov.br/copa2014). Mais informações sobre a reunião em http://goo.gl/KCBD6.
As alternativas apresentadas foram discutidas e deliberadas pela CMA com aprovação da colaboração oficial entre o poder Legislativo e o poder Executivo para executar a integração da entrada de dados dos respectivos portais da copa do mundo. Notícias sobre essa colaboração podem ser encontradas em goo.gl/N8cbr, goo.gl/RVMGd, goo.gl/Ze3uJ, goo.gl/6o7BZ e goo.gl/C1CFv.
Título:
O que é e como usar dados abertos governamentais
Resumo:
A Web Semântica visa associar os dados disponibilizados na Web aos seus significados de forma a possibilitar que esses dados sejam compreensíveis tanto por humanos quanto por máquinas. Isso permitirá que tarefas, antes realizadas apenas por humanos, possam agora ser delegadas a máquinas. Técnicas de Web Semântica têm se difundido com o significativo aumento no número de aplicações que fazem uso de ontologias e semântica através de tecnologias como RDF, OWL, dentre outras, e as várias iniciativas espalhadas pelo mundo referente à disponibilização de dados abertos, em especial, de dados abertos governamentais. Dados abertos governamentais são definidos pela W3C – Consórcio da Web, como “a publicação e disseminação na Web de dados gerados pelo Setor Público, compartilhados em formato bruto e aberto, compreensíveis logicamente, de modo a permitir sua reutilização em aplicações digitais desenvolvidas pela sociedade”. O objetivo dessa palestra é apresentar os principais conceitos que norteiam as diversas iniciativas de dados abertos governamentais, a situação atual dessa iniciativa no Brasil, os benefícios que essa iniciativa traz para a sociedade como o uso desses dados abertos para contribuir com a melhoria e transparência da gestão pública.
Palestrante:
Dr. Rommel Novaes Carvalho, Ph.D
Postdoctoral Research Associate – C4I Center @ GMU
Analista de Finanças e Controle – CGU
http://mason.gmu.edu/~rcarvalh
CV resumido:
Rommel Novaes Carvalho é bacharel em Ciência da Computação e Mestre em Informática pela Universidade de Brasília, e doutor em Engenharia de Sistemas e Pesquisa Operacional pela Universidade George Mason, Estados Unidos. Pesquisador em Inteligência Artificial (IA) e membro do Grupo de Pesquisa em Inteligência Artificial da Universidade de Brasília (GIA). Suas áreas de interesse abrangem representação e raciocínio com incerteza na Web Semântica usando inferência bayesiana, mineração de dados, e engenharia de software. Desenvolvedor Java certificado, com experiência em implementação de sistemas de redes probabilísticas, sendo o arquiteto principal do projeto UnBBayes – Framework para raciocino probabilístico, em desenvolvimento pelo GIA desde 2000. Em seu doutorado propôs e implementou a versão 2 para o PR-OWL – Probabilistic OWL, para permitir o reuso de ontologias determinísticas existentes, sua interoperabilidade com ontologias probabilísticas representadas em PR-OWL, e raciocínio misto ontológico e probabilístico. Desde 2005 trabalha na Controladoria-Geral da União como especialista em Tecnologia da Informação. Em 2011, tornou-se pesquisador associado de Pós-Doutorado na George Mason University.
Modeling a Probabilistic Ontology for Maritime Domain AwarenessRommel Carvalho
The document describes developing a probabilistic ontology for maritime domain awareness. It aims to develop an ontology capable of reasoning with evidence from different domains to provide situational awareness. It discusses ontologies, probabilistic ontologies, and using the Probabilistic Web Ontology Language and other techniques. It also presents an uncertainty modeling process and incremental methodology for modeling the probabilistic ontology, including modeling cycles with goals, queries, evidence and assumptions.
SWRL-F - A Fuzzy Logic Extension of the Semantic Web Rule LanguageRommel Carvalho
Presentation given by Tomasz Wlodarczyk at the 6th Uncertainty Reasoning for the Semantic Web Workshop at the 9th International Semantic Web Conference in 2010.
Paper: SWRL-F - A Fuzzy Logic Extension of the Semantic Web Rule Language
Abstract: Enhancing Semantic Web technologies with an ability to express uncertainty and imprecision is widely discussed topic. While SWRL can provide additional expressivity to OWL-based ontologies, it does not provide any way to handle uncertainty or imprecision. We introduce an extension of SWRL called SWRL-F that is based on SWRL rule language and uses SWRL's strong semantic foundation as its formal underpinning. We extend it with a SWRL-F ontology to enable fuzzy reasoning in the rule base. The resulting language provides small but powerful set of fuzzy operations that do not introduce inconsistencies in the host ontology.
Default Logics for Plausible Reasoning with Controversial AxiomsRommel Carvalho
Presentation given by Thomas Scharrenbach at the 6th Uncertainty Reasoning for the Semantic Web Workshop at the 9th International Semantic Web Conference in 2010.
Paper: Default Logics for Plausible Reasoning with Controversial Axioms
Abstract: Using a variant of Lehmann's Default Logics and Probabilistic Description Logics we recently presented a framework that invalidates those unwanted inferences that cause concept unsatisfiability without the need to remove explicitly stated axioms. The solutions of this methods were shown to outperform classical ontology repair w.r.t. the number of inferences invalidated. However, conflicts may still exist in the knowledge base and can make reasoning ambiguous. Furthermore, solutions with a minimal number of inferences invalidated do not necessarily minimize the number of conflicts. In this paper we provide an overview over finding solutions that have a minimal number of conflicts while invalidating as few inferences as possible. Specifically, we propose to evaluate solutions w.r.t. the quantity of information they convey by recurring to the notion of entropy and discuss a possible approach towards computing the entropy w.r.t. an ABox.
Tractability of the Crisp Representations of Tractable Fuzzy Description LogicsRommel Carvalho
Presentation given by Fernando Bobillo at the 6th Uncertainty Reasoning for the Semantic Web Workshop at the 9th International Semantic Web Conference in 2010.
Paper: Tractability of the Crisp Representations of Tractable Fuzzy Description Logics
Abstract: An important line of research within the field of fuzzy DLs is the computation of an equivalent crisp representation of a fuzzy ontology. In this short paper, we discuss the relation between tractable fuzzy DLs and tractable crisp representations. This relation heavily depends on the family of fuzzy operators considered.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
Physiology and chemistry of skin and pigmentation, hairs, scalp, lips and nail, Cleansing cream, Lotions, Face powders, Face packs, Lipsticks, Bath products, soaps and baby product,
Preparation and standardization of the following : Tonic, Bleaches, Dentifrices and Mouth washes & Tooth Pastes, Cosmetics for Nails.
বাংলাদেশের অর্থনৈতিক সমীক্ষা ২০২৪ [Bangladesh Economic Review 2024 Bangla.pdf] কম্পিউটার , ট্যাব ও স্মার্ট ফোন ভার্সন সহ সম্পূর্ণ বাংলা ই-বুক বা pdf বই " সুচিপত্র ...বুকমার্ক মেনু 🔖 ও হাইপার লিংক মেনু 📝👆 যুক্ত ..
আমাদের সবার জন্য খুব খুব গুরুত্বপূর্ণ একটি বই ..বিসিএস, ব্যাংক, ইউনিভার্সিটি ভর্তি ও যে কোন প্রতিযোগিতা মূলক পরীক্ষার জন্য এর খুব ইম্পরট্যান্ট একটি বিষয় ...তাছাড়া বাংলাদেশের সাম্প্রতিক যে কোন ডাটা বা তথ্য এই বইতে পাবেন ...
তাই একজন নাগরিক হিসাবে এই তথ্য গুলো আপনার জানা প্রয়োজন ...।
বিসিএস ও ব্যাংক এর লিখিত পরীক্ষা ...+এছাড়া মাধ্যমিক ও উচ্চমাধ্যমিকের স্টুডেন্টদের জন্য অনেক কাজে আসবে ...
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
Main Java[All of the Base Concepts}.docxadhitya5119
This is part 1 of my Java Learning Journey. This Contains Custom methods, classes, constructors, packages, multithreading , try- catch block, finally block and more.
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Dr. Vinod Kumar Kanvaria
Exploiting Artificial Intelligence for Empowering Researchers and Faculty,
International FDP on Fundamentals of Research in Social Sciences
at Integral University, Lucknow, 06.06.2024
By Dr. Vinod Kumar Kanvaria
it describes the bony anatomy including the femoral head , acetabulum, labrum . also discusses the capsule , ligaments . muscle that act on the hip joint and the range of motion are outlined. factors affecting hip joint stability and weight transmission through the joint are summarized.
URSW 2009 - Probabilistic Ontology and Knowledge Fusion for Procurement Fraud Detection in Brazil
1. Probabilistic Ontology and Knowledge
Fusion for Procurement Fraud Detection
in Brazil
Rommel Carvalho, Kathryn Laskey, and Paulo Costa
George Mason University
Marcelo Ladeira, Laécio Santos, and Shou Matsumoto
Universidade de Brasília
Paper - Uncertainty Reasoning for the Semantic Web
URSW - ISWC
9. Introduction
Brazilian Office of the Comptroller General
(CGU) primary mission
Prevent and detect irregularities (corruption)
Gather information from a variety of sources
Combine the information
Then evaluate whether further action is necessary
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 4
10. Introduction
Brazilian Office of the Comptroller General
(CGU) primary mission
Prevent and detect irregularities (corruption)
Gather information from a variety of sources
Combine the information
Then evaluate whether further action is necessary
Problem
Information explosion
Growing Acceleration Program (PAC)
250 billion dollars - 1,000+ projects only in SP
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 4
12. Introduction
MEBN
Represent and reason with uncertainty about any
propositions that can be expressed in first-order logic
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 5
13. Introduction
MEBN
Represent and reason with uncertainty about any
propositions that can be expressed in first-order logic
PR-OWL
Uses MEBN logic to provide a framework for building
probabilistic ontologies
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 5
14. Introduction
MEBN
Represent and reason with uncertainty about any
propositions that can be expressed in first-order logic
PR-OWL
Uses MEBN logic to provide a framework for building
probabilistic ontologies
Fraud Detection and Prevention Model
Uses MEBN and PR-OWL
Proof of concept
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 5
28. Procurement Fraud Detection
A major source of corruption is the procurement
process
Laws attempt to ensure a competitive and fair process
Perpetrators find ways to turn the process to their
advantage while appearing to be legitimate
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 11
29. Procurement Fraud Detection
A major source of corruption is the procurement
process
Laws attempt to ensure a competitive and fair process
Perpetrators find ways to turn the process to their
advantage while appearing to be legitimate
Specialist from CGU (Mário Spinelli)
Structured the different kinds of procurement frauds found
in the past years
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 11
31. Procurement Fraud Detection
Types of fraud
Characterized by criteria
Principle of competition is violated when we have
Owners who work as a front (usually someone with little or no education)
Use of accounting indices that are not common
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 12
32. Procurement Fraud Detection
Types of fraud
Characterized by criteria
Principle of competition is violated when we have
Owners who work as a front (usually someone with little or no education)
Use of accounting indices that are not common
Ultimate goal
Structure the specialist knowledge in a way that an
automated system can reason with the evidence in a manner
similar to the specialist
Support current specialists
Train new ones
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 12
34. Procurement Fraud Detection
Realistic goal (this paper)
Proof of concept
Selected just a few criteria
Why Semantic Web?
Propose an overall architecture for collecting data, reasoning
with uncertainty (model designed), and reporting alerts
Ask specialists to analyze results (subjective)
No massive data used
Show that new criteria can be easily incorporated
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 13
35. Why Semantic Web?
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
36. Why Semantic Web?
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
37. Why Semantic Web?
audits and inspections
(Procurement)
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
38. Why Semantic Web?
audits and inspections
(Procurement) socio-economic
(Person)
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
39. Why Semantic Web?
audits and inspections
(Procurement) socio-economic
(Person)
criminal history
(Person)
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
40. Why Semantic Web?
audits and inspections
(Procurement) socio-economic
(Person)
? criminal history
(Person) (Person)
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion
14
* AAA, open world, and nonunique naming - RIS environment
43. Architecture
Informaion Gathering
Public Notices - Data
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
44. Architecture
Informaion Gathering
DB - Information
Public Notices - Data
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
45. Architecture
Informaion Gathering
DB - Information
Public Notices - Data Design - UnBBayes
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
46. Architecture
Informaion Gathering
DB - Information
Public Notices - Data Design - UnBBayes
Inference - Knowledge
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
47. Architecture
Informaion Gathering
DB - Information
Public Notices - Data Design - UnBBayes
Report for Decision Makers Inference - Knowledge
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
48. Architecture
Informaion Gathering
DB - Information
Public Notices - Data Design - UnBBayes
Report for Decision Makers Inference - Knowledge
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 15
54. Results
Non suspect procurement:
0.01% that the procurement was directed to a specific company by using accounting
indices;
0.10% that the procurement was directed to a specific company.
Suspect procurement:
55.00% that the procurement was directed to a specific company by using accounting
indices;
29.77%, when the information about demanding experience in only one contract was
omitted, and 72.00%, when it was given, that the procurement was directed to a specific
company.
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 20
59. Conclusion
Correct conclusion for both suspicious and non-
suspicious cases
Results are encouraging
Suggesting that a fuller development of our proof of concept is promising
Needs more testing, especially with real data for validating the conclusions
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 24
60. Conclusion
Correct conclusion for both suspicious and non-
suspicious cases
Results are encouraging
Suggesting that a fuller development of our proof of concept is promising
Needs more testing, especially with real data for validating the conclusions
Advantages
Impartiality in the judgment
Scalability
Joint analysis of large volumes of indicators
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 24
62. Conclusion
Future work
Choose/add new criteria
Collect more data for validation of the model
Will probably required fusion of data from different
agencies
Good for assessing the usefulness of ontologies and the SW
Introduction - MEBN and PR-OWL - Procurement Fraud Detection - Conclusion 25