Опис
http://dataconf.com.ua/speaker-page/andrii-malenko.php
Відео
https://www.youtube.com/watch?v=tBgNBeO5-rA&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=13
This document discusses 5 learnings from using Clojure at a text analytics startup that processes 3 million documents per day. It notes that Clojure provides great operability and powerful yet simple parallelism tools. However, using Clojure alone does not guarantee good design and the language lacks a relevant machine learning ecosystem compared to Python. Overall, Clojure is concluded to be one of the best multi-purpose languages but not a silver bullet on its own.
Clojure's three killer design decisions that boosted its adoption and maturity were:
1) Solving everyday problems through its small core and terse syntax which addresses boilerplate code and provides primitives for concurrency and parallelism.
2) Being hosted on the JVM to leverage the trusted and performant runtime as well as the large Java ecosystem.
3) Having effective governance through significant innovations within the first 10 years and an openness to adopting ideas from other communities.
OpenPonk (formerly DynaCASE). The open modeling platformESUG
OpenPonk (formerly DynaCASE) is an open modeling platform for conceptual modeling. It allows for diagramming, live modeling where changes are immediately visible, simulations for more visual modeling, bi-directional DSLs to link models and text, validations to ensure model quality, code generation from models, and is fully open source. The document introduces OpenPonk, discusses conceptual modeling, and outlines several features of the modeling platform.
Roberto Navigli - From Text to Concepts and Back: Going Multilingual with Bab...MeetupDataScienceRoma
The document discusses Roberto Navigli giving a talk on BabelNet, a multilingual semantic network created by merging various knowledge resources, and Babelfy, a state-of-the-art multilingual word sense disambiguation system that leverages BabelNet. The talk outlines BabelNet and Babelfy, demonstrates Babelfy for word sense disambiguation, and discusses how working with BabelNet provides coverage of numerous resources by annotating with their combined knowledge.
Multilingual Access to Cultural Heritage Content on the Semantic Web - Acl2013Mariana Damova, Ph.D
The document discusses building an ontology-based application to communicate museum content in multiple languages on the Semantic Web. It aims to make cultural heritage accessible to both humans and computers by generating natural language descriptions from semantic data. The application uses Grammatical Framework to linearly multiple museum datasets and ontologies into 15 languages. It addresses challenges in cross-linguistically representing classes, properties, word order, tense, and reference. The system was demonstrated to generate descriptions of paintings from the Louvre museum in English and French.
Modeling software systems at a macroscopic scaleRalf Laemmel
This document discusses megamodeling, which involves creating models that represent other models, languages, and technologies used in software projects. A megamodel captures the relationships between these different elements at a macro scale. The document proposes using MegaL, a general-purpose megamodeling language, to precisely model software systems in terms of the languages, technologies, and concepts involved and their relationships. MegaL aims to help manage diversity and heterogeneity in software while providing cognitive value through abstract understanding and documentation of designs.
The Global Environment Facility (GEF) initiated the International Waters: Learning Exchange and Resource Network (IW:LEARN) project in 1997 to enhance knowledge sharing for GEF international waters projects. IW:LEARN is a UNEP/UNDP project that works to improve online sharing of data and information relevant to managing international waters. The IW:LEARN team is spread globally and maintains a website to serve as an information hub, document repository, and portal for learning materials and project visualization.
This document discusses 5 learnings from using Clojure at a text analytics startup that processes 3 million documents per day. It notes that Clojure provides great operability and powerful yet simple parallelism tools. However, using Clojure alone does not guarantee good design and the language lacks a relevant machine learning ecosystem compared to Python. Overall, Clojure is concluded to be one of the best multi-purpose languages but not a silver bullet on its own.
Clojure's three killer design decisions that boosted its adoption and maturity were:
1) Solving everyday problems through its small core and terse syntax which addresses boilerplate code and provides primitives for concurrency and parallelism.
2) Being hosted on the JVM to leverage the trusted and performant runtime as well as the large Java ecosystem.
3) Having effective governance through significant innovations within the first 10 years and an openness to adopting ideas from other communities.
OpenPonk (formerly DynaCASE). The open modeling platformESUG
OpenPonk (formerly DynaCASE) is an open modeling platform for conceptual modeling. It allows for diagramming, live modeling where changes are immediately visible, simulations for more visual modeling, bi-directional DSLs to link models and text, validations to ensure model quality, code generation from models, and is fully open source. The document introduces OpenPonk, discusses conceptual modeling, and outlines several features of the modeling platform.
Roberto Navigli - From Text to Concepts and Back: Going Multilingual with Bab...MeetupDataScienceRoma
The document discusses Roberto Navigli giving a talk on BabelNet, a multilingual semantic network created by merging various knowledge resources, and Babelfy, a state-of-the-art multilingual word sense disambiguation system that leverages BabelNet. The talk outlines BabelNet and Babelfy, demonstrates Babelfy for word sense disambiguation, and discusses how working with BabelNet provides coverage of numerous resources by annotating with their combined knowledge.
Multilingual Access to Cultural Heritage Content on the Semantic Web - Acl2013Mariana Damova, Ph.D
The document discusses building an ontology-based application to communicate museum content in multiple languages on the Semantic Web. It aims to make cultural heritage accessible to both humans and computers by generating natural language descriptions from semantic data. The application uses Grammatical Framework to linearly multiple museum datasets and ontologies into 15 languages. It addresses challenges in cross-linguistically representing classes, properties, word order, tense, and reference. The system was demonstrated to generate descriptions of paintings from the Louvre museum in English and French.
Modeling software systems at a macroscopic scaleRalf Laemmel
This document discusses megamodeling, which involves creating models that represent other models, languages, and technologies used in software projects. A megamodel captures the relationships between these different elements at a macro scale. The document proposes using MegaL, a general-purpose megamodeling language, to precisely model software systems in terms of the languages, technologies, and concepts involved and their relationships. MegaL aims to help manage diversity and heterogeneity in software while providing cognitive value through abstract understanding and documentation of designs.
The Global Environment Facility (GEF) initiated the International Waters: Learning Exchange and Resource Network (IW:LEARN) project in 1997 to enhance knowledge sharing for GEF international waters projects. IW:LEARN is a UNEP/UNDP project that works to improve online sharing of data and information relevant to managing international waters. The IW:LEARN team is spread globally and maintains a website to serve as an information hub, document repository, and portal for learning materials and project visualization.
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...Yuki Tomo
12/22 Deep Learning勉強会@小町研 にて
"Learning Character-level Representations for Part-of-Speech Tagging" C ́ıcero Nogueira dos Santos, Bianca Zadrozny
を紹介しました。
IW:LEARN is a joint UNEP/UNDP project which works with GEF (Gloabal Envrionment Facility) International Waters projects to improve online sharing of data and information relevant to managing international waters, including marine, coastal and freshwater ecosystems.
Event Details
The IW:LEARN website acts a document clearinghouse and information and news hub for the GEF IW Projects. IW:LEARN also supplies Plone training and hosting for projects.
The IW:LEARN Team is spread around the globe (Bratislava, Nairobi, Bangkok and formerly Washington DC).
This talk will give an overview of the challenges and experiences using Plone in an internationally distributed project with requirements that are a good match for Plones capabilities and exceed what Plone offers out of the box.
For the needs of our content editors we have developed a set of tools that make metadata entry and extraction easier to encourage them to provide better quality metadata. This improves the User experience and discover-ability.
collective.ots to extract a meaningful description from the content
collective.simserver to automatically relate similar items
collective.langdet to determine the language of content
IW:LEARN integrates and visualizes geospatial information about the projects and their area of work which uses collective.geo extensively. In this talk we will show how we use Choropleth and Cluster maps for thematic mapping with collective.geo.
We also provide services for which Plone is not a good match so we use other technologies as well such as KARL (Knowledge management and sharing) or GeoNode (Sharing of geospatial data)
This document discusses contextual word embeddings and how they address the limitations of context-free word embeddings. It begins by explaining that context-free word embeddings cannot model polysemy since words have the same embedding regardless of context. It then introduces contextual word embeddings as a solution, discussing early approaches like CoVe and ELMo that learn contextual embeddings from language models. The document emphasizes that contextual embeddings allow words to have different representations depending on the surrounding context.
The Datalift Project aims to publish and interconnect government open data. It develops tools and methodologies to transform raw datasets into interconnected semantic data. The project's first phase focuses on opening data by developing an infrastructure to ease publication. The second phase will validate the platform by publishing real datasets. The goal of Datalift is to move data from its raw published state to being fully interconnected on the Semantic Web.
This document discusses Universal Design for Learning (UDL) and how Web 2.0 tools can support its three principles of multiple means of engagement, representation, and expression. It provides examples of several free Web 2.0 tools that meet UDL criteria by allowing flexible learning solutions and collaboration. These include Wallwisher for online corkboards, Glogster for multimedia presentations, and Piratepad for shared document editing. Contact information is given for further resources on UDL and Web 2.0 tools.
WIDOCO: A Wizard for Documenting Ontologiesdgarijo
WIDOCO is a WIzard for DOCumenting Ontologies that guides users through the documentation process of their vocabularies. Given an RDF vocabulary, WIDOCO detects missing vocabulary metadata and creates a documentation with diagrams, human readable descriptions of the ontology terms and a summary of
changes with respect to previous versions of the ontology. The documentation consists on a set of linked enriched HTML pages that can be further extended by end users. WIDOCO is open source and builds on well established Semantic Web tools. So far, it has been used to document more than one hundred ontologies in different domains.
A Controlled Natural Language Interface for Semantic MediaWikiJie Bao
This document proposes using a controlled natural language interface for the semantic wiki Semantic MediaWiki. It aims to improve usability and expressivity. Key points:
- Using a controlled natural language instead of formal logic improves ease of use for non-experts and allows knowledge input without thinking in "subject-property-object" terms.
- An ontology meta-model extends Semantic MediaWiki to support the full range of OWL/RDF constructs like class domains and ranges.
- Forms, templates, and a natural language generation module allow editing knowledge in controlled natural language and translating between the wiki, ontology meta-model, and RDF formats.
- The approach supports multiple controlled natural
Oxford University Press (OUP) publishes dictionaries and other reference works in over 70 languages. It aims to further the University of Oxford's research and education through worldwide publishing. OUP provides dictionary content and data to power various digital applications, including tools that track word frequencies over time, analyze text, map metaphors, and visualize the development of specialized vocabularies like surfing terms. Its datasets cover dozens of languages and include information on headwords, definitions, and etymologies. OUP seeks to enable rich language communication globally and reduce the "digital divide" through its language solutions and datasets.
This document discusses static analysis of programs written in domain-specific languages (DSLs) developed with Xtext. It presents a model-driven strategy to bridge the gap between Xtext and SonarQube grammar formats, allowing quality analysis of DSL programs. The strategy involves automatically generating the Java code infrastructure for a new language in SonarQube based on the Xtext grammar. An example DSL for teaching programming called Vary is presented to demonstrate the approach.
Analyzing the Evolution of Vocabulary Terms and Their Impact on the LOD CloudMOVING Project
Vocabularies are used for modeling data in Knowledge Graphs
(KGs) like the Linked Open Data Cloud and Wikidata. During their life-time, vocabularies are subject to changes. New terms are coined while existing terms are modified or deprecated. We first quantify the amount and frequency of changes in vocabularies. Subsequently, we investigate to which extend and when the changes are adopted in the evolution of KGs.
We conduct our experiments on three large-scale KGs for which time-stamped information is available, namely the Billion Triples Challenge datasets, Dynamic Linked Data Observatory dataset, and Wikidata. Our results show that the change frequency of terms is rather low, but can have high impact due to the large amount of distributed graph data on the web. Furthermore, not all coined terms are used and most of the
deprecated terms are still used by data publishers. The adoption time of terms coming from different vocabularies ranges from very fast (few days) to very slow (few years). Surprisingly, we could observe some adoptions before the vocabulary changes were published. Understanding the evolution of vocabulary terms is important to avoid wrong assumptions about the modeling status of data published on the web, which may result in difficulties when querying the data from distributed sources.
Can Deep Learning Techniques Improve Entity Linking?Julien PLU
Julien Plu presented on using deep learning techniques to improve entity linking. He discussed using word embeddings and neural networks to better recognize and link entities in documents by understanding semantic relationships between words. Current supervised methods require large training sets and are not robust to new entity types, while unsupervised methods have difficulties computing relatedness between candidate entities. Deep learning approaches may help address these issues through their ability to learn complex patterns from large amounts of unlabeled text data.
Video game controlled vocabulary in wikidatapeterchanws
Peter Chan is a digital archivist at Stanford Libraries who has led several projects related to preserving born-digital archival materials and developing software to support processing digital archives. He is involved with organizations working on video game preservation and proposed using Wikidata to publish controlled vocabularies and game metadata. The presentation outlined steps to include the OLAC video game genre vocabulary in Wikidata, including proposing it as an external identifier property, creating missing terms, and adding broader relationships between concepts. Issues with Wikidata like lack of control over edits and the need for multiple models were also discussed.
This presentation begins with a specific issue in text mining that connect it with word embeddings. Later, the importance of the Wikipedia is highlighted and finally, lessons to be learned from the Wikipedia are discussed.
This document summarizes an agenda for a responsive design roundtable discussion. The roundtable will cover topics including what responsive design is, the user continuum, and mobile web vs responsive design vs apps. It will take place on June 12, 2013 from 8am to 10am and be led by Christian Glover Wilson. The agenda includes 4 topics to discuss and time for Q&A.
The document discusses the latest programming languages as of 2011-2012. It provides an example report format listing Boo and D as two of the latest languages. Boo, developed in 2011, is a general-purpose language for the .NET platform seeking to support Unicode and web applications. D, developed in 2012, is a general systems language for Unix-like, Windows, and Mac OS X platforms using an object-oriented approach and compiler.
Overview of text classification approaches algorithms & software v lyubin...Olga Zinkevych
The main points of the presentation:Overview of text classification approaches: algorithms & software
Summary: For the last 2 month I've been building a system for classifying customer support tickets into several categories in terms of product area, importance, etc. Throughout that time I've tried several approaches and benchmarked them against each other. In this talk I would like to showcase some of my findings, including research algorithms that perform well and relevant software. This talk would be useful for someone who needs to build a text categorization system, or someone who just wants to get an overview of one of the most popular NLP research problems (classification).
In this talk you will learn:
* About various approaches used for text classification (e.g. approaches based on TF-IDF, or approaches based on word embeddings and RNNs - recurrent neural nets).
* How these approaches perform against each other on a real-world data.
* Software that is useful for implementing these approaches.
* Research behind some of these approaches.
http://dataconf.com.ua/speaker-page/volodymyr-lyubinets.php
https://www.youtube.com/watch?v=shmc-MI-xbo&index=5&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu
What it takes to build a model for detecting patients that defaults from medi...Olga Zinkevych
Topic of presentation: What it takes to build a model for detecting patients that defaults from medication
The main points of the presentation:
Why data exploration is important?
Clean data is a half of success
why subject knowledge experts are crucial in healthcare project
feature engineerings as a way to make you model more accurate
We will talk about how using clinical data tyr to predict if patients will or will not defect from their medication.
http://dataconf.com.ua/speaker-page/jaya-plmanabhan.php
https://www.youtube.com/watch?v=vjvwzhyLOX4&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=7
http://dataconf.com.ua/speaker-page/khrystyna-kosenko.php
More Related Content
Similar to Evolution of words through time a malenko dataconf 21 04_18
Deep Learning勉強会@小町研 "Learning Character-level Representations for Part-of-Sp...Yuki Tomo
12/22 Deep Learning勉強会@小町研 にて
"Learning Character-level Representations for Part-of-Speech Tagging" C ́ıcero Nogueira dos Santos, Bianca Zadrozny
を紹介しました。
IW:LEARN is a joint UNEP/UNDP project which works with GEF (Gloabal Envrionment Facility) International Waters projects to improve online sharing of data and information relevant to managing international waters, including marine, coastal and freshwater ecosystems.
Event Details
The IW:LEARN website acts a document clearinghouse and information and news hub for the GEF IW Projects. IW:LEARN also supplies Plone training and hosting for projects.
The IW:LEARN Team is spread around the globe (Bratislava, Nairobi, Bangkok and formerly Washington DC).
This talk will give an overview of the challenges and experiences using Plone in an internationally distributed project with requirements that are a good match for Plones capabilities and exceed what Plone offers out of the box.
For the needs of our content editors we have developed a set of tools that make metadata entry and extraction easier to encourage them to provide better quality metadata. This improves the User experience and discover-ability.
collective.ots to extract a meaningful description from the content
collective.simserver to automatically relate similar items
collective.langdet to determine the language of content
IW:LEARN integrates and visualizes geospatial information about the projects and their area of work which uses collective.geo extensively. In this talk we will show how we use Choropleth and Cluster maps for thematic mapping with collective.geo.
We also provide services for which Plone is not a good match so we use other technologies as well such as KARL (Knowledge management and sharing) or GeoNode (Sharing of geospatial data)
This document discusses contextual word embeddings and how they address the limitations of context-free word embeddings. It begins by explaining that context-free word embeddings cannot model polysemy since words have the same embedding regardless of context. It then introduces contextual word embeddings as a solution, discussing early approaches like CoVe and ELMo that learn contextual embeddings from language models. The document emphasizes that contextual embeddings allow words to have different representations depending on the surrounding context.
The Datalift Project aims to publish and interconnect government open data. It develops tools and methodologies to transform raw datasets into interconnected semantic data. The project's first phase focuses on opening data by developing an infrastructure to ease publication. The second phase will validate the platform by publishing real datasets. The goal of Datalift is to move data from its raw published state to being fully interconnected on the Semantic Web.
This document discusses Universal Design for Learning (UDL) and how Web 2.0 tools can support its three principles of multiple means of engagement, representation, and expression. It provides examples of several free Web 2.0 tools that meet UDL criteria by allowing flexible learning solutions and collaboration. These include Wallwisher for online corkboards, Glogster for multimedia presentations, and Piratepad for shared document editing. Contact information is given for further resources on UDL and Web 2.0 tools.
WIDOCO: A Wizard for Documenting Ontologiesdgarijo
WIDOCO is a WIzard for DOCumenting Ontologies that guides users through the documentation process of their vocabularies. Given an RDF vocabulary, WIDOCO detects missing vocabulary metadata and creates a documentation with diagrams, human readable descriptions of the ontology terms and a summary of
changes with respect to previous versions of the ontology. The documentation consists on a set of linked enriched HTML pages that can be further extended by end users. WIDOCO is open source and builds on well established Semantic Web tools. So far, it has been used to document more than one hundred ontologies in different domains.
A Controlled Natural Language Interface for Semantic MediaWikiJie Bao
This document proposes using a controlled natural language interface for the semantic wiki Semantic MediaWiki. It aims to improve usability and expressivity. Key points:
- Using a controlled natural language instead of formal logic improves ease of use for non-experts and allows knowledge input without thinking in "subject-property-object" terms.
- An ontology meta-model extends Semantic MediaWiki to support the full range of OWL/RDF constructs like class domains and ranges.
- Forms, templates, and a natural language generation module allow editing knowledge in controlled natural language and translating between the wiki, ontology meta-model, and RDF formats.
- The approach supports multiple controlled natural
Oxford University Press (OUP) publishes dictionaries and other reference works in over 70 languages. It aims to further the University of Oxford's research and education through worldwide publishing. OUP provides dictionary content and data to power various digital applications, including tools that track word frequencies over time, analyze text, map metaphors, and visualize the development of specialized vocabularies like surfing terms. Its datasets cover dozens of languages and include information on headwords, definitions, and etymologies. OUP seeks to enable rich language communication globally and reduce the "digital divide" through its language solutions and datasets.
This document discusses static analysis of programs written in domain-specific languages (DSLs) developed with Xtext. It presents a model-driven strategy to bridge the gap between Xtext and SonarQube grammar formats, allowing quality analysis of DSL programs. The strategy involves automatically generating the Java code infrastructure for a new language in SonarQube based on the Xtext grammar. An example DSL for teaching programming called Vary is presented to demonstrate the approach.
Analyzing the Evolution of Vocabulary Terms and Their Impact on the LOD CloudMOVING Project
Vocabularies are used for modeling data in Knowledge Graphs
(KGs) like the Linked Open Data Cloud and Wikidata. During their life-time, vocabularies are subject to changes. New terms are coined while existing terms are modified or deprecated. We first quantify the amount and frequency of changes in vocabularies. Subsequently, we investigate to which extend and when the changes are adopted in the evolution of KGs.
We conduct our experiments on three large-scale KGs for which time-stamped information is available, namely the Billion Triples Challenge datasets, Dynamic Linked Data Observatory dataset, and Wikidata. Our results show that the change frequency of terms is rather low, but can have high impact due to the large amount of distributed graph data on the web. Furthermore, not all coined terms are used and most of the
deprecated terms are still used by data publishers. The adoption time of terms coming from different vocabularies ranges from very fast (few days) to very slow (few years). Surprisingly, we could observe some adoptions before the vocabulary changes were published. Understanding the evolution of vocabulary terms is important to avoid wrong assumptions about the modeling status of data published on the web, which may result in difficulties when querying the data from distributed sources.
Can Deep Learning Techniques Improve Entity Linking?Julien PLU
Julien Plu presented on using deep learning techniques to improve entity linking. He discussed using word embeddings and neural networks to better recognize and link entities in documents by understanding semantic relationships between words. Current supervised methods require large training sets and are not robust to new entity types, while unsupervised methods have difficulties computing relatedness between candidate entities. Deep learning approaches may help address these issues through their ability to learn complex patterns from large amounts of unlabeled text data.
Video game controlled vocabulary in wikidatapeterchanws
Peter Chan is a digital archivist at Stanford Libraries who has led several projects related to preserving born-digital archival materials and developing software to support processing digital archives. He is involved with organizations working on video game preservation and proposed using Wikidata to publish controlled vocabularies and game metadata. The presentation outlined steps to include the OLAC video game genre vocabulary in Wikidata, including proposing it as an external identifier property, creating missing terms, and adding broader relationships between concepts. Issues with Wikidata like lack of control over edits and the need for multiple models were also discussed.
This presentation begins with a specific issue in text mining that connect it with word embeddings. Later, the importance of the Wikipedia is highlighted and finally, lessons to be learned from the Wikipedia are discussed.
This document summarizes an agenda for a responsive design roundtable discussion. The roundtable will cover topics including what responsive design is, the user continuum, and mobile web vs responsive design vs apps. It will take place on June 12, 2013 from 8am to 10am and be led by Christian Glover Wilson. The agenda includes 4 topics to discuss and time for Q&A.
The document discusses the latest programming languages as of 2011-2012. It provides an example report format listing Boo and D as two of the latest languages. Boo, developed in 2011, is a general-purpose language for the .NET platform seeking to support Unicode and web applications. D, developed in 2012, is a general systems language for Unix-like, Windows, and Mac OS X platforms using an object-oriented approach and compiler.
Overview of text classification approaches algorithms & software v lyubin...Olga Zinkevych
The main points of the presentation:Overview of text classification approaches: algorithms & software
Summary: For the last 2 month I've been building a system for classifying customer support tickets into several categories in terms of product area, importance, etc. Throughout that time I've tried several approaches and benchmarked them against each other. In this talk I would like to showcase some of my findings, including research algorithms that perform well and relevant software. This talk would be useful for someone who needs to build a text categorization system, or someone who just wants to get an overview of one of the most popular NLP research problems (classification).
In this talk you will learn:
* About various approaches used for text classification (e.g. approaches based on TF-IDF, or approaches based on word embeddings and RNNs - recurrent neural nets).
* How these approaches perform against each other on a real-world data.
* Software that is useful for implementing these approaches.
* Research behind some of these approaches.
http://dataconf.com.ua/speaker-page/volodymyr-lyubinets.php
https://www.youtube.com/watch?v=shmc-MI-xbo&index=5&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu
What it takes to build a model for detecting patients that defaults from medi...Olga Zinkevych
Topic of presentation: What it takes to build a model for detecting patients that defaults from medication
The main points of the presentation:
Why data exploration is important?
Clean data is a half of success
why subject knowledge experts are crucial in healthcare project
feature engineerings as a way to make you model more accurate
We will talk about how using clinical data tyr to predict if patients will or will not defect from their medication.
http://dataconf.com.ua/speaker-page/jaya-plmanabhan.php
https://www.youtube.com/watch?v=vjvwzhyLOX4&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=7
http://dataconf.com.ua/speaker-page/khrystyna-kosenko.php
Topic of presentation: Variational autoencoders for speech processing
The main points of the presentation: Variational autoencoders (or VAE) have become one of the most popular unsupervised learning techniques for modelling complex data distributions, such as images and audio. In this talk I'll begin with a general introduction to VAEs and then review a recent technique called VQ-VAE which is capable of learning rundimentary phoneme-level language model from raw audio without any supervision.
http://dataconf.com.ua/speaker-page/dmytro-bielievtsov.php
https://www.youtube.com/watch?v=euYSAL-aKMI&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=9
Dataservices based on mesos and kafka kostiantyn bokhan dataconf 21 04 18Olga Zinkevych
Topic of presentation: Dataservices based on mesos and kafka
The main points of the presentation: У своїй доповіді Костянтин поділіться досвідом побудови датасервісів на основі таких технологій як: Kafka, Docker, Mesos, Aerospike та Spark. Будуть розглянуті наступні питання: оркестрація, ізоляція, управляння ресурсами, service discovery and load balancing, взаємодія датасервісів. Будуть обговорені проблеми управління ресурсами Java-based та Spark-based сервісів під керування mesos кластера, а також реалізація CI та CD датасервісів.
*CI - continuous integration, CD - continuous delivery
http://dataconf.com.ua/speaker-page/kostiantyn-bokhan.php
https://www.youtube.com/watch?v=4d41DDyKuwU&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=3
Azure data catalog your data your way eugene polonichko dataconf 21 04 18Olga Zinkevych
Topic of presentation: Azure Data Catalog: your data, your way
The main points of the presentation:It’s a fully-managed service that lets you—from analyst to data scientist to data developer—register, enrich, discover, understand, and consume data sources
http://dataconf.com.ua/speaker-page/eugene-polonichko.php
https://www.youtube.com/watch?v=wceGzcQcPOo&list=PL5_LBM8-5sLjbRFUtXaUpg84gtJtyc4Pu&t=0s&index=4
Aibdconference chat bot for every product Maksym VolchenkoOlga Zinkevych
This document discusses conversational interfaces and chatbots. It begins with an introduction to the author's background and experience in artificial intelligence and as an Android developer. It then discusses why conversational interfaces are becoming more popular as people prefer interacting with products and services through human conversation versus many separate apps. Chatbots provide a cross-platform solution for this using natural language processing. The document defines key terms like bots, chatbots, virtual assistants and describes common NLP techniques. It provides examples of chatbot architectures and development tools like API.ai and discusses analytics. It concludes that while conversing with bots is interesting, human interaction is more meaningful.
Ai&bigdataconference oleksandr saienko machine learning use cases in telecomOlga Zinkevych
Topic of presentation: Machine Learning use cases in Telecom
The main points of the presentation: Oleksandr will talk about some interesting examples of using Machine Learning in Telecom: optimizing the cellular network, improving customer experience, models for mobile devices location predicting, customer churn prediction, detecting fraud and others. He will consider the main modern approaches based no machine learning.
http://dataconf.com.ua/oleksandr-saienko.php
Ai big dataconference_volodymyr getmanskyi colorization distance measuringOlga Zinkevych
Topic of presentation: Deep learning for satellite imagery colorization and distance measuring.
The main points of the presentation:
Using modern techniques we compared existing methods for colorization from the perspective of satelite maps. After this we built our own engine for measuring the distances on the maps.
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_taras firman how to build advanced prediction with addi...Olga Zinkevych
Topic of presentation: How to build advanced prediction with adding external data.
The main points of the presentation:
We will discuss different types of time series, main approaches for building forecasting, how to work with missing data, how to add external data using Machine Learning technics. After that, we will consider existing Python's forecasting libraries
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_sparkinonehour_vitalii bashunOlga Zinkevych
Topic of presentation: First Spark application in one hour
Are you a beginner in Big Data world? Do not know where to start from? This session is for you. Introduction to distributed computations, Hadoop and the most popular and powerful framework in Big Data world - Apache Spark. This session aims to explain Big Data from scratch in simple words and to show how you can write and run your first Spark application in one hour.
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_semantic image segmentatation using word embeddings_ole...Olga Zinkevych
Topic of presentation: Semantic image segmentation using word embeddings
The main points of the presentation:
Semantic image segmentation
Word embeddings
Unsupervised learning
Object detection
Multimodal learning
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_ml_fastdata_vitalii bondarenkoOlga Zinkevych
This document discusses machine learning on fast data. It presents an agenda covering ML on production systems, TensorFlow, Kafka, Docker and Kubernetes. It then describes the machine learning process and shows how an enterprise analytics platform can integrate data sources, a machine learning cluster using Kafka, and data destinations. It provides details on TensorFlow and how it can be used for linear regression and neural networks. It also explains Apache Kafka as a streaming data service bus and how Confluent Platform extends it. Finally, it briefly introduces Docker and Kubernetes.
Ai big dataconference_krakovetskyi_microsoft ai a new era of smart solutionsOlga Zinkevych
Topic of presentation: Microsoft AI: a new era of smart solutions
The main points of the presentation: In this presentation we will talk about Microsoft's tools and products that will add some intelligence to your apps and solutions.We will talk about Cognitive Services, chatbots, Cortana and Alexa, Deep Learning and Azure Machine Learning.
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_jeffrey ricker_kappa_architectureOlga Zinkevych
Topic of presentation: Kappa architecture (and beyond)
The main points of the presentation:
We will discuss the evolution of big data architecture, from batch to Lambda to Kappa. I will walk through how to implement a Kappa architecture with practical examples, focusing on how to reach full potential and avoid the pitfalls. We will finish with reviewing what lies ahead, including the inevitable consolidation between microservices, GPGPU and Hadoop.
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Ai big dataconference_eugene_polonichko_azure data lake Olga Zinkevych
Topic of presentation: Azure Data Lake: what is it? why is it? where is it?
The main points of the presentation:
What is Azure Data Lake? Why does this technology call Microsoft Big Data? Azure Data Lake includes all the capabilities required to make it easy for developers, data scientists, and analysts to store data of any size, shape, and speed, and do all types of processing and analytics across platforms and languages. It removes the complexities of ingesting and storing all of your data while making it faster to get up and running with batch, streaming, and interactive analytics.
http://dataconf.com.ua/index.php#agenda
#dataconf
#AIBDConference
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
A review of the growth of the Israel Genealogy Research Association Database Collection for the last 12 months. Our collection is now passed the 3 million mark and still growing. See which archives have contributed the most. See the different types of records we have, and which years have had records added. You can also see what we have for the future.
How to Fix the Import Error in the Odoo 17Celine George
An import error occurs when a program fails to import a module or library, disrupting its execution. In languages like Python, this issue arises when the specified module cannot be found or accessed, hindering the program's functionality. Resolving import errors is crucial for maintaining smooth software operation and uninterrupted development processes.
2. The idea
● Meaning of words in natural language change slowly but surely
● Word’s meaning can be estimated by context or nearest neighbours
occuring in texts
● Language model represent relations between words in numeric way
● Change of language model shows evolution of natural language in
different time periods
● New concepts in mass consciousness appear as a synthesis of
previous concepts