Here are the intro to OpenCalais workshop slides presented at WeMedia 2010 in Miami (2.10.10) by Krista Thomas of The Thomson Reuters OpenCalais Initiative.
Opportunities for AI in Intelligent Web-based Technology-Supported LearningCarsten Ullrich
Talk I gave at WISM’09-AICI'09 (2009 International Conference on Web Information Systems and Mining and the 2009 International Conference on Artificial Intelligence and Computational Intelligence) where I describe research opportunities offered by today's Web 2.0 and Semantic Web.
Knowledge Infrastructure for Global Systems ScienceDavid De Roure
Presentation at the First Open Global Systems Science Conference, Brussels, 8-10 November 2012
http://www.gsdp.eu/nc/news/news/date/2012/10/31/first-open-global-systems-science-conference/
Here are the intro to OpenCalais workshop slides presented at WeMedia 2010 in Miami (2.10.10) by Krista Thomas of The Thomson Reuters OpenCalais Initiative.
Opportunities for AI in Intelligent Web-based Technology-Supported LearningCarsten Ullrich
Talk I gave at WISM’09-AICI'09 (2009 International Conference on Web Information Systems and Mining and the 2009 International Conference on Artificial Intelligence and Computational Intelligence) where I describe research opportunities offered by today's Web 2.0 and Semantic Web.
Knowledge Infrastructure for Global Systems ScienceDavid De Roure
Presentation at the First Open Global Systems Science Conference, Brussels, 8-10 November 2012
http://www.gsdp.eu/nc/news/news/date/2012/10/31/first-open-global-systems-science-conference/
The Semantic Web is a mesh of information linked up in such a way as to be easily processable by machines, on a global scale. You can think of it as being an efficient way of representing data on the World Wide Web, or as a globally linked database.
Building a distributed search system with Hadoop and LuceneMirko Calvaresi
This work analyses the problem coming from the so called Big Data scenario, which can be defined as the technological challenge to manage and administer quantity of information with global dimension in the order of Terabyte (〖10〗^12 bytes) or Petabyte (〖10〗^15 bytes) and with an exponential growth rate.
We’ll explore a technological and algorithmic approach to handle and calculate theses amounts of data that exceed the limit of computation of a traditional architecture based on real-time request processing: in particular we’ll analyze a singular open source technology, called Apache Hadoop, which implements the approach described as Map and Reduce.
We’ll describe also how to distribute a cluster of common server to create a Virtual File System and use this environment to populate a centralized search index (realized using another open source technology, called Apache Lucene).
The practical implementation will be a web based application which offers to the user a unified searching interface against a collection of technical papers.
The scope is to demonstrate that a performant search system can be obtained pre-processing the data using the Map and Reduce paradigm, in order to obtain a real time response, which is independent to the underlying amount of data.
Finally we’ll compare this solutions to different approaches based on clusterization or No SQL solutions, with the scope to describe the characteristics of concrete scenarios, which suggest the adoption of those technologies.
Elasticsearch : petit déjeuner du 13 mars 2014ALTER WAY
Elasticsearch est un moteur de recherche Open Source très puissant basé sur
Apache Lucene. Il permet l'indexation de millions de données, leur recherche et leur
analyse en temps réel. Les outils Elascticsearch sont déjà utilisés par des acteurs de
référence tels que FourSquare, GitHub, OpenDataSoft ou encore Dailymotion.
Alter Way et Elasticsearch vous convient à venir découvrir la suite Elasticsearch
enfin disponible en version 1.0 et prête pour la production !
See the WEBCAST as well!! mms://wmedia.it.su.se/SUB/NordLib/3.wmv
Presentation at Nordlib 2.0 in Stockholm, November 21th 2008
http://www.nordlib20.org/programme/
Built to grow: scalability factors to consider before commencing your next di...Marcus Emmanuel Barnes
Discussion piece on scalability factors to consider before commencing your next digital library software project presented at the Third Annual Code4Lib on November 27, 2015 at the UBC Irving K. Barber Learning Centre.
2012 03-28 Wf4ever, preserving workflows as digital research objectsStian Soiland-Reyes
Presented on 2012-03-28 at EGI Community Forum 2012, Munich.
http://www.wf4ever-project.org/
http://purl.org/wf4ever/model
http://cf2012.egi.eu/
https://www.egi.eu/indico/sessionDisplay.py?sessionId=66&confId=679#20120328
Using schema.org to improve SEO presented at DrupalCamp Asheville in August 2014.
http://drupalasheville.com/drupal-camp-asheville-2014/sessions/using-schemaorg-improve-seo
Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
Information Extraction from Text, presented @ DeloitteDeep Kayal
Useful unstructured text occurs in plentiful amounts, and often is central to the success of a business. The benefits of being able to successfully decipher unstructured text can be direct or derived. Companies which offer products for medical differential diagnosis are directly benefitted by the ability to correctly extract drug-disease interactions from publications, for example. As for derived benefits of text processing, we need to look no further than cases of improving process flows by analyzing the sentiment of the emails a company receives from its customers.
Being at the frontier of natural language processing, information representation and retrieval, information extraction has been the subject of extensive research for several decades and there are plenty of existing techniques to help with the understanding of unstructured textual content. This presentation will introduce and summarize useful techniques that are helpful in tackling sub-domains of information extraction, such as named entity recognition, keyword extraction and document summarization for efficient retrieval. Additionally, the talk will also emphasize low-resource cases, when not much useful labelled information is available.
Big Data Analysis : Deciphering the haystack Srinath Perera
A primary outcome of Bigdata is to derive useful and actionable insights from large or challenges data collections. The goal is to run the transformations from data, to information, to knowledge, and finally to insights. This includes calculating simple analytics like Mean, Max, and Median, to derive overall understanding about data by building models, and finally to derive predictions from data. Some cases we can afford to wait to collect and processes them, while in other cases we need to know the outputs right away. MapReduce has been the defacto standard for data processing, and we will start our discussion from there. However, that is only one side of the problem. There are other technologies like Apache Spark and Apache Drill graining ground, and also realtime processing technologies like Stream Processing and Complex Event Processing. Finally there are lot of work on porting decision technologies like Machine learning into big data landscape. This talk discusses big data processing in general and look at each of those different technologies comparing and contrasting them.
Workplace-based Learning in Industry 4.0 -- Multi-perspective approaches and ...Carsten Ullrich
Keynote at the 3rd Annual International Conference of the Immersive Learning Research Network, iLRN 2017
Today’s shop floor, the area of a factory where operatives assemble products, is a complex and demanding work environment. The employed and produced technology becomes ever more complex, and employees are responsible for an increasing amount of tasks. As a consequence, the employee is under constant pressure to solve problems occurring on the shop floor as fast as possible, and simultaneously to improve his work-related knowledge, skills, and capabilities. This keynotes presents the outcome of the APPsist project, which investigated how adaptive technology can support the employee on the shop floor in this challenging environment.
An Ontology for Learning Services on the Shop FloorCarsten Ullrich
An ontology expresses a common understanding of a domain that serves as a basis of communication between people or systems, and enables knowledge sharing, reuse of domain knowledge, reasoning and thus problem solving. In Technology-Enhanced Learning, especially in Intelligent Tutoring Systems and Adaptive Learning Environments, ontologies serve as the basis of adaptivity and personalization. For mathematics learning and similarly structured domains, ontologies and their usage for adaptive learning are well understood and established. This contribution presents an ontology for the industrial shop floor (the area of a factory where operatives assemble products) and illustrates its usage in several learning services.
More Related Content
Similar to Rapid Prototyping of a Semantic-Web-based Research Workbench
The Semantic Web is a mesh of information linked up in such a way as to be easily processable by machines, on a global scale. You can think of it as being an efficient way of representing data on the World Wide Web, or as a globally linked database.
Building a distributed search system with Hadoop and LuceneMirko Calvaresi
This work analyses the problem coming from the so called Big Data scenario, which can be defined as the technological challenge to manage and administer quantity of information with global dimension in the order of Terabyte (〖10〗^12 bytes) or Petabyte (〖10〗^15 bytes) and with an exponential growth rate.
We’ll explore a technological and algorithmic approach to handle and calculate theses amounts of data that exceed the limit of computation of a traditional architecture based on real-time request processing: in particular we’ll analyze a singular open source technology, called Apache Hadoop, which implements the approach described as Map and Reduce.
We’ll describe also how to distribute a cluster of common server to create a Virtual File System and use this environment to populate a centralized search index (realized using another open source technology, called Apache Lucene).
The practical implementation will be a web based application which offers to the user a unified searching interface against a collection of technical papers.
The scope is to demonstrate that a performant search system can be obtained pre-processing the data using the Map and Reduce paradigm, in order to obtain a real time response, which is independent to the underlying amount of data.
Finally we’ll compare this solutions to different approaches based on clusterization or No SQL solutions, with the scope to describe the characteristics of concrete scenarios, which suggest the adoption of those technologies.
Elasticsearch : petit déjeuner du 13 mars 2014ALTER WAY
Elasticsearch est un moteur de recherche Open Source très puissant basé sur
Apache Lucene. Il permet l'indexation de millions de données, leur recherche et leur
analyse en temps réel. Les outils Elascticsearch sont déjà utilisés par des acteurs de
référence tels que FourSquare, GitHub, OpenDataSoft ou encore Dailymotion.
Alter Way et Elasticsearch vous convient à venir découvrir la suite Elasticsearch
enfin disponible en version 1.0 et prête pour la production !
See the WEBCAST as well!! mms://wmedia.it.su.se/SUB/NordLib/3.wmv
Presentation at Nordlib 2.0 in Stockholm, November 21th 2008
http://www.nordlib20.org/programme/
Built to grow: scalability factors to consider before commencing your next di...Marcus Emmanuel Barnes
Discussion piece on scalability factors to consider before commencing your next digital library software project presented at the Third Annual Code4Lib on November 27, 2015 at the UBC Irving K. Barber Learning Centre.
2012 03-28 Wf4ever, preserving workflows as digital research objectsStian Soiland-Reyes
Presented on 2012-03-28 at EGI Community Forum 2012, Munich.
http://www.wf4ever-project.org/
http://purl.org/wf4ever/model
http://cf2012.egi.eu/
https://www.egi.eu/indico/sessionDisplay.py?sessionId=66&confId=679#20120328
Using schema.org to improve SEO presented at DrupalCamp Asheville in August 2014.
http://drupalasheville.com/drupal-camp-asheville-2014/sessions/using-schemaorg-improve-seo
Keynote presentation delivered at ELAG 2013 in Gent, Belgium, on May 29 2013. Discusses Research Objects and the relationship to work my team has been involved in during the past couple of years: OAI-ORE, Open Annotation, Memento.
Information Extraction from Text, presented @ DeloitteDeep Kayal
Useful unstructured text occurs in plentiful amounts, and often is central to the success of a business. The benefits of being able to successfully decipher unstructured text can be direct or derived. Companies which offer products for medical differential diagnosis are directly benefitted by the ability to correctly extract drug-disease interactions from publications, for example. As for derived benefits of text processing, we need to look no further than cases of improving process flows by analyzing the sentiment of the emails a company receives from its customers.
Being at the frontier of natural language processing, information representation and retrieval, information extraction has been the subject of extensive research for several decades and there are plenty of existing techniques to help with the understanding of unstructured textual content. This presentation will introduce and summarize useful techniques that are helpful in tackling sub-domains of information extraction, such as named entity recognition, keyword extraction and document summarization for efficient retrieval. Additionally, the talk will also emphasize low-resource cases, when not much useful labelled information is available.
Big Data Analysis : Deciphering the haystack Srinath Perera
A primary outcome of Bigdata is to derive useful and actionable insights from large or challenges data collections. The goal is to run the transformations from data, to information, to knowledge, and finally to insights. This includes calculating simple analytics like Mean, Max, and Median, to derive overall understanding about data by building models, and finally to derive predictions from data. Some cases we can afford to wait to collect and processes them, while in other cases we need to know the outputs right away. MapReduce has been the defacto standard for data processing, and we will start our discussion from there. However, that is only one side of the problem. There are other technologies like Apache Spark and Apache Drill graining ground, and also realtime processing technologies like Stream Processing and Complex Event Processing. Finally there are lot of work on porting decision technologies like Machine learning into big data landscape. This talk discusses big data processing in general and look at each of those different technologies comparing and contrasting them.
Workplace-based Learning in Industry 4.0 -- Multi-perspective approaches and ...Carsten Ullrich
Keynote at the 3rd Annual International Conference of the Immersive Learning Research Network, iLRN 2017
Today’s shop floor, the area of a factory where operatives assemble products, is a complex and demanding work environment. The employed and produced technology becomes ever more complex, and employees are responsible for an increasing amount of tasks. As a consequence, the employee is under constant pressure to solve problems occurring on the shop floor as fast as possible, and simultaneously to improve his work-related knowledge, skills, and capabilities. This keynotes presents the outcome of the APPsist project, which investigated how adaptive technology can support the employee on the shop floor in this challenging environment.
An Ontology for Learning Services on the Shop FloorCarsten Ullrich
An ontology expresses a common understanding of a domain that serves as a basis of communication between people or systems, and enables knowledge sharing, reuse of domain knowledge, reasoning and thus problem solving. In Technology-Enhanced Learning, especially in Intelligent Tutoring Systems and Adaptive Learning Environments, ontologies serve as the basis of adaptivity and personalization. For mathematics learning and similarly structured domains, ontologies and their usage for adaptive learning are well understood and established. This contribution presents an ontology for the industrial shop floor (the area of a factory where operatives assemble products) and illustrates its usage in several learning services.
Rules for Adaptive Learning and Assistance on the Shop FloorCarsten Ullrich
Today’s shop floor, the area of a factory where operatives assemble products, is a complex and demanding work environment. The employed and produced technology becomes ever more complex, and employees are responsible for an increasing amount of tasks. As a consequence, the employee is under constant pressure to solve problems occurring on the shop floor as fast as possible, and simultaneously to improve his work-related knowledge, skills, and capabilities.
This paper gives an account how adaptive technology can support the employee on the shop floor in these respects. It describes the organizational requirements to take into consideration and describes a set of rules that support the employee in problem solving and knowledge acquisition.
Assistance- and Knowledge-Services for Smart ProductionCarsten Ullrich
The transformation towards Smart Manufacturing results in machines that are increasingly complex to use and to maintain, as well as in ever-complicated production processes. Coupled with a continuing reduction of staff, this leads to an increasing demand for information needs and work expertise. At the same time, these challenges offer the opportunity to enhance the employee’s leeway with respect to designing and organizing their work. The project APPsist focuses on how this transformation can be supported technically and organizationally. This paper presents the technical approach: an architecture for intelligent-adaptive assistance and knowledge services. The paper describes how process mapping identified the requirements of the APPsist system, and presents the identified services and their communication, as well as the intelligent-adaptive functionality of the services.
A description of work done by our lab for the seminar "Sharing Inspiration in Higher Education" at the Finland Pavilion on the World Expo site in Shanghai.
Slides presented at a keynote at the Bayer Innovation Fair in Singapore.
The slides give a high-level introduction into the concepts and motivations of the Semantic Web, including lots of examples.
A large part was shown live during the keynote and thus is not properly rendered in the slides.
See <a></a> for more details.
Slides for our talk at Barcamp Shanghai 2009. The talk was mostly a discussion between us, our Australian partner and the audience, so you won't get much from the slides :(
Microblogging for Language Learning: Using Twitter to Train Communicative and...Carsten Ullrich
Our work analyzes the usefulness of microblogging in second language learning using the example of the social network Twitter. Most learners of English do not require even more passive input in form of texts, lectures or videos, etc. This input is readily available in numerous forms on the Internet. What learners of English need is the chance to actively produce language and the chance to use English as tool of communication. This calls for instructional methods and tools promoting ‘active’ learning that present opportunities for students to express themselves and interact in the target language. In this paper we describe how we used Twitter with students of English at the Distant College of Shanghai Jiao Tong University. We analyze the students’ messages and show how the usage of Twitter trained communicative and cultural competence.
Paper: http://www.carstenullrich.net/pubs/Borau09Microblogging.pdf
Video killed the radiostar, but will Web 3.0 kill the teacher?Carsten Ullrich
Slides from a presentation given at Spark09.
Technology disrupts. The WWW makes established business models and institutions obsolete, often almost overnight. Nobody foresaw that the proud Encyclopaedia Britannica would be replaced by Wikipedia, a Website to which everyone can contribute.
But what about schools and education? Will teachers by succeeded by avatars living in virtual worlds? Or will their role change, from an instructor to a moderator, from a know-it-all to a mediator of knowledge?
Web technology, from the collaborative and user-driven world of the Web 2.0 to the Web 3.0, the machine-understandable Semantic Web transforms education.
Supporting Active Learning and Education by Artificial Intelligence and Web 2.0Carsten Ullrich
At lecture for fresh(wo)men on learning theories, learning theories and language learning, using Twitter for language learning, building a personal learning environment in 10 minutes, building a personal learning environment in 10 seconds, Artificial Intelligence for learning support.
At lecture for fresh(wo)men on learning theories, learning theories and language learning, using Twitter for language learning, building a personal learning environment in 10 minutes, building a personal learning environment in 10 seconds, Artificial Intelligence for learning support.
In this version, there are problems with the images. Please use the pdf version for online viewing
http://www.slideshare.net/ullrich/sjtu221107-176532
The ppt download is better for the animations, though.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
Rapid Prototyping of a Semantic-Web-based Research Workbench
1. Rapid Prototyping of a
Semantic-Web-based
Research Workbench
Carsten Ullrich
Dept. of Computer Science and
Engineering, SJTU
2. Overview
• Project done with Totuba, Inc.
• Goal: develop a research workbench
– bibliography manager
– research network
– support while writing research papers
• Sorry, no new pure research results
• But: overview on state-of-the-art of
existing Web services / Web data
4. Entity Extraction
The term "Web 2.0" is used to describe applications that
distinguish themselves from previous generations of software by
a number of principles. Existing work shows that Web 2.0
applications can be successfully exploited for technology-
enhanced learning. However, in-depth analyses of the
relationship between Web 2.0 technology on the one hand and
teaching and learning on the other hand are still rare.
5. Entity Extraction
Gur grez "Jro 2.0" vf hfrq gb qrfpevor nccyvpngvbaf gung
qvfgvathvfu gurzfryirf sebz cerivbhf trarengvbaf bs fbsgjner ol n
ahzore bs cevapvcyrf. Rkvfgvat jbex fubjf gung Jro 2.0
nccyvpngvbaf pna or fhpprffshyyl rkcybvgrq sbe grpuabybtl-
raunaprq yrneavat. Ubjrire, va-qrcgu nanylfrf bs gur
eryngvbafuvc orgjrra Jro 2.0 grpuabybtl ba gur bar unaq naq
grnpuvat naq yrneavat ba gur bgure unaq ner fgvyy ener.
6. Entity Extraction
Gur grez "Jro 2.0" vf hfrq gb qrfpevor nccyvpngvbaf gung
qvfgvathvfu gurzfryirf sebz cerivbhf trarengvbaf bs fbsgjner ol n
ahzore bs cevapvcyrf. Rkvfgvat jbex fubjf gung Jro 2.0
nccyvpngvbaf pna or fhpprffshyyl rkcybvgrq sbe grpuabybtl-
raunaprq yrneavat. Ubjrire, va-qrcgu nanylfrf bs gur
eryngvbafuvc orgjrra Jro 2.0 grpuabybtl ba gur bar unaq naq
grnpuvat naq yrneavat ba gur bgure unaq ner fgvyy ener.
OpenCalais
• Jro 2.0
• grpuabybtl-raunaprq yrneavat
7. Open Calais
• Thomson Reuters company
• Web Service
• Extracts entities, facts,
events (about 100 types)
• Free for noncommercial and
commercial use
Entities
Anniversary, City, Company, Continent, Country, Currency, EmailAddress,
EntertainmentAwardEvent, Facility, FaxNumber, Holiday, IndustryTerm, MarketIndex,
MedicalCondition, MedicalTreatment, Movie, MusicAlbum, MusicGroup, NaturalFeature,
OperatingSystem, Organization, Person, PhoneNumber, Position, Product, ProgrammingLanguage,
ProvinceOrState, PublishedMedium, RadioProgram, RadioStation, Region, SportsEvent,
SportsGame, SportsLeague, Technology, TVShow, TVStation, URL
8. Semantifying
The term "Web 2.0“...
OpenCalais
• Web 2.0
• technology-supported learning
DBPedia (others: Yago, Freebase, UMBEL)
• http://dbpedia.org/resource/Web_2.0
• http://dbpedia.org/resource/Technology-Enhanced_Learning
11. Reuse
• Highly efficient entity extraction
• Enormous databases
– describe the entities
– link to related entities
• Give a high-level starting position to explore new
challenges
– how to put this data into use?
– context: what is relevant for user/current usage
12. Lessons Learned
• Reuse enables progress
– no duplication of work
– focus on problems relevant for you
• Having a landscape that encourages reuse
creates advantages for research / commercial
applications
• Problems
– mostly only English
– few Chinese services / programming libraries
• e.g., named entity extraction
16. Questions
• I have some:
– opinion mining
– information extraction
• Any toolkits available? RASCALLI?
• Contact me in case you find this
interesting
• ullrich_c@sjtu.edu.cn