The SpokenMedia project’s goal is to increase the effectiveness of web-based lecture media by improving the search and discoverability of specific, relevant media segments. SpokenMedia creates media-linked transcripts that will enable users to find contextually relevant video segments to improve their teaching and learning. The SpokenMedia project envisions a number of tools and services layered on top of, and supporting, these media-linked transcripts to enable users to interact with the media in more educationally relevant ways. Presented by Brandon Muramatsu to the IEEE-Computer Society Bangalore Section on August 6, 2009. (Unfortunately I didn't record the audio from this presentation :(, I thought it went really well and would have made a great slidecast.)
Project Greenfield: A New Way of Thinking about MIT OpenCourseWareBrandon Muramatsu
“Wouldn’t it be great if MIT OpenCourseWare…” How many times have you thought about this? What if there were a sandbox in which we could test and experiment with the tools and services that might prove useful to OpenCourseWare visitors, and more generally open educational resource content providers? The Office of Educational Innovation and Technology (OEIT) at MIT is setting up just such a sandbox, through Project Greenfield, to experiment with innovative experiences built on a mirror copy of MIT OpenCourseWare. Presented by Brandon Muramatsu at the Open Education 2010 Conference, Barcelona, Spain, November 2, 2010.
SpokenMedia Project: Media-Linked Transcripts and Rich Media Notebooks for Le...Brandon Muramatsu
The SpokenMedia project’s goal is to increase the effectiveness of web-based lecture media by improving the search and discoverability of specific, relevant media segments. SpokenMedia creates media-linked transcripts that will enable users to find contextually relevant video segments to improve their teaching and learning. The SpokenMedia project envisions a number of tools and services layered on top of, and supporting, these media-linked transcripts to enable users to interact with the media in more educationally relevant ways. Presented by Brandon Muramatsu at the Technology For Education 2009 Workshop, Bangalore, India, August 4, 2009.
Improving the OER Experience: Enabling Rich Media Notebooks of OER Video and ...Brandon Muramatsu
The SpokenMedia project at MIT is developing a web-based service to enable automatic lecture transcription. And it is developing a suite of tools and services to improve interaction with OER webcasts and podcasts enabling students and faculty to create rich media notebooks to support their learning and teaching. Presented by Brandon Muramatsu at OER 10, Cambridge, UK, March 23, 2010.
Automated Lecture Transcription at OCW Consortium Global Meeting 2009Brandon Muramatsu
Introduction and background to the automated lecture transcription/lecture transcription service project by MIT's Office of Educational Innovation and Technology (OEIT). Presented by Brandon Muramatsu at the OCW Consortium Global Meeting in Monterrey Mexico, April 22, 2009.
Building Community for Rich Media Notebooks: The SpokenMedia Project at NMC 2009Brandon Muramatsu
Moving toward a community around the SpokenMedia Project--for automatic transcription of lecture-style videos--by MIT's Office of Educational Innovation and Technology (OEIT) and the University of Queensland Centre for Educational Innovation and Technology. Presented by Brandon Muramatsu and Phillip Long at the New Media Consortium Summer Conference, Monterey, CA, June 12, 2009.
Implementing SpokenMedia for the Indian Institute for Human SettlementsBrandon Muramatsu
The SpokenMedia project implemented a proof-of-concept demonstration of its automatic lecture transcription and video player technologies for the Indian Institute for Human Settlements. This presentation describes the development of the technologies, production of the automatic lecture transcripts, deployment of the transcripts via a video player, initial reactions and future directions. Presented remotely by Brandon Muramatsu at the Technology for Education Conference, Mumbai, India, July 1, 2010
The Indian Institute for Human Settlements (IIHS, www.iihs.co.in) and the SpokenMedia (spoken- media.mit.edu) team from the MIT Office of Educational Innovation and Technology (OEIT) have been discussing how SpokenMedia technologies might be used by IIHS to provide cost effective ways of making video/audio course materials accessible to the diversity of students expected by IIHS. This presentation provides a case study of the proof-of-concept demonstration SpokenMedia developed for IIHS. Presented by Brandon Muramatsu at OCWC Global 2010, Hanoi, Vietnam, May 5, 2010.
SpokenMedia: Content, Content Everywhere...What video? Where? at OpenEd 2009Brandon Muramatsu
Need to find a specific video segment in an hour-long OCW/OER lecture? Try the tools and services developed by MIT OEIT to automatically transcribe lectures for search and discovery. Presented by Brandon Muramatsu at the OpenEd 2009, Vancouver, BC, Canada, August 12, 2009.
Project Greenfield: A New Way of Thinking about MIT OpenCourseWareBrandon Muramatsu
“Wouldn’t it be great if MIT OpenCourseWare…” How many times have you thought about this? What if there were a sandbox in which we could test and experiment with the tools and services that might prove useful to OpenCourseWare visitors, and more generally open educational resource content providers? The Office of Educational Innovation and Technology (OEIT) at MIT is setting up just such a sandbox, through Project Greenfield, to experiment with innovative experiences built on a mirror copy of MIT OpenCourseWare. Presented by Brandon Muramatsu at the Open Education 2010 Conference, Barcelona, Spain, November 2, 2010.
SpokenMedia Project: Media-Linked Transcripts and Rich Media Notebooks for Le...Brandon Muramatsu
The SpokenMedia project’s goal is to increase the effectiveness of web-based lecture media by improving the search and discoverability of specific, relevant media segments. SpokenMedia creates media-linked transcripts that will enable users to find contextually relevant video segments to improve their teaching and learning. The SpokenMedia project envisions a number of tools and services layered on top of, and supporting, these media-linked transcripts to enable users to interact with the media in more educationally relevant ways. Presented by Brandon Muramatsu at the Technology For Education 2009 Workshop, Bangalore, India, August 4, 2009.
Improving the OER Experience: Enabling Rich Media Notebooks of OER Video and ...Brandon Muramatsu
The SpokenMedia project at MIT is developing a web-based service to enable automatic lecture transcription. And it is developing a suite of tools and services to improve interaction with OER webcasts and podcasts enabling students and faculty to create rich media notebooks to support their learning and teaching. Presented by Brandon Muramatsu at OER 10, Cambridge, UK, March 23, 2010.
Automated Lecture Transcription at OCW Consortium Global Meeting 2009Brandon Muramatsu
Introduction and background to the automated lecture transcription/lecture transcription service project by MIT's Office of Educational Innovation and Technology (OEIT). Presented by Brandon Muramatsu at the OCW Consortium Global Meeting in Monterrey Mexico, April 22, 2009.
Building Community for Rich Media Notebooks: The SpokenMedia Project at NMC 2009Brandon Muramatsu
Moving toward a community around the SpokenMedia Project--for automatic transcription of lecture-style videos--by MIT's Office of Educational Innovation and Technology (OEIT) and the University of Queensland Centre for Educational Innovation and Technology. Presented by Brandon Muramatsu and Phillip Long at the New Media Consortium Summer Conference, Monterey, CA, June 12, 2009.
Implementing SpokenMedia for the Indian Institute for Human SettlementsBrandon Muramatsu
The SpokenMedia project implemented a proof-of-concept demonstration of its automatic lecture transcription and video player technologies for the Indian Institute for Human Settlements. This presentation describes the development of the technologies, production of the automatic lecture transcripts, deployment of the transcripts via a video player, initial reactions and future directions. Presented remotely by Brandon Muramatsu at the Technology for Education Conference, Mumbai, India, July 1, 2010
The Indian Institute for Human Settlements (IIHS, www.iihs.co.in) and the SpokenMedia (spoken- media.mit.edu) team from the MIT Office of Educational Innovation and Technology (OEIT) have been discussing how SpokenMedia technologies might be used by IIHS to provide cost effective ways of making video/audio course materials accessible to the diversity of students expected by IIHS. This presentation provides a case study of the proof-of-concept demonstration SpokenMedia developed for IIHS. Presented by Brandon Muramatsu at OCWC Global 2010, Hanoi, Vietnam, May 5, 2010.
SpokenMedia: Content, Content Everywhere...What video? Where? at OpenEd 2009Brandon Muramatsu
Need to find a specific video segment in an hour-long OCW/OER lecture? Try the tools and services developed by MIT OEIT to automatically transcribe lectures for search and discovery. Presented by Brandon Muramatsu at the OpenEd 2009, Vancouver, BC, Canada, August 12, 2009.
This presentation introduce how to create a small group of students engaged in a Whatsapp group and how to integrate vocabulary learning with it.
Feel free to use it and let me know by email: mustbek@yahoo.com
The Future of Closed Captioning in Higher Education3Play Media
With recent lawsuits, evolving legal requirements, and continuous advancements in technology, the question of closed captioning in higher education is one that is on a lot of people's minds. What does the future of captioning hold?
In this webinar, Sean Zdenek, author of the book Reading Sounds: Closed Captioned Media and Popular Culture and an Associate Professor at Texas Tech University, will answer exactly that question. Given the legal landscape, he will first focus on the hurdles and challenges of developing an infrastructure for closed captioning at the university level. Sean will then take a closer look at where closed captioning is going, focusing on the likely future requirements for, advancements in, and features of captioning.
This presentation will cover:
Developing an infrastructure for captioning at the university level
Training faculty & addressing faculty resistance
Economic analysis of captioning
Integrating closed captioning with lecture capture & video platforms
Faculty response to new captioning mandates
Future requirements for captioning
Current & future advancements in closed captioning technology
Advanced features that make captioning beneficial to all users
Teacher training my brainshark - 1 introductionPhil Longwell
Introduction to a proposal to use the cloud-based software, MyBrainshark, in an English Language Teaching setting - The China Central Academy Of Fine Arts, Beijing. Note: This was created for an ICT module on an MA course at the University of Warwick. Part 1 of 6.
Strategies for Getting Administrative and Faculty Buy-In for UDL3Play Media
The twentieth century saw access to higher education broaden in several significant ways. Formerly under-served populations of students were the targets of concentrated efforts to provide opportunities for college study: women, veterans, first-generation college learners, students from disadvantaged socioeconomic backgrounds, people with disabilities. To help make educational materials and teaching practices inclusive for all learners, this webinar radically reflects on how to motivate and inspire colleges and universities to adopt Universal Design for Learning (UDL).
Presented by Thomas Tobin, the Coordinator of Learning Technologies in the Center for Teaching and Learning (CTL) at Northeastern Illinois University in Chicago, this webinar will explore strategies for getting administrative and faculty buy-in for UDL. Thomas will focus on key shifts to make at your institution that will help demonstrate a measurable return on the investment of UDL.
This presentation will cover:
Training staff in Universal Design for Learning (UDL)
How to use UDL principles to increase student retention, persistence, and satisfaction
Motivating the adoption of UDL as part of campus culture
Getting administrative, budgetary, and faculty buy-in for UDL
How to expand the use of UDL elements beyond the legally required minimum
SpokenMedia: Automatic Lecture Transcription and Rich Media NotebooksBrandon Muramatsu
Need to find a specific segment in an hour-long web video, webcast or podcast of a lecture? Want to read a transcript of that lecture? Want to bookmark, annotate, or discuss video or audio clips from an entire lecture? The SpokenMedia project at MIT is developing a web-based service to enable automatic lecture transcription. The project is also developing a suite of tools and services to improve interaction with webcasts and podcasts enabling students and faculty to create rich media notebooks to support their learning and teaching. Presented by Brandon Muramatsu, Andrew McKinney and Peter Wilkins at the NERCOMP 2010, Providence, Rhode Island, March 9, 2010.
SpokenMedia automatically transcribes IIIHS video, and enables a process to edit and translate transcripts. Presented by Brandon Muramatsu at the IIHS Curriculum Conference, Bangalore, India, January 5, 2010.
This presentation introduce how to create a small group of students engaged in a Whatsapp group and how to integrate vocabulary learning with it.
Feel free to use it and let me know by email: mustbek@yahoo.com
The Future of Closed Captioning in Higher Education3Play Media
With recent lawsuits, evolving legal requirements, and continuous advancements in technology, the question of closed captioning in higher education is one that is on a lot of people's minds. What does the future of captioning hold?
In this webinar, Sean Zdenek, author of the book Reading Sounds: Closed Captioned Media and Popular Culture and an Associate Professor at Texas Tech University, will answer exactly that question. Given the legal landscape, he will first focus on the hurdles and challenges of developing an infrastructure for closed captioning at the university level. Sean will then take a closer look at where closed captioning is going, focusing on the likely future requirements for, advancements in, and features of captioning.
This presentation will cover:
Developing an infrastructure for captioning at the university level
Training faculty & addressing faculty resistance
Economic analysis of captioning
Integrating closed captioning with lecture capture & video platforms
Faculty response to new captioning mandates
Future requirements for captioning
Current & future advancements in closed captioning technology
Advanced features that make captioning beneficial to all users
Teacher training my brainshark - 1 introductionPhil Longwell
Introduction to a proposal to use the cloud-based software, MyBrainshark, in an English Language Teaching setting - The China Central Academy Of Fine Arts, Beijing. Note: This was created for an ICT module on an MA course at the University of Warwick. Part 1 of 6.
Strategies for Getting Administrative and Faculty Buy-In for UDL3Play Media
The twentieth century saw access to higher education broaden in several significant ways. Formerly under-served populations of students were the targets of concentrated efforts to provide opportunities for college study: women, veterans, first-generation college learners, students from disadvantaged socioeconomic backgrounds, people with disabilities. To help make educational materials and teaching practices inclusive for all learners, this webinar radically reflects on how to motivate and inspire colleges and universities to adopt Universal Design for Learning (UDL).
Presented by Thomas Tobin, the Coordinator of Learning Technologies in the Center for Teaching and Learning (CTL) at Northeastern Illinois University in Chicago, this webinar will explore strategies for getting administrative and faculty buy-in for UDL. Thomas will focus on key shifts to make at your institution that will help demonstrate a measurable return on the investment of UDL.
This presentation will cover:
Training staff in Universal Design for Learning (UDL)
How to use UDL principles to increase student retention, persistence, and satisfaction
Motivating the adoption of UDL as part of campus culture
Getting administrative, budgetary, and faculty buy-in for UDL
How to expand the use of UDL elements beyond the legally required minimum
SpokenMedia: Automatic Lecture Transcription and Rich Media NotebooksBrandon Muramatsu
Need to find a specific segment in an hour-long web video, webcast or podcast of a lecture? Want to read a transcript of that lecture? Want to bookmark, annotate, or discuss video or audio clips from an entire lecture? The SpokenMedia project at MIT is developing a web-based service to enable automatic lecture transcription. The project is also developing a suite of tools and services to improve interaction with webcasts and podcasts enabling students and faculty to create rich media notebooks to support their learning and teaching. Presented by Brandon Muramatsu, Andrew McKinney and Peter Wilkins at the NERCOMP 2010, Providence, Rhode Island, March 9, 2010.
SpokenMedia automatically transcribes IIIHS video, and enables a process to edit and translate transcripts. Presented by Brandon Muramatsu at the IIHS Curriculum Conference, Bangalore, India, January 5, 2010.
Video Captioning for Accessibility: University of Florida and Regis Universit...3Play Media
With the proliferation of education video, captioning has become an essential part of many university accessibility policies. Although captioning is sometimes perceived as obtrusive and expensive, in this session University of Florida and Regis University demonstrate their cost-effective, streamlined captioning workflows that provide push-button simplicity for instructors and administrators campus-wide.
Watch this session to learn about the efficient and cost-effective ways to implement video captioning solutions. We will also cover the basics of how to create closed captions, accessibility laws, captions formats including emerging formats for HTML5 and mobile, video player compatibility, and automated workflows with Kaltura.
This session was part of the Kaltura Education Video Summit held on December 6, 2012.
Presenters:
Nicole Croy
eLearning Technologist | Regis University
Jason Neely
Office of Distance Learning | University of Florida
Tole Khesin
VP of Marketing | 3Play Media
University of Wisconsin: Captioning and Transcription Policies, Uses and Work...3Play Media
This slideshow comes from a presentation at UB Tech in June, 2014. The presentation was led by Josh Miller of 3Play Media and Patrick Wirth of University of Wisconsin-Extension, Continuing Education, Outreach & E-Learning (CEOEL). 3Play Media provides captioning and transcription services for UW-Extension.
CEOEL serves as the coordinator for University of Wisconsin continuing education programs at all 26 campuses. The presentation covered:
- Overview of CEOEL accessibility policies and procedures
- Overview of media team captioning and transcription workflows
- Discussion of CEOEL strategies for addressing accessibility issues/needs
- Showcase of creative uses for transcripts and captions around marketing initiatives
- Access to media players and templates (audio, video, narrations, etc.)
Tools for Designing Distance Learning InstructionMarsha J. Chan
Differences between live and online classrooms require instructors to adapt familiar tools and adopt new ones to create a learning environment in cyberspace. The presenter will describe one module in an online certification course for college faculty that examines technology tools to design and teach a web-based distance learning course.
Slides for podcasting workshop at Leeds Met, 08 Sept 09.
http://www.leedscarnegiefestivals.co.uk/EventDetails.aspx?eid=412
Podcasts In Higher Education – Learning On The Move?
Podcasts can be of particular interest in order to promote flexibility in teaching and learning. In our presentation, accompanied by a hands-on workshop, we will share our experience of creating and evaluating the use of podcasts based on projects conducted through CETL ALiC (Centre for Excellence in Teaching and Learning - Active Learning in Computing) and the TEL Team.
In this workshop you will be introduced to podcasts and their practical application in a learning and teaching environment. You will use Audacity software to create your own podcast and then upload it to the podcast server. This will give you the option to make your podcast available for downloading via RSS feed (e.g. iTunes) to your intended audience.
Low-Cost, High-Gain Technologies for Teachingbrettssu
Slideset from seminar given to faculty of Middle East Technical University. Goals was to share some interesting examples of technologies for teaching and learning that were relatively simple and inexpensive.
Implementing Universal and Inclusive Design for Online Learning Accessibility3Play Media
Accessibility is a critical component of any online learning content. With legal requirements stronger than ever, colleges and universities must find tangible ways to improve their web accessibility. This webinar will discuss how the principles of universal and inclusive design can be applied to the online learning environment, with a particular focus on the accessibility of course content and materials.
Howard Kramer, an Access Specialist at University of Colorado at Boulder, and Sheryl Burgstahler, the Director of Accessible Technology Services at the University of Washington, will explain what universal design is, the importance of incorporating universal design principles into online courses, and strategies for doing so.
This webinar will cover:
What is universal and inclusive design?
Strategies for implementing universal design
Best practices for the presentation of information and resources
Incorporating inclusiveness into a syllabus
Creating accessible documents and media
Providing information through multiple mediums
Resources and tools for incorporating inclusive design into the online environment
Bridging the Gap: Mixing approaches, content and tools to help college studentsBrandon Muramatsu
The Next Generation Learning Challenge has provided a call to action for those involved in Open Educational Resources to meet the needs of the US education system. One of the challenges is to deploy open core courseware to address the retention and completion issues in community colleges. In the Open Learning: Bridge to Success (B2S) initiative The Open University working in partnership with MIT, Anne Arundel Community College (AACC) and University of Maryland University College (UMUC) will develop open bridging modules to help learners having problems in coping with credit bearing courses. Modules focussed on learning to learn and understanding mathematics will be released as complete open educational resource packages from The Open University's existing successful programme of entry-level (pre-degree) "Openings" modules. The Open University has an established open presence through its OpenLearn open content site which offers a wide range of units, and the courses will be developed in the open to benefit not only students in the partner institutions but any learners who wish to use them.
The project will run its first pilots with Community College students from September and this presentation focuses on the early phase of the project including: release of the initial materials, augmentation with other OER, design of the research methodology and early lessons from working together as partners. Already working in the open is changing how we think about the provision of content and the instruction of practical experiences alongside meeting curriculum needs. We anticipate presenting the design requirements and how they have been met through open provision, reflections from those involved in the projects, the first feedback from students at the pilot colleges, and the indications from the additional users in the open.
Presented by Dale Allen, Jeff Merriman, Brandon Muramatsu, Domy Raymond, and Mike Reilly, Grantmakers for Education, San Francisco, CA, October 22-23, 2015
We share a potential model for online recitation sessions for MIT residential courses based on our experiences running similar sessions for courses in the MITx MicroMasters Program in Statistics and Data Science.
The Best of Both Worlds: Transforming OpenCourseWare in an age of InteractivityBrandon Muramatsu
The Best of Both Worlds: Transforming OpenCourseWare in an age of Interactivity presented by Peter Pinch and Brandon Muramatsu in Arlington, VA on November 20, 2014.
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
A review of the growth of the Israel Genealogy Research Association Database Collection for the last 12 months. Our collection is now passed the 3 million mark and still growing. See which archives have contributed the most. See the different types of records we have, and which years have had records added. You can also see what we have for the future.
Delivering Micro-Credentials in Technical and Vocational Education and TrainingAG2 Design
Explore how micro-credentials are transforming Technical and Vocational Education and Training (TVET) with this comprehensive slide deck. Discover what micro-credentials are, their importance in TVET, the advantages they offer, and the insights from industry experts. Additionally, learn about the top software applications available for creating and managing micro-credentials. This presentation also includes valuable resources and a discussion on the future of these specialised certifications.
For more detailed information on delivering micro-credentials in TVET, visit this https://tvettrainer.com/delivering-micro-credentials-in-tvet/
Normal Labour/ Stages of Labour/ Mechanism of LabourWasim Ak
Normal labor is also termed spontaneous labor, defined as the natural physiological process through which the fetus, placenta, and membranes are expelled from the uterus through the birth canal at term (37 to 42 weeks
Acetabularia Information For Class 9 .docxvaibhavrinwa19
Acetabularia acetabulum is a single-celled green alga that in its vegetative state is morphologically differentiated into a basal rhizoid and an axially elongated stalk, which bears whorls of branching hairs. The single diploid nucleus resides in the rhizoid.
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
South African Journal of Science: Writing with integrity workshop (2024)
SpokenMedia Project: Media-Linked Transcripts and Rich Media Notebooks for Learning and Teaching at IEEE-CS Bangalore
1. The SpokenMedia Project: Toward Rich Media Notebooks for Teaching and Learning Brandon Muramatsu [email_address] MIT, Office of Educational Innovation and Technology Andrew McKinney, MIT OEIT Phillip Long and John Zornig, University of Queensland Citation: Muramatsu, B., McKinney, A., Long, P. D., & Zornig, J. (2009). The SpokenMedia Project: Toward Rich Media Notebooks for Teaching and Learning. Presented to the IEEE Computer Soceity: Bangalore, India, August 6, 2009. Unless otherwise specified, this work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License
2.
3.
4.
5. What about Bing? Bing Search for “angular momentum” Performed August 2009
6.
7.
8.
9.
10.
11. How Does it Work? Lecture Transcription Workflow
24. Thanks! oeit.mit.edu/spokenmedia Brandon Muramatsu [email_address] MIT, Office of Educational Innovation and Technology Andrew McKinney, MIT OEIT Phillip Long and John Zornig, University of Queensland Citation: Muramatsu, B., McKinney, A., Long, P. D., & Zornig, J. (2009). The SpokenMedia Project: Toward Rich Media Notebooks for Teaching and Learning. Presented to the IEEE Computer Society: Bangalore, India, August 6, 2009. Unless otherwise specified, this work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License
Editor's Notes
Citation: Muramatsu, B., McKinney, A., Long, P. D., & Zornig, J. (2009). The SpokenMedia Project: Toward Rich Media Notebooks for Teaching and Learning. Presented to the IEEE Computer Society: Bangalore, India, August 6, 2009. Unless otherwise specified, this work is licensed under a Creative Commons Attribution-Noncommercial-Share Alike 3.0 United States License
Why are we doing this? In the last few years, we’ve seen an explosion of videos on the web. Self publishing by millions on YouTube. Universities recording course lectures and putting them on the web. Students are relying upon web video for learning. Common statistic mentioned by folks like UC Berkeley (which has been doing course webcasts since 1999) is that usage spikes as students prepare for tests, and that they tend to focus on small segments of the video Also, cultural organizations (museums, foundations, non-profit organizations) sharing their interviews on the web. Other similar single speaker web video, cost of technology has come down.
What are the challenges? Large volume of material to search through! Search results—approximately 3 Million in Google (April 2009): Wikipedia, Angular and Conservation of Angular Momentum links might be useful Quantum mechanics link is probably too advanced Angular Momentum (company) probably not useful But no videos Oh, there’s a way of just doing a video search at Google, search is segmented by media type Google Video Search results—only 400 (April 2009), that’s better: All appear to be relevant Two are lecture length (i.e. 20+ minutes or longer): Mechanical Universe, and Lecture 21 from MIT OCW Four are probably demos relating angular momentum to physical examples (tennis, ice skating) Search results are based on: Metadata Title of video/link Text description of video (typically short), or the text surrounding an embedded video
Additional Challenges Interaction and Use Get the full length video, over 50 minutes There may or may not be a transcript, which may or may not be displayed as captioning for accessibility Policy Implications Technology allows for bookmarking and comments, they aren’t enabled
What about Bing? Fewer Web search results, only 1 Million (August 2009) Three of top six are for companies (two for watchmaker, one for other) Still segmented searching (web, video) Much less Video search results, only 2,400 (August 2009) Video search results much less relevant, First five are for watches, Next three are educational, Does not include Mechanical Universe or MIT OCW videos in first 20 results, NPTEL video is result 19
Why do we want these tools? MIT as the customer Lots of materials, 1900+ OCW courses, some with video/audio Opportunities for positive change: improving presentation and user experience, advocate for new methods of interaction
What can we do today? Demo of Spoken Lecture Browser
Spoken Lecture Browser Demo Search of video library Search results show search query in context, query highlighted in yellow Results show “conceptual chunks” Player controls video and bouncing ball in the transcript Uses 2006 technology, Real Player – needs to get updated for current technlogy Much of the work for the Spoken Lecture browser and player completed in 2006 (though recognizer research continues) In 2006, Real could allow within movie bookmarks and playback In 2006, Real was the preferred high-quality video playback mechanism on the Web (lots has changed since then)
Lecture Transcription Jim Glass and his group have years of research experience for spoken languages Lectures are a different type of spoken language Much of the speech recognition research has focused on real time transcription of news broadcasts, or interactive voice response systems (telephone) Broadcast news has something like 300 unique words in an hour long broadcast Broadcast news is well structured, prepared copy (in studio via teleprompters), clear transitions between speakers, etc. Lectures are conversational and spontaneous Can use highly specialized vocabularies, engineering, physical sciences, mathematics
Spoken Lecture Project Supported by iCampus Includes the browser (which was just demo’d) the processor (back end lecture transcription) and a hand workflow to do the processing Approximately 400 hours of video indexed
How does it work? Audio System only needs audio (waveform), extracts from video Domain Model (base is generic domain model) System needs to know what words it can expect to find in the audio Syllabus, lecture notes, index from text book, research papers Build library of domains Separate sub-process for text for domain model Speaker model (base is generic speaker model) If multiple lectures by the same author, best to create a speaker model Separate sub-process for speaker model Process—With audio, domain and speaker models Output Time coded transcript (standard formats) Links media and transcript Applications Search/retrieval Player
Recognizer Accuracy Base accuracy is approximately 50% (generic domain and speaker models) Increase accuracy with speaker model up to 80-85%, and specific domain model This approach is good for courses with multiple lectures by the same speaker Domain models get more useful as more relevant text documents are indexed (keyword/noun phrase extraction) Initial results indicate that doing one 99% accurate (by hand/manual) transcript can help immensely for additional lectures by the same speaker Better use of limited resources Search accuracy is closer to 90%, searches tend to be for unique words which the processor is better at recognizing
Transcript “Errors” Recall, processor has 85% accuracy Here are two examples of recognizer errors… In the first case, looking at the transcript, it’s hard to say what the speaker (Lewin) might have said Continuing … it’s unlikely that he used the word “fork” twice Let’s listen…ok. It’s torque not fork Recognizer can recognize when it’s guessing—that’s not exposed in a public interface, but could be
What we have today It’s not perfect, but a pretty good start Prototype has a number of useful features that demonstrate search interfaces and interaction interfaces
Where are we heading? Transition from research project to service Explore new interactions—what we’re calling Rich Media Notebooks
Towards a Lecture Transcription Service OEIT at MIT’s goal is to transition from research to production First priority to get running on our servers Prototype a transcript production service—second priority For MIT Automate a mostly hand process Considering integration with local Podcast Producer workflow engine (Apple) Integrate into media production workflow, as a plugin Partner with other content producers to test service—tied for third priority See how it meets needs of other content producers See how it plays with Opencast Matterhorn, distributed service
A Lecture Transcription Service? Caveats Full disclosure, limitations we know about or think are important We’ve been asked about other languages Should be possible Jim Glass is experimenting with Chinese Would have to create a language model, not sure what’s involved with that Current plan to host a web service from MIT Contribution will be important aspect of participation
Toward Rich Media Notebooks Updating the playback—third priority Flash and H.264 encoded video Implement other common video features (e.g., from YouTube and other commercial video sites) Bookmarking, annotations and comments (timestamp, text fields) Clip creation (ala XMAS cross media annotation system) Down the road Social editing to improve transcripts, wiki interfaces, trust systems Concept and semantic searching—current system breaks text up into “logical” chunks, and lets users search for a term or phrase, but doesn’t really get to concepts.
OCW Highlights for High School Links MIT OCW course materials to Advanced Placement tests Created by undergrads and high school students Hand-created link between topics in AP Physics and existing course resources Links concepts to specific video segments
Look Listen Learn Hand crafted from MIT OCW video Created equations that are written on the board (in LaTeX) Displays equations as they are written Displays transcripts as well Must have taken quite a bit of time to do this Probably a really good interface for the material
Google Audio Indexing Google Labs Project Indexed CSPAN on economy and 2008 U.S. Presidential elections Similar search results and display of segments (though less text for context) Takes advantage of playback from within Flash at YouTube (which they haven’t made available to the general public as far as I know yet)