Successfully reported this slideshow.
We use your LinkedIn profile and activity data to personalize ads and to show you more relevant ads. You can change your ad preferences anytime.

Generating a synthetic video dataset


Published on

Presenter: Mark Whiting
BELIV 2010 Workshop

  • Be the first to comment

  • Be the first to like this

Generating a synthetic video dataset

  1. 1. Creating a Synthetic Video Dataset for the VAST 2009 Challenge Mark A. Whiting, Carrie Varley, Jereme Haack Presented by Jean Scholtz BELIV 2010 03-28-2010
  2. 2. IEEE Visual Analytics Science and Technology (VAST) Annual Challenge – if you don’t know about it, you should review it <ul><li>Why look at theVASTChallenge? There are scenarios, tasks, and data available for use and modification; plus award-winning write-ups and software solutions. </li></ul><ul><li>When known ground truth is available, a new world opens up for evaluation of analytic software. </li></ul><ul><ul><li>Accuracy measures now become possible. </li></ul></ul><ul><li>See http:// to start with. Catherine Plaisant and her assistants have done a great job in capturing the 2006 – 2009 challenge information </li></ul>
  3. 3. VAST 2009 Features the First Video Challenge <ul><li>Multimedia and video analysis are becoming increasingly important in information analysis, as people do more with these media </li></ul><ul><li>First video analytics workshop at VAST 2009; special edition of IEEE CG&A on multimedia analytics soon </li></ul><ul><li>TRECVID looks at event detection; we wanted to look at events within scenarios </li></ul><ul><li>Questions can be asked: Why are those people there, and what does it have to do with the story being investigated? </li></ul>
  4. 4. Was this a “synthetic” video? <ul><li>Of course. See Herb Simon’s discussion of artificial and synthetic in The Sciences of the Artificial . We needed to engineer known ground truth into the scenes. </li></ul><ul><li>What did we do? We sent our actors in front of an existing webcam to act out a couple of scenes. </li></ul>
  5. 5. What was the scenario? <ul><li>A US embassy employee in “Flovania” was meeting their “handler,” a person who was acting friendly to the employee while really working for a criminal organization. The task was to find instances of possible meetings between the two as well as other events that might indicate espionage within a reasonably long segment of recorded video. </li></ul><ul><li>Evidence planted in the video was the illegal transfer of information from the employee to the handler as well as other meetings. </li></ul>
  6. 6. What webcam did we pick? <ul><li>Walla-Walla Washington. </li></ul>
  7. 7. Analysis Issues <ul><li>Had multiple (4) views to contend with </li></ul><ul><li>Only needed to analyze small segments of the video field of view </li></ul>
  8. 8. What did we plant? A scene where the Embassy employee “dupe” was meeting the handler outside the coffee shop and handing off information A scene where the handler was meeting another member of her criminal organization. They did the old briefcase “switcheroo”. See the light and dark cases.
  9. 9. How did we coordinate all the activities? <ul><li>One team was dispatched to Walla Walla; another was in the lab recording the web cam video off the web. </li></ul><ul><ul><li>Teams were connected via cell phone. </li></ul></ul><ul><li>The WW team has a timed script. When we needed to pass a message to the actors, we had a “non-actor” walk by and whisper instructions. </li></ul><ul><li>The lab team recorded several hours of web cam video overall, to ensure enough “noise” video to make the search challenging . </li></ul><ul><ul><li>Video duration was 8 hours </li></ul></ul><ul><ul><li>Each “scene” was viewed for several seconds before the camera switched to another view. </li></ul></ul><ul><ul><li>Three segments that had to be recognized for accuracy (7 seconds, 5 seconds and 7 seconds long) </li></ul></ul>
  10. 10. Contestant Results: one example University of Stuttgart – using video perpetuograms to track people from between scenes and views to enhance continuity
  11. 11. Lessons Learned & Next Steps <ul><li>Participants liked the challenge but few teams were prepared to do video. </li></ul><ul><li>We would like to coordinate more closely with groups like the TRECVID organizers for future video analytics challenges. </li></ul><ul><ul><li>And encourage teams entering VAST Challenge to seek out groups with expertise. </li></ul></ul><ul><li>This was a very new kind of challenge for both our visualization and information analysts – there wasn’t much of a baseline for either group to base assessments on. </li></ul><ul><ul><li>Little software available. </li></ul></ul><ul><ul><li>Question by analysts was what might be missed due to automatic recognition. </li></ul></ul><ul><li>As always it is a challenge to fit the mini challenges into the overall Grand Challenge. </li></ul>