Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save people's lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visual data collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
Links:
http://infolab.usc.edu/DocsDemos/to_ieeebigdata2015.pdf
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=7363814
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save peoples lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visualdata collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-kim
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Minyoung Kim, Senior Research Engineer at Panasonic Silicon Valley Laboratory, presents the "A Fast Object Detector for ADAS using Deep Learning" tutorial at the May 2017 Embedded Vision Summit.
Object detection has been one of the most important research areas in computer vision for decades. Recently, deep neural networks (DNNs) have led to significant improvement in several machine learning domains, including computer vision, achieving the state-of-the-art performance thanks to their theoretically proven modeling and generalization capabilities. However, it is still challenging to deploy such DNNs on embedded systems, for applications such as advanced driver assistance systems (ADAS), where computation power is limited.
Kim and her team focus on reducing the size of the network and required computations, and thus building a fast, real-time object detection system. They propose a fully convolutional neural network that can achieve at least 45 fps on 640x480 frames with competitive performance. With this network, there is no proposal generation step, which can cause a speed bottleneck; instead, a single forward propagation of the network approximates the locations of objects directly.
Applying Photonics to User Needs: The Application ChallengeLarry Smarr
05.02.28
Invited Talk to the 4th Annual On*VECTOR International Photonics Workshop
Sponsored by NTT Network Innovation Laboratories
Title: Applying Photonics to User Needs: The Application Challenge
University of California, San Diego
The OptIPortal, a Scalable Visualization, Storage, and Computing Termination ...Larry Smarr
The OptIPortal is a scalable visualization, storage, and computing termination device for high bandwidth campus bridging. It is built from commodity PC clusters and LCDs to create a 10Gbps scalable termination device. OptIPortals provide end-to-end cyberinfrastructure for petascale end users and can display high resolution portals over dedicated optical channels to global science data.
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
In this deck from the Stanford HPC Conference, Katie Lewis from Lawrence Livermore National Laboratory presents: The Incorporation of Machine Learning into Scientific Simulations at Lawrence Livermore National Laboratory.
"Scientific simulations have driven computing at Lawrence Livermore National Laboratory (LLNL) for decades. During that time, we have seen significant changes in hardware, tools, and algorithms. Today, data science, including machine learning, is one of the fastest growing areas of computing, and LLNL is investing in hardware, applications, and algorithms in this space. While the use of simulations to focus and understand experiments is well accepted in our community, machine learning brings new challenges that need to be addressed. I will explore applications for machine learning in scientific simulations that are showing promising results and further investigation that is needed to better understand its usefulness."
Watch the video: https://youtu.be/NVwmvCWpZ6Y
Learn more: https://computing.llnl.gov/research-area/machine-learning
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Looking Back, Looking Forward NSF CI Funding 1985-2025Larry Smarr
This document provides an overview of the development of national research platforms (NRPs) from 1985 to the present, with a focus on the Pacific Research Platform (PRP). It describes the evolution of the PRP from early NSF-funded supercomputing centers to today's distributed cyberinfrastructure utilizing optical networking, containers, Kubernetes, and distributed storage. The PRP now connects over 15 universities across the US and internationally to enable data-intensive science and machine learning applications across multiple domains. Going forward, the document discusses plans to further integrate regional networks and partner with new NSF-funded initiatives to develop the next generation of NRPs through 2025.
The document provides an overview of the Pacific Research Platform (PRP) and discusses its role in connecting researchers across institutions and enabling new applications. It summarizes the PRP's key components like Science DMZs, Data Transfer Nodes (FIONAs), and use of Kubernetes for container management. Several examples are given of how the PRP facilitates high-performance distributed data analysis, access to remote supercomputers, and sensor networks coupled to real-time computing. Upcoming work on machine learning applications and expanding the PRP internationally is also outlined.
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save peoples lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visualdata collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/embedded-vision-alliance/embedded-vision-training/videos/pages/may-2017-embedded-vision-summit-kim
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Minyoung Kim, Senior Research Engineer at Panasonic Silicon Valley Laboratory, presents the "A Fast Object Detector for ADAS using Deep Learning" tutorial at the May 2017 Embedded Vision Summit.
Object detection has been one of the most important research areas in computer vision for decades. Recently, deep neural networks (DNNs) have led to significant improvement in several machine learning domains, including computer vision, achieving the state-of-the-art performance thanks to their theoretically proven modeling and generalization capabilities. However, it is still challenging to deploy such DNNs on embedded systems, for applications such as advanced driver assistance systems (ADAS), where computation power is limited.
Kim and her team focus on reducing the size of the network and required computations, and thus building a fast, real-time object detection system. They propose a fully convolutional neural network that can achieve at least 45 fps on 640x480 frames with competitive performance. With this network, there is no proposal generation step, which can cause a speed bottleneck; instead, a single forward propagation of the network approximates the locations of objects directly.
Applying Photonics to User Needs: The Application ChallengeLarry Smarr
05.02.28
Invited Talk to the 4th Annual On*VECTOR International Photonics Workshop
Sponsored by NTT Network Innovation Laboratories
Title: Applying Photonics to User Needs: The Application Challenge
University of California, San Diego
The OptIPortal, a Scalable Visualization, Storage, and Computing Termination ...Larry Smarr
The OptIPortal is a scalable visualization, storage, and computing termination device for high bandwidth campus bridging. It is built from commodity PC clusters and LCDs to create a 10Gbps scalable termination device. OptIPortals provide end-to-end cyberinfrastructure for petascale end users and can display high resolution portals over dedicated optical channels to global science data.
The Incorporation of Machine Learning into Scientific Simulations at Lawrence...inside-BigData.com
In this deck from the Stanford HPC Conference, Katie Lewis from Lawrence Livermore National Laboratory presents: The Incorporation of Machine Learning into Scientific Simulations at Lawrence Livermore National Laboratory.
"Scientific simulations have driven computing at Lawrence Livermore National Laboratory (LLNL) for decades. During that time, we have seen significant changes in hardware, tools, and algorithms. Today, data science, including machine learning, is one of the fastest growing areas of computing, and LLNL is investing in hardware, applications, and algorithms in this space. While the use of simulations to focus and understand experiments is well accepted in our community, machine learning brings new challenges that need to be addressed. I will explore applications for machine learning in scientific simulations that are showing promising results and further investigation that is needed to better understand its usefulness."
Watch the video: https://youtu.be/NVwmvCWpZ6Y
Learn more: https://computing.llnl.gov/research-area/machine-learning
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Looking Back, Looking Forward NSF CI Funding 1985-2025Larry Smarr
This document provides an overview of the development of national research platforms (NRPs) from 1985 to the present, with a focus on the Pacific Research Platform (PRP). It describes the evolution of the PRP from early NSF-funded supercomputing centers to today's distributed cyberinfrastructure utilizing optical networking, containers, Kubernetes, and distributed storage. The PRP now connects over 15 universities across the US and internationally to enable data-intensive science and machine learning applications across multiple domains. Going forward, the document discusses plans to further integrate regional networks and partner with new NSF-funded initiatives to develop the next generation of NRPs through 2025.
The document provides an overview of the Pacific Research Platform (PRP) and discusses its role in connecting researchers across institutions and enabling new applications. It summarizes the PRP's key components like Science DMZs, Data Transfer Nodes (FIONAs), and use of Kubernetes for container management. Several examples are given of how the PRP facilitates high-performance distributed data analysis, access to remote supercomputers, and sensor networks coupled to real-time computing. Upcoming work on machine learning applications and expanding the PRP internationally is also outlined.
Global Research Platforms: Past, Present, FutureLarry Smarr
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise boosts blood flow, releases endorphins, and promotes changes in the brain which help regulate emotions and stress levels.
PEARC17: Data Access for LIGO on the OSGDerek Weitzel
This document discusses how LIGO collaborated with the Open Science Grid (OSG) to distribute large-scale gravitational wave data analysis computations across OSG resources. It describes how the initial implementation faced issues at scale due to data transfer bottlenecks. An improved solution deployed a distributed data caching system called StashCache using XRootD and CVMFS to provide secure POSIX access to LIGO data across OSG sites, enabling production of the PyCBC analysis workflow at scale on the OSG.
LambdaGrids--Earth and Planetary Sciences Driving High Performance Networks a...Larry Smarr
05.02.04
Invited Talk to the NASA Jet Propulsion Laboratory
Title: LambdaGrids--Earth and Planetary Sciences Driving High Performance Networks and High Resolution Visualizations
Pasadena, CA
Toward a Global Interactive Earth Observing CyberinfrastructureLarry Smarr
The document discusses the need for a new generation of cyberinfrastructure to support interactive global earth observation. It outlines several prototyping projects that are building examples of systems enabling real-time control of remote instruments, remote data access and analysis. These projects are driving the development of an emerging cyber-architecture using web and grid services to link distributed data repositories and simulations.
The document discusses the growing carbon footprint of information and communication technologies (ICT) and efforts to make cyberinfrastructure more energy efficient and environmentally sustainable. Specifically, it mentions that (1) ICT energy usage is growing rapidly and accounts for 2% of global greenhouse gas emissions, (2) universities are working on initiatives like the GreenLight project to reduce ICT energy usage through techniques like dynamic power management, and (3) further research is needed to develop more energy-efficient computing technologies, data center designs, and videoconferencing solutions to reduce the need for travel.
Complexity Neural Networks for Estimating Flood Process in Internet-of-Things...Dr. Amarjeet Singh
With the advancement of the Internet of Things (IoT)-based water conservation computerization, hydrological data is increasingly enriched. Considering the ability of deep learning on complex features extraction, we proposed a flood process forecastin gmodel based on Convolution Neural Network(CNN) with two-dimension(2D) convolutional operation. At first, we imported the spatial-temporal rainfall features of the Xixian basin. Subsequently, extensive experiments were carried out to determine the optimal hyperparameters of the proposed CNN flood forecasting model.
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemLarry Smarr
The Pacific Research Platform (PRP) is a multi-institutional partnership that establishes a high-capacity "big data freeway system" spanning the University of California campuses and other research universities in California to facilitate rapid data access and sharing between researchers and institutions. Fifteen multi-campus application teams in fields like particle physics, astronomy, earth sciences, biomedicine, and visualization drive the technical design of the PRP over five years. The goal of the PRP is to extend campus "Science DMZ" networks to allow high-speed data movement between research labs, supercomputer centers, and data repositories across campus, regional
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Global grid of master events for waveform cross correlation: design and testingIvan Kitov
This document describes the design and testing of a global grid of master events for waveform cross correlation to improve seismic monitoring capabilities. The grid uses real and synthetic master events from seismic arrays and other stations. Testing in February 2013 using 134 events found 92 matches using the grid, demonstrating its potential to detect more small events through cross correlation across the global seismic network. Future versions aim to optimize the use of real data, principal components, and machine learning to further enhance monitoring.
- Digital mirror worlds are software models of physical systems that are continuously updated with real-time data, allowing them to closely mimic and predict the behavior of the real system.
- Advances in computing power and sensors are enabling increasingly detailed digital twins of objects, human bodies, cities, wildfires, and even the observable universe.
- One trillion sensors are expected to feed the planetary computer within the next decade, driving a global industrial internet and $15 trillion in economic value through digital twins of manufactured products.
- Digital twins powered by consumer sensor data may one day provide early disease detection and personalized health coaching at scale.
National Federated Compute Platforms: The Pacific Research PlatformLarry Smarr
The Pacific Research Platform (PRP) is a multi-institution hypercluster that connects science DMZs across 25 partner campuses using FIONA data transfer nodes and 10-100Gbps networks. PRP adopted Kubernetes and Rook to orchestrate petabytes of distributed storage and GPUs for data science applications. A CHASE-CI grant added machine learning capabilities. PRP is working to federate with the Open Science Grid and become a prototype for a future National Research Platform connecting regional networks.
The Pacific Research Platform: Building a Distributed Big-Data Machine-Learni...Larry Smarr
The document summarizes the Pacific Research Platform (PRP) which connects researchers across multiple universities with high-speed networks and computing resources for big data and machine learning applications. Key points:
- PRP connects 15 universities with optical networks, distributed storage devices (FIONAs), and over 350 GPUs for data analysis and AI training.
- It allows researchers to rapidly share and analyze large datasets, with one example reducing a workflow from 19 days to 52 minutes.
- Other projects using PRP resources include climate modeling, astrophysics simulations, and machine learning courses involving thousands of students.
Analyzing Large Earth Data Sets: New Tools from the OptiPuter and LOOKING Pro...Larry Smarr
The document discusses two projects, OptIPuter and LOOKING, that aim to analyze large earth data sets using optical networking and grid technologies. OptIPuter extends grid middleware to dedicated optical circuits for earth and medical sciences. LOOKING builds on OptIPuter to provide real-time control of ocean observatories through web and grid services integrated over optical networks. Both projects represent efforts to develop cyberinfrastructure for interactive analysis of remote earth science data and instruments.
Calit2-a Persistent UCSD/UCI Framework for CollaborationLarry Smarr
05.02.16
Invited Talk
Sun Microsystems Global Education and Research
Conference 2005
Title: Calit2-a Persistent UCSD/UCI Framework for Collaboration
San Francisco, CA
Peering The Pacific Research Platform With The Great Plains NetworkLarry Smarr
The Pacific Research Platform (PRP) connects research institutions across the western United States with high-speed networks to enable data-intensive science collaborations. Key points:
- The PRP connects 15 campuses across California and links to the Great Plains Network, allowing researchers to access remote supercomputers, share large datasets, and collaborate on projects like analyzing data from the Large Hadron Collider.
- The PRP utilizes Science DMZ architectures with dedicated data transfer nodes called FIONAs to achieve high-speed transfer of large files. Kubernetes is used to manage distributed storage and computing resources.
- Early applications include distributed climate modeling, wildfire science, plankton imaging, and cancer genomics. The PR
The Pacific Research Platform: Building a Distributed Big Data Machine Learni...Larry Smarr
This document summarizes Dr. Larry Smarr's invited talk about the Pacific Research Platform (PRP) given at the San Diego Supercomputer Center in April 2019. The PRP is building a distributed big data machine learning supercomputer by connecting high-performance computing and data resources across multiple universities in California and beyond using high-speed networks. It provides researchers with petascale computing power, distributed storage, and tools like Kubernetes to enable collaborative data-intensive science across institutions.
The document discusses the Pacific Research Platform (PRP), a distributed cyberinfrastructure that connects researchers and data across multiple campuses in California and beyond using optical fiber networking. Key points:
- The PRP uses high-speed networking infrastructure like the CENIC network to connect data generators and consumers across 15+ campuses, creating an integrated "big data freeway system".
- It deploys specialized data transfer nodes called FIONAs to enable high-speed transfer of large datasets between sites at near the full network speed.
- Recent additions include using Kubernetes to orchestrate containers across the PRP infrastructure and integrating machine learning resources through the CHASE-CI grant to support data-intensive AI applications.
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save people's lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visual data collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
IRJET- A Real Time Yolo Human Detection in Flood Affected Areas based on Vide...IRJET Journal
This document proposes a method for real-time human detection in flood-affected areas using video content analysis and the YOLO object detection algorithm. It trains YOLO on the COCO Human dataset to detect and localize humans in video frames from surveillance cameras. The results show that YOLO can accurately detect multiple humans, even with occlusion, and single humans under varying illumination. This approach aims to help rescue operations quickly identify affected areas and prioritize aid.
Global Research Platforms: Past, Present, FutureLarry Smarr
The document discusses the benefits of exercise for mental health. Regular physical activity can help reduce anxiety and depression and improve mood and cognitive functioning. Exercise boosts blood flow, releases endorphins, and promotes changes in the brain which help regulate emotions and stress levels.
PEARC17: Data Access for LIGO on the OSGDerek Weitzel
This document discusses how LIGO collaborated with the Open Science Grid (OSG) to distribute large-scale gravitational wave data analysis computations across OSG resources. It describes how the initial implementation faced issues at scale due to data transfer bottlenecks. An improved solution deployed a distributed data caching system called StashCache using XRootD and CVMFS to provide secure POSIX access to LIGO data across OSG sites, enabling production of the PyCBC analysis workflow at scale on the OSG.
LambdaGrids--Earth and Planetary Sciences Driving High Performance Networks a...Larry Smarr
05.02.04
Invited Talk to the NASA Jet Propulsion Laboratory
Title: LambdaGrids--Earth and Planetary Sciences Driving High Performance Networks and High Resolution Visualizations
Pasadena, CA
Toward a Global Interactive Earth Observing CyberinfrastructureLarry Smarr
The document discusses the need for a new generation of cyberinfrastructure to support interactive global earth observation. It outlines several prototyping projects that are building examples of systems enabling real-time control of remote instruments, remote data access and analysis. These projects are driving the development of an emerging cyber-architecture using web and grid services to link distributed data repositories and simulations.
The document discusses the growing carbon footprint of information and communication technologies (ICT) and efforts to make cyberinfrastructure more energy efficient and environmentally sustainable. Specifically, it mentions that (1) ICT energy usage is growing rapidly and accounts for 2% of global greenhouse gas emissions, (2) universities are working on initiatives like the GreenLight project to reduce ICT energy usage through techniques like dynamic power management, and (3) further research is needed to develop more energy-efficient computing technologies, data center designs, and videoconferencing solutions to reduce the need for travel.
Complexity Neural Networks for Estimating Flood Process in Internet-of-Things...Dr. Amarjeet Singh
With the advancement of the Internet of Things (IoT)-based water conservation computerization, hydrological data is increasingly enriched. Considering the ability of deep learning on complex features extraction, we proposed a flood process forecastin gmodel based on Convolution Neural Network(CNN) with two-dimension(2D) convolutional operation. At first, we imported the spatial-temporal rainfall features of the Xixian basin. Subsequently, extensive experiments were carried out to determine the optimal hyperparameters of the proposed CNN flood forecasting model.
The Pacific Research Platform: a Science-Driven Big-Data Freeway SystemLarry Smarr
The Pacific Research Platform (PRP) is a multi-institutional partnership that establishes a high-capacity "big data freeway system" spanning the University of California campuses and other research universities in California to facilitate rapid data access and sharing between researchers and institutions. Fifteen multi-campus application teams in fields like particle physics, astronomy, earth sciences, biomedicine, and visualization drive the technical design of the PRP over five years. The goal of the PRP is to extend campus "Science DMZ" networks to allow high-speed data movement between research labs, supercomputer centers, and data repositories across campus, regional
In this deck from the Stanford HPC Conference, Peter Dueben from the European Centre for Medium-Range Weather Forecasts (ECMWF) presents: Machine Learning for Weather Forecasts.
"I will present recent studies that use deep learning to learn the equations of motion of the atmosphere, to emulate model components of weather forecast models and to enhance usability of weather forecasts. I will than talk about the main challenges for the application of deep learning in cutting-edge weather forecasts and suggest approaches to improve usability in the future."
Peter is contributing to the development and optimization of weather and climate models for modern supercomputers. He is focusing on a better understanding of model error and model uncertainty, on the use of reduced numerical precision that is optimised for a given level of model error, on global cloud- resolving simulations with ECMWF's forecast model, and the use of machine learning, and in particular deep learning, to improve the workflow and predictions. Peter has graduated in Physics and wrote his PhD thesis at the Max Planck Institute for Meteorology in Germany. He worked as Postdoc with Tim Palmer at the University of Oxford and has taken up a position as University Research Fellow of the Royal Society at the European Centre for Medium-Range Weather Forecasts (ECMWF) in 2017.
Watch the video: https://youtu.be/ks3fkRj8Iqc
Learn more: https://www.ecmwf.int/
and
http://www.hpcadvisorycouncil.com/events/2020/stanford-workshop/
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Global grid of master events for waveform cross correlation: design and testingIvan Kitov
This document describes the design and testing of a global grid of master events for waveform cross correlation to improve seismic monitoring capabilities. The grid uses real and synthetic master events from seismic arrays and other stations. Testing in February 2013 using 134 events found 92 matches using the grid, demonstrating its potential to detect more small events through cross correlation across the global seismic network. Future versions aim to optimize the use of real data, principal components, and machine learning to further enhance monitoring.
- Digital mirror worlds are software models of physical systems that are continuously updated with real-time data, allowing them to closely mimic and predict the behavior of the real system.
- Advances in computing power and sensors are enabling increasingly detailed digital twins of objects, human bodies, cities, wildfires, and even the observable universe.
- One trillion sensors are expected to feed the planetary computer within the next decade, driving a global industrial internet and $15 trillion in economic value through digital twins of manufactured products.
- Digital twins powered by consumer sensor data may one day provide early disease detection and personalized health coaching at scale.
National Federated Compute Platforms: The Pacific Research PlatformLarry Smarr
The Pacific Research Platform (PRP) is a multi-institution hypercluster that connects science DMZs across 25 partner campuses using FIONA data transfer nodes and 10-100Gbps networks. PRP adopted Kubernetes and Rook to orchestrate petabytes of distributed storage and GPUs for data science applications. A CHASE-CI grant added machine learning capabilities. PRP is working to federate with the Open Science Grid and become a prototype for a future National Research Platform connecting regional networks.
The Pacific Research Platform: Building a Distributed Big-Data Machine-Learni...Larry Smarr
The document summarizes the Pacific Research Platform (PRP) which connects researchers across multiple universities with high-speed networks and computing resources for big data and machine learning applications. Key points:
- PRP connects 15 universities with optical networks, distributed storage devices (FIONAs), and over 350 GPUs for data analysis and AI training.
- It allows researchers to rapidly share and analyze large datasets, with one example reducing a workflow from 19 days to 52 minutes.
- Other projects using PRP resources include climate modeling, astrophysics simulations, and machine learning courses involving thousands of students.
Analyzing Large Earth Data Sets: New Tools from the OptiPuter and LOOKING Pro...Larry Smarr
The document discusses two projects, OptIPuter and LOOKING, that aim to analyze large earth data sets using optical networking and grid technologies. OptIPuter extends grid middleware to dedicated optical circuits for earth and medical sciences. LOOKING builds on OptIPuter to provide real-time control of ocean observatories through web and grid services integrated over optical networks. Both projects represent efforts to develop cyberinfrastructure for interactive analysis of remote earth science data and instruments.
Calit2-a Persistent UCSD/UCI Framework for CollaborationLarry Smarr
05.02.16
Invited Talk
Sun Microsystems Global Education and Research
Conference 2005
Title: Calit2-a Persistent UCSD/UCI Framework for Collaboration
San Francisco, CA
Peering The Pacific Research Platform With The Great Plains NetworkLarry Smarr
The Pacific Research Platform (PRP) connects research institutions across the western United States with high-speed networks to enable data-intensive science collaborations. Key points:
- The PRP connects 15 campuses across California and links to the Great Plains Network, allowing researchers to access remote supercomputers, share large datasets, and collaborate on projects like analyzing data from the Large Hadron Collider.
- The PRP utilizes Science DMZ architectures with dedicated data transfer nodes called FIONAs to achieve high-speed transfer of large files. Kubernetes is used to manage distributed storage and computing resources.
- Early applications include distributed climate modeling, wildfire science, plankton imaging, and cancer genomics. The PR
The Pacific Research Platform: Building a Distributed Big Data Machine Learni...Larry Smarr
This document summarizes Dr. Larry Smarr's invited talk about the Pacific Research Platform (PRP) given at the San Diego Supercomputer Center in April 2019. The PRP is building a distributed big data machine learning supercomputer by connecting high-performance computing and data resources across multiple universities in California and beyond using high-speed networks. It provides researchers with petascale computing power, distributed storage, and tools like Kubernetes to enable collaborative data-intensive science across institutions.
The document discusses the Pacific Research Platform (PRP), a distributed cyberinfrastructure that connects researchers and data across multiple campuses in California and beyond using optical fiber networking. Key points:
- The PRP uses high-speed networking infrastructure like the CENIC network to connect data generators and consumers across 15+ campuses, creating an integrated "big data freeway system".
- It deploys specialized data transfer nodes called FIONAs to enable high-speed transfer of large datasets between sites at near the full network speed.
- Recent additions include using Kubernetes to orchestrate containers across the PRP infrastructure and integrating machine learning resources through the CHASE-CI grant to support data-intensive AI applications.
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save people's lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visual data collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
IRJET- A Real Time Yolo Human Detection in Flood Affected Areas based on Vide...IRJET Journal
This document proposes a method for real-time human detection in flood-affected areas using video content analysis and the YOLO object detection algorithm. It trains YOLO on the COCO Human dataset to detect and localize humans in video frames from surveillance cameras. The results show that YOLO can accurately detect multiple humans, even with occlusion, and single humans under varying illumination. This approach aims to help rescue operations quickly identify affected areas and prioritize aid.
Secure IoT Systems Monitor Framework using Probabilistic Image EncryptionIJAEMSJORNAL
In recent years, the modeling of human behaviors and patterns of activity for recognition or detection of special events has attracted considerable research interest. Various methods abounding to build intelligent vision systems aimed at understanding the scene and making correct semantic inferences from the observed dynamics of moving targets. Many systems include detection, storage of video information, and human-computer interfaces. Here we present not only an update that expands previous similar surveys but also a emphasis on contextual abnormal detection of human activity , especially in video surveillance applications. The main purpose of this survey is to identify existing methods extensively, and to characterize the literature in a manner that brings to attention key challenges.
Preliminary Evaluation of TinyYOLO on a New Dataset for Search-And-Rescue wit...Gennaro Vessio
This document discusses using the lightweight TinyYOLOv3 model for object detection on drones to assist with search and rescue operations. It proposes using drones equipped with computer vision to autonomously detect people in disaster areas to reduce human workload and increase detection rates. The researchers collected a new dataset of aerial images from mountains and beaches for testing and fine-tuned the TinyYOLOv3 model using an existing drone dataset. Preliminary results show the model can run in real-time on drone hardware and successfully detect people in the new search and rescue test images.
VIDEO STREAM ANALYSIS IN CLOUDS: AN OBJECT DETECTION AND CLASSIFICATION FRAME...Nexgen Technology
The document presents a cloud-based video analytics framework for automated object detection and classification from large volumes of video stream data. The framework analyzes video streams uploaded to cloud storage and performs computationally intensive detection and classification using GPU servers. Testing on 21,600 video streams totaling 175GB showed the framework could analyze the data in 6.52 hours using a 15 node cloud, and analysis was twice as fast (3 hours) when using GPUs.
For the full video of this presentation, please visit:
https://www.embedded-vision.com/platinum-members/pathpartner/embedded-vision-training/videos/pages/may-2019-embedded-vision-summit
For more information about embedded vision, please visit:
http://www.embedded-vision.com
Praveen Nayak, Tech Lead at PathPartner Technology, presents the "Using Deep Learning for Video Event Detection on a Compute Budget" tutorial at the May 2019 Embedded Vision Summit.
Convolutional neural networks (CNNs) have made tremendous strides in object detection and recognition in recent years. However, extending the CNN approach to understanding of video or volumetric data poses tough challenges, including trade-offs between representation quality and computational complexity, which is of particular concern on embedded platforms with tight computational budgets. This presentation explores the use of CNNs for video understanding.
Nayak reviews the evolution of deep representation learning methods involving spatio- temporal fusion from C3D to Conv-LSTMs for vision-based human activity detection. He proposes a decoupled alternative to this fusion, describing an approach that combines a low-complexity predictive temporal segment proposal model and a fine-grained (perhaps high- complexity) inference model. PathPartner Technology finds that this hybrid approach, in addition to reducing computational load with minimal loss of accuracy, enables effective solutions to these high complexity inference tasks.
ROAD POTHOLE DETECTION USING YOLOV4 DARKNETIRJET Journal
This document presents research on detecting potholes using the YOLOv4 object detection model in the Darknet framework. It begins with an introduction to the importance of road maintenance and automated pothole detection. It then describes the implementation process, which involves storing and preprocessing the dataset, training the YOLOv4 model, generating weights, and allowing users to upload images for detection. Case studies demonstrate the model successfully detecting potholes in test images. The document concludes that this method provides a cost-effective solution for government agencies to identify and repair potholes, improving road safety.
Moving Object Detection for Video SurveillanceIJMER
Video surveillance has long been in use to monitor security sensitive areas such as banks, department stores, highways, crowded public places and borders. The advance in computing power, availability of large-capacity storage devices and high speed network infrastructure paved the way for cheaper, multi sensor video surveillance systems. Traditionally, the video outputs are processed online by human operators and are usually saved to tapes for later use only after a forensic event. The increase in the number of cameras in ordinary surveillance systems overloaded both the human operators and the storage devices with high volumes of data and made it infeasible to ensure proper monitoring of sensitive areas for long times. In order to filter out redundant information generated by an array of cameras, and increase the response time to forensic events, assisting the human operators with identification of important events in video by the use of “smart” video surveillance systems has become a critical requirement. The making of video surveillance systems “smart” requires fast, reliable and robust algorithms for moving object detection, classification, tracking and activity analysis.
Human motion is fundamental to understanding behaviour. In spite of advancement on single image 3 Dimensional pose and estimation of shapes, current video-based state of the art methods unsuccessful to produce precise and motion of natural sequences due to inefficiency of ground-truth 3 Dimensional motion data for training. Recognition of Human action for programmed video surveillance applications is an interesting but forbidding task especially if the videos are captured in an unpleasant lighting environment. It is a Spatial-temporal feature-based correlation filter, for concurrent observation and identification of numerous human actions in a little-light environment. Estimated the presentation of a proposed filter with immense experimentation on night-time action datasets. Tentative results demonstrate the potency of the merging schemes for vigorous action recognition in a significantly low light environment.
Exploring visual and motion saliency for automatic video object extractionMuthu Samy
Sybian Technologies Pvt Ltd
Final Year Projects & Real Time live Projects
JAVA(All Domains)
DOTNET(All Domains)
ANDROID
EMBEDDED
VLSI
MATLAB
Project Support
Abstract, Diagrams, Review Details, Relevant Materials, Presentation,
Supporting Documents, Software E-Books,
Software Development Standards & Procedure
E-Book, Theory Classes, Lab Working Programs, Project Design & Implementation
24/7 lab session
Final Year Projects For BE,ME,B.Sc,M.Sc,B.Tech,BCA,MCA
PROJECT DOMAIN:
Cloud Computing
Networking
Network Security
PARALLEL AND DISTRIBUTED SYSTEM
Data Mining
Mobile Computing
Service Computing
Software Engineering
Image Processing
Bio Medical / Medical Imaging
Contact Details:
Sybian Technologies Pvt Ltd,
No,33/10 Meenakshi Sundaram Building,
Sivaji Street,
(Near T.nagar Bus Terminus)
T.Nagar,
Chennai-600 017
Ph:044 42070551
Mobile No:9790877889,9003254624,7708845605
Mail Id:sybianprojects@gmail.com,sunbeamvijay@yahoo.com
Exploring visual and motion saliency for automatic video object extractionMuthu Samy
Sybian Technologies Pvt Ltd
Final Year Projects & Real Time live Projects
JAVA(All Domains)
DOTNET(All Domains)
ANDROID
EMBEDDED
VLSI
MATLAB
Project Support
Abstract, Diagrams, Review Details, Relevant Materials, Presentation,
Supporting Documents, Software E-Books,
Software Development Standards & Procedure
E-Book, Theory Classes, Lab Working Programs, Project Design & Implementation
24/7 lab session
Final Year Projects For BE,ME,B.Sc,M.Sc,B.Tech,BCA,MCA
PROJECT DOMAIN:
Cloud Computing
Networking
Network Security
PARALLEL AND DISTRIBUTED SYSTEM
Data Mining
Mobile Computing
Service Computing
Software Engineering
Image Processing
Bio Medical / Medical Imaging
Contact Details:
Sybian Technologies Pvt Ltd,
No,33/10 Meenakshi Sundaram Building,
Sivaji Street,
(Near T.nagar Bus Terminus)
T.Nagar,
Chennai-600 017
Ph:044 42070551
Mobile No:9790877889,9003254624,7708845605
Mail Id:sybianprojects@gmail.com,sunbeamvijay@yahoo.com
Object Detection and Localization for Visually Impaired People using CNNIRJET Journal
The document presents a system to help visually impaired people by using CNNs (Convolutional Neural Networks) for object detection and localization based on camera images. The system aims to achieve over 95% accuracy in identifying objects captured by the camera and conveying this information to users through voice messages. It discusses existing assistive systems that use techniques like CNNs, speech recognition and custom object detection. The proposed system is intended to enable real-time object recognition and localization using a CNN model to improve awareness of the indoor environment for visually impaired individuals. Test results showed the CNN program for object recognition was implemented effectively.
This project constructed a low-cost airglow imager to involve undergraduate students in aeronomy research. Students successfully built an imaging system inside a 6-inch tube featuring a CCD, lenses, filter, temperature control, and Raspberry Pi computer. First light images from the system captured the bright OH airglow emissions over Capitol Reef National Park in Utah. The imager will support student research projects involving topics like light pollution and atmospheric waves.
A Survey on Person Detection for Social Distancing and Safety Violation Alert...IRJET Journal
This document discusses methods for monitoring social distancing using video surveillance and deep learning techniques. It describes how faster R-CNN, single shot detector (SSD) and YOLO v3 deep learning models can be used to detect people in video frames and calculate the distance between individuals to determine if social distancing guidelines are being followed. If distances between people are found to be unsafe, the system can send alerts or cautions. The methodology is intended to help prevent the spread of COVID-19 by monitoring adherence to social distancing and triggering warnings if safety violations are detected.
Global Network Advancement Group - Next Generation Network-Integrated SystemsLarry Smarr
This document summarizes a presentation on global petascale to exascale workflows for data intensive sciences. It discusses a partnership convened by the GNA-G Data Intensive Sciences Working Group with the mission of meeting challenges faced by data-intensive science programs. Cornerstone concepts that will be demonstrated include integrated network and site resource management, model-driven frameworks for resource orchestration, end-to-end monitoring with machine learning-optimized data transfers, and integrating Qualcomm's GradientGraph with network services to optimize applications and science workflows.
Desing on wireless intelligent seneor network on cloud computing system for s...csandit
Sensors on (or attached to) mobile phones can enable attractive sensing applications in
different domains such as environmental monitoring, social networking, healthcare, etc. In this
paper we propose a cloud computing system dedicated on smart home applications. We design
the proposed wireless vision sensor network (WVSN) with its algorithm and hardware
implementation. In WVSN, The partial-vision camera strategy is applied to allocate the
computation task between the sensor node and the central server. Then we propose a high
performance segmentation algorithm. Meanwhile, an efficient binary data compression method
is proposed to cope with the result on labeling information. The proposed algorithm can provide
high precision rate for the smart home applications such as the gesture recognition and
humanoid tracking. To realize the physical system, we implement it on the embedded platform
and the central server with their transmission work
Traffic Congestion Detection Using Deep Learningijtsrd
Despite the huge amount of traffic surveillance videos and images have been accumulated in the daily monitoring, deep learning approaches have been underutilized in the application of traffic intelligent management and control. Traffic images, including various illumination, weather conditions, and vast scenarios are considered and preprocessed to set up a proper training dataset. In order to detect traffic congestion, a network structure is proposed based on residual learning to be pre trained and fine tuned. The network is then transferred to the traffic application and retrained with self established training dataset to generate the Traffic Net. The accuracy of Traffic Net to classify congested and uncongested road states reaches 99 for the validation dataset and 95 for the testing dataset. The trained model is stored in cloud storage for easy access for application from anywhere. The proposed Traffic Net can be used by a regional detection of traffic congestion on a large scale surveillance system. The effectiveness and efficiencies are magnificently demonstrated with quick detection in the high accuracy in the case study. The experimental trial could extend its successful application to traffic surveillance system and has potential enhancement for intelligent transport system in future. Anusha C | Dr. J. Bhuvana "Traffic Congestion Detection Using Deep Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49401.pdf Paper URL: https://www.ijtsrd.com/computer-science/other/49401/traffic-congestion-detection-using-deep-learning/anusha-c
Similar to Crowdsourcing the Acquisition and Analysis of Mobile Videos for Disaster Response (20)
The first 24 hours after disaster are critical
Timely collecting and accurately analyzing data from multiple sources are important
Providing an overview of the situation for the decision makers is critical
In recent years, augmented-reality (AR) has been
attracting extensive attentions from both the research community
and industry as a new form of media, mixing virtual content into
the physical world. However, the scarcity of the AR content and
the lack of user contexts are major impediments to providing
and representing rich and dynamic multimedia content on AR
applications. In this study, we propose an approach to search
and filter big multimedia data, specifically geo-tagged mobile
videos, for context-aware AR applications. The challenge is to
automatically search for interesting video segments out of a
huge amount of user-generated mobile videos, which is one of
the biggest multimedia data, to be efficiently incorporated into
AR applications. We model the significance of video segments
as AR content adopting camera shooting patterns defined in
filming, such as panning, zooming, tracking and arching. Then,
several efficient algorithms are proposed to search for such
patterns using fine granular geospatial properties of the videos
such as camera locations and viewing directions over time.
Experiments with real-world geo-tagged video dataset show that
the proposed algorithms effectively search for a large collection of
user-generated mobile videos to identify top K significant video
segments.
Spatial Crowdsourcing (SC) is a transformative platform that engages individuals, groups and communities in the act of collecting, analyzing, and disseminating environmental, social and other spatio-temporal information. The objective of SC is to outsource a set of spatio-temporal tasks to a set of workers, i.e., individuals with mobile devices that perform the tasks by physically traveling to specified locations of interest. However, current solutions require the workers, who in many cases are simply volunteering for a cause, to disclose their locations to untrustworthy entities. In this paper, we introduce a framework for protecting location privacy of workers participating in SC tasks. We argue that existing location privacy techniques are not sufficient for SC, and we propose a mechanism based on differential privacy and geocasting that achieves effective SC services while offering privacy guarantees to workers. We investigate analytical models and task assignment strategies that balance multiple crucial aspects of SC functionality, such as task completion rate, worker travel distance and system overhead. Extensive experimental results on real-world datasets show that the proposed technique protects workers' location privacy without incurring significant performance metrics penalties.
Link: http://dl.acm.org/citation.cfm?id=2732966
Histograms have been extensively used for selectivity estimation by academics and have successfully been adopted by database industry. However, the estimation error is usually large for skewed distributions and biased attributes, which are typical in real-world data. Therefore, we propose effective models to quantitatively measure bias and selectivity based on information entropy. These models together with the principles of maximum entropy are then used to develop a class of entropy-based histograms. Moreover, since entropy can be computed incrementally, we present the incremental variations of our algorithms that reduce the complexities of the histogram construction from quadratic to linear. We conducted an extensive set of experiments with both synthetic and real-world datasets to compare the accuracy and efficiency of our proposed techniques with many other histogram-based techniques, showing the superiority of the entropy-based approaches for both equality and range queries.
Links:
http://dl.acm.org/citation.cfm?id=2505756
http://infolab.usc.edu/DocsDemos/to-CIKM13.pdf
With the popularity of mobile devices, spatial crowdsourcing is rising as a new framework that enables human workers to solve tasks in the physical world. With spatial crowdsourcing, the goal is to crowdsource a set of spatiotemporal tasks (i.e., tasks related to time and location) to a set of workers, which requires the workers to physically travel to those locations in order to perform the tasks. In this article, we focus on one class of spatial crowdsourcing, in which the workers send their locations to the server and thereafter the server assigns to every worker tasks in proximity to the worker’s location with the aim of maximizing the overall number of assigned tasks. We formally define this maximum task assignment (MTA) problem in spatial crowdsourcing, and identify its challenges. We propose alternative solutions to address these challenges by exploiting the spatial properties of the problem space, including the spatial distribution and the travel cost of the workers. MTA is based on the assumptions that all tasks are of the same type and all workers are equally qualified in performing the tasks. Meanwhile, different types of tasks may require workers with various skill sets or expertise. Subsequently, we extend MTA by taking the expertise of the workers into consideration. We refer to this problem as the maximum score assignment (MSA) problem and show its practicality and generality. Extensive experiments with various synthetic and two real-world datasets show the applicability of our proposed framework.
Links:
http://dl.acm.org/citation.cfm?id=2729713
http://infolab.usc.edu/DocsDemos/to_TSAS15.pdf
http://dl.acm.org/citation.cfm?doid=2729713
Spatial crowdsourcing has gained emerging interest from both research communities and industries. Most of current spatial crowdsourcing frameworks assume independent and atomic tasks. However, there could be some cases that one needs to crowdsource a spatial complex task which consists of some spatial sub-tasks (i.e., tasks related to a specific location). The spatial complex task's assignment requires assignments of all of its sub-tasks. The currently available frameworks are inapplicable to such kind of tasks. In this paper, we introduce a novel approach to crowdsource spatial complex tasks. We first formally define the Maximum Complex Task Assignment (MCTA) problem and propose alternative solutions. Subsequently, we perform various experiments using both real and synthetic datasets to investigate and verify the usability of our proposed approach.
Links:
http://dl.acm.org/citation.cfm?id=2539243
http://dl.acm.org/citation.cfm?id=2539243
http://dl.acm.org/citation.cfm?doid=2539150.2539243
This document discusses research on spatial crowdsourcing and differential privacy. It introduces the PriGeoCrowd framework which aims to preserve worker location privacy in spatial crowdsourcing using differential privacy. The framework uses a WorkerPSD technique to guide efficient task assignment and differentially private spatial decompositions. It also proposes a geocast query algorithm that expands the query using the WorkerPSD to navigate worker search and assignment.
We propose some interesting ideas, e.g., small amount of metadata is transmitted to a server in real-time while video content will remain on the device; the server will ping the device when it wants the user to upload the video.
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save people's lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visual data collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
Motivation: the popularity of smart phones, and availability of data allow us to create an application that provide students with useful services, and allow them to share their experience
Integrate various types of data, including public data (Google Map, Facebook, Twitter), user data (picture, location), USC's private data (event, alert)
Uses the four dimensions of "what, when, where and who“, and allow USC community members to easily add their own apps
Mending Clothing to Support Sustainable Fashion_CIMaR 2024.pdfSelcen Ozturkcan
Ozturkcan, S., Berndt, A., & Angelakis, A. (2024). Mending clothing to support sustainable fashion. Presented at the 31st Annual Conference by the Consortium for International Marketing Research (CIMaR), 10-13 Jun 2024, University of Gävle, Sweden.
(June 12, 2024) Webinar: Development of PET theranostics targeting the molecu...Scintica Instrumentation
Targeting Hsp90 and its pathogen Orthologs with Tethered Inhibitors as a Diagnostic and Therapeutic Strategy for cancer and infectious diseases with Dr. Timothy Haystead.
PPT on Direct Seeded Rice presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
PPT on Alternate Wetting and Drying presented at the three-day 'Training and Validation Workshop on Modules of Climate Smart Agriculture (CSA) Technologies in South Asia' workshop on April 22, 2024.
Anti-Universe And Emergent Gravity and the Dark UniverseSérgio Sacani
Recent theoretical progress indicates that spacetime and gravity emerge together from the entanglement structure of an underlying microscopic theory. These ideas are best understood in Anti-de Sitter space, where they rely on the area law for entanglement entropy. The extension to de Sitter space requires taking into account the entropy and temperature associated with the cosmological horizon. Using insights from string theory, black hole physics and quantum information theory we argue that the positive dark energy leads to a thermal volume law contribution to the entropy that overtakes the area law precisely at the cosmological horizon. Due to the competition between area and volume law entanglement the microscopic de Sitter states do not thermalise at sub-Hubble scales: they exhibit memory effects in the form of an entropy displacement caused by matter. The emergent laws of gravity contain an additional ‘dark’ gravitational force describing the ‘elastic’ response due to the entropy displacement. We derive an estimate of the strength of this extra force in terms of the baryonic mass, Newton’s constant and the Hubble acceleration scale a0 = cH0, and provide evidence for the fact that this additional ‘dark gravity force’ explains the observed phenomena in galaxies and clusters currently attributed to dark matter.
Evidence of Jet Activity from the Secondary Black Hole in the OJ 287 Binary S...Sérgio Sacani
Wereport the study of a huge optical intraday flare on 2021 November 12 at 2 a.m. UT in the blazar OJ287. In the binary black hole model, it is associated with an impact of the secondary black hole on the accretion disk of the primary. Our multifrequency observing campaign was set up to search for such a signature of the impact based on a prediction made 8 yr earlier. The first I-band results of the flare have already been reported by Kishore et al. (2024). Here we combine these data with our monitoring in the R-band. There is a big change in the R–I spectral index by 1.0 ±0.1 between the normal background and the flare, suggesting a new component of radiation. The polarization variation during the rise of the flare suggests the same. The limits on the source size place it most reasonably in the jet of the secondary BH. We then ask why we have not seen this phenomenon before. We show that OJ287 was never before observed with sufficient sensitivity on the night when the flare should have happened according to the binary model. We also study the probability that this flare is just an oversized example of intraday variability using the Krakow data set of intense monitoring between 2015 and 2023. We find that the occurrence of a flare of this size and rapidity is unlikely. In machine-readable Tables 1 and 2, we give the full orbit-linked historical light curve of OJ287 as well as the dense monitoring sample of Krakow.
Evidence of Jet Activity from the Secondary Black Hole in the OJ 287 Binary S...
Crowdsourcing the Acquisition and Analysis of Mobile Videos for Disaster Response
1. 1
Crowdsourcing the Acquisition and Analysis of
Mobile Videos for Disaster Response
IEEE Big Data 2015, October 31, 2015
Presented by Hien To
Integrated Media Systems Center
University of Southern California
Dr. Seon Ho Kim Prof. Cyrus Shahabi
2. 2
GeoQ by National Geospatial-Intelligence Agency (NGA)
GeoQ is a crowdsourcing platform to analyze disaster dataEach project contains a set of jobsEach job contains a set of work cells, which are assigned to analystsAnalysts evaluate mobile videos within their assigned work cells
3. 3
Motivation
Obtain real-time information to enhance
situational awareness during and after disasters
Live streaming of mobile video, e.g., YouTube,
Periscope
Analyze collected data in timely manner
Prioritize data collection with limited transmission
4. 4
Haiti Earthquake
Disasters
Assess damage across large geographical area
Cable Cuts in Vietnam
Disrupted communication networks or data overload!
Haiti earthquake, 85% Haitians could still access to their mobile
phones but 70% cell towers were destroyed
5. 5
Crowdsourcing Disaster Response
Early disaster data collection focus on VGI
Disaster surveillance and response system
Processing and integration of disaster data by
leveraging Linked Open Data
[Goodchild et al. 2010]
[Chu et al. 2012]
[Ortmann et al. 2011]
Combine human and machine intelligence [Schulz et al. 2012]
Focuses on both data collection and data analysis
Collects huge amount of videos concerning disasters
Improves resilience and responsiveness of computer systems
Our study:
Infrastructure damage caused by Earthquake [Kobayashi 2014]
Resilient communication technologies, e.g., drones [Nemoto and
Hamaguchi 2014]
7. 7
MediaQ and GeoQ
MediaQ [1]: mobile crowdsourcing
platform to collect videos/images
FOV Metadata:
p: camera location
d: camera direction
α: viewable angle
R: viewable distance
t: timestamp
𝑁
α
𝑑
p
R
Field Of View (FOV)
GeoQ [2]: crowdsourcing platform to
analyze disaster data
[1] By USC Integrated Media System Center
[2] By US National Geospatial-Intelligence Agency (NGA)
GeoQ’s Work Cells used in
Nepal Earthquake 2015
MediaQ workers are on-site people; GeoQ workers are off-site analysts.
MediaQ provides visual data to GeoQ.
Data Collection Data Analysis
Work cells are assigned to analysts
8. 8
Proposed Techniques
Coverage Map
Video metadata
Upload metadata (FOV) first then
identify videos with high visual
awareness to be uploaded
Partition space based on
video locations
Video content
Collect data fast because time
is critical
Analyze the collected data
effectively
9. 9
Proposed Framework
0 0
0
1 3
2 5
2
0 0
4. Evaluate
work cells
3. Assign videos to analysts
1. Crowdsource
videos from
community
Urgency Map
Analyst
Server
2. Upload relevant videos
Coverage Map Space Decomposition
Video content
Video metadata
10. 10
1. Acquisition of Video Metadata and Content
Real-time acquisition of metadata via Internet, SMS, WLAN
Size of metadata is thousands times smaller than content
Hold video in the device and have server ping the device
when it wants the owner to upload the video
Real-time analysis on the uploaded metadata, e.g.,
visualization
Support data governance, preserve privacy and strict
access control [Dey and Chinchwadkar, ICDE 2015]
11. 11
2. Visual Awareness Maximization
Visual Awareness (VA) of a Video
f1
f2
f3
VA(v) =Urgency(w)
area(v)
area(w)
Selects videos that maximize total VA without exceeding budget
(i.e., constrained bandwidth)
This problem is 0-1 Knapsack.
Video location
Work cell
A video has high VA if
Enclosing work cell is urgent
Aggregate video coverage is large
3-seconds video within a work cell
12. 12
3. Assign Videos to Analysts
Each analyst handles equal number of videos
Terminates when #workcells > #analysts-3
Quadtree (Gaussian dist.)Kd-tree (Gaussian dist.)
GeoQ uses of grid-based partition
Some analysts may be overloaded
Space partitioning algorithm based on Quadtree and Kd-tree
13. 13
Minimum Redundant Coverage
Upload key frames and metadata to server and minimize
redundant coverage
Select frames that maximizes total Visual Awareness without
exceeding budget
This problem is Maximum Coverage Problem.
f2
f1 f3
FOV overlap
f4
14. 14
Statistics Value
Spatial distributions Uniform, Gaussian and Zipfian
#videos 1000 with varying length
Region size 10 × 10 square km in Los Angeles area
Unit grid cells 500×500
FOV # per second 1
Viewable angle 60 degrees
Visible distance 200 meters
Synthetic Datasets
Simulation Results
1. Partitioning techniques: Grid, Quadtree, Kd-tree
2. Visual awareness at video level vs. frame level
[Ay et. Al. SIGSPATIAL’2010]
15. 15
Visual Awareness Maximization at Video Level
Varying the number of analysts
Uniform Dataset
0
0.5
1
1.5
2
2.5
3
3.5
16 25 36 49 64
VisualAwareness
Analyst count
Grid
Quadtree
Kd-tree
Kd-tree increases visual
awareness by up to 300%
0.1
1
10
100
1000
Grid Quadtree Kd-tree
Variance(logscale)
Uniform
Gaussian
Zipfian
Variance of #videos per analyst
Kd-tree produces similar
#videos per work cell
Gaussian Dataset
16. 16
Frame Level
0
20
40
60
80
100
120
16 25 36 49 64
VisualAwareness
Analyst count
Grid
Quadtree
Kd-tree
Obtained visual awareness is an
order of magnitude higher in
comparison to video-level opt
Runtime Measurements
5.2
7
7.4
0
1
2
3
4
5
6
7
8
Grid Quadtree Kd-tree
Constructiontime(ms)
Construction times are
small and the differences
are insignificant
Uniform Dataset Uniform Dataset
O(n)
O(nlogn)O(logAn)
17. 17
Conclusion
Proposed crowdsourcing framework for collection and
analysis of video data under disaster situations
Introduced more effective spatial decomposition to assign
videos to analysts
Developed an analytical model to quantify the visual
awareness of a particular video or frame
Introduced two variants visual awareness maximization
problem: uploading the entire videos and individual frames
Introduced a mechanism to enable time-sensitive acquisition
and analysis of spatial metadata
20. 20
Location/Region Entropy
Location entropy (LE)
– Measures the diversity of unique visitors of a location
– High if many users were observed at the location with equal
proportion
Cranshaw, J., et al., (2010). Bridging the gap between physical locations and online social networks. UBICOMP, 119-
128
1u 2u
1l
2l
1000 1
500 500
LE
0.35
0.69
lOlFreq )(
luOlUserCount ,)(
)(log)()( uPuPlLE
lUu ll
l
lu
l
O
O
uP
,
)(where
: total number of visits
: number of visits by worker u
21. 21
Work Cell Urgency
Urgency(w)= Priority(w)a +RegionEntropy(w.r)b
Compute urgency value of a work cell
Region Entropy – geosocial factor
of a region. RE is high if the region
are visited by many people many
times, e.g., school, hospital.
USC Campus
The priority of an area,
e.g., nuclear plant areas
are more important than
residence areas.
Nuclear plant
22. 22
Future Work
1. Use user-generated videos from
MediaQ
2. Use near-real-time ShakeMap
from USGS
23. 23
Data Analysis
Work Cell Analysis
Analysts measure the severity of their assigned work cells and set
urgency values to them
Urgency of the work cells can be automatically calculated
Editor's Notes
Collecting disaster data efficiently and analyzing these data effectively are important for disaster response because it can save people lives.
There are many kinds of disaster-related data, text, audio, in which videos and images are most effective in understanding the disaster situation because they can be watched and easily understood by people, who may not know the local language.
However, collecting big video data is challenging because communication network can be damaged during disaster. Also, having the analysts to process the videos quickly is important
In this study, we focus on crowdsourcing mobile videos during and after disaster and solving such 2 unique challenges
The application we are interested assessing damage across large geographical area of disaster.
There are many kinds of disaster, natural disaster or man-made disaster.
In all of these disasters either communication network is disrupted or we may have the issue of having too many data. For example, in Haitii earthquake, 85% Haitians could still access to their mobile phones but 70% of cell towers were destroyed. In such a case, uploading a lot of user-generated videos to the server is challenging.
Crowdsourcing disaster response is a growing research area.
Crowdsourcing has been shown to be a cost-effective and time-efficient way to acquire disaster data. There have been many studies on this topic, just to name a few here.
However, these studies focus on either data collection or data analysis while our study aims to deal with both issues. We will propose a framework that collects a huge amount of videos concerning disasters. The framework improves resilience and responsiveness of computer systems to facilitate realtime data collection.
There are lots of efforts in disaster response.
We now introduce two existing platforms that inspire our study, one for data acquisition and one for data analysis.
MediaQ is mobile crowdsourcing platform to collect video metadata and content. The geospatial metadata are transparently associated with each video frame. We represent a video as a sequence of frames, each frame is modeled as a field of view or FOV. An FOV includes camera location, direction, viewable angle and viewable distance and timestamp.
On the other hand, GeoQ is another crowdsourcing platform that allows analysts to assess disaster damage across a large area. The idea is to divide disaster area into small regions namely work cells and assign them to the analysts. The role of the analysts is to evaluate the available data sources in their allocated work cells to identify relevant incident, for example a building on fire, or road block.
Two reasons MediaQ and GeoQ are compliment to each others.
We identify two challenges with MediaQ and GeoQ.
The first one is to collect the data fast because in disaster time is critical. Also bandwidth is limited during and after disaster, so we may not be able to upload all videos.
So there is a need of uploading metadata first. Then, from the metadata, we identify videos to be collected. We prefer the videos with high visual awareness.
The second challenge is to analyze the collected data effectively. GeoQ makes use of grid-based partition. However, the shortcoming of such approach is that depending on the data distribution, some analysts may be assigned the workcells that containing a lots of videos while the others are assigned zero videos. So we propose to partition space based on video locations.
We now introduce a unified framework that adopt such techniques.
The framework has 4 steps.
First, videos can be collected voluntarily from community or on-demand manner.
When users record a video, its metadata is automatically captured and uploaded to the server but not the actual content. Then, basing on the uploaded metadata, the server uploads the potentially relevant videos, given a bandwidth constraint.
The server will then assign the collected data to the analysts by partitioning the space into small regions namely work cells.
Finally, the analysts watch the videos in their workcells and assign an importance value to the workcell. The higher the value, the more urgent the situation in the cell. The urgency values can also be automatically calculated. However, we will not discuss here, you can find details in our paper.
I will talk about steps 1,2,3 in details.
We propose the idea of holding the video in the device, and having a server ping the device when it wants the owner to upload the video. Being small in size, the metadata can be can be transmitted in realtime through various communication channels such as Internet, SMS and WLAN. To illustrate, the size of metadata of a ten-seconds video is only 1-3KB with a sampling rate of one FOV per second. However, the size of a video is typically a few MBs.
This idea has various applications, for example showing the coverage map of the videos from the collected metadata. Other reasons for handling metadata separately are shown in ICDE 1015 paper, including supporting data governance as metadata often lives longer than its content, preserving privacy and strict access control.
Based on the collected metadata, we will identify the potentially relevant videos to be uploaded. We define visual awareness as the importance of a video. It is important to compute visual awareness based on the geospatial metadata only because analyzing the actual content may be expensive in terms of human or computational resources.
Urgency(w) is the urgency value of the workcell, between 0 and 5 and given by the analyst. The aggregated the coverage area of the video wrt the containing work cell is a value between 0 and 1. This equation means that the visual awareness of a video is high if the workcell is urgent OR the coverage area of the video is large.
We formulate the problem of selecting a set of videos with the maximum total visual awareness under bandwidth constraints. We prove that this problem is 0-1 knapsack. So, we use dynamic programming to optimally solve it.
When the videos are uploaded, they are distributed to analysts. GeoQ use equal-size grid, w soorkload distribution to analysts are not even. Some analysts may be assigned a lots of videos while the others are assigned zero video.
Because human analysts are the most expensive resource, to facilitate timely evaluation on the acquired data, we propose an algorithm for space partitioning based on Quadtree and Kd-tree. The algorithm adaptively decomposes the space based on the spatial distribution of the videos. We modify the traditional point quadtree or kd-tree such that the algorithm terminates when the number of workcells is larger than the number of analysts. By doing so, each workcell can be assigned to at least one analyst. As the result, each analyst has similar number of videos to analyze and no one becomes the bottleneck in collective crowdsourcing efforts.
Our last technique is to upload individual frames rather than entire video. The reason is that, many times several frames from a video are enough to understand the situation. Also, time is critical. So, we propose to upload only keyframes to the server. At the same time, we also minimize redundant coverage of these frames to reduce the bandwidth usage.
In this example, if we can select two FOVs to maximize total coverage, those would be f1 and f3.
Our goal now is to select a set of frames that maximizes total VA given a budget constraint. We formally define the problem and prove that it is NP-hard by reduction from the MCP problem. Details can be found in the paper. So, we adopt the greedy algorithm, which is the best-possible polynomial time approximation algorithm for MCP.
More analysts look at the same amount of video VA increases
More analysts work cells become smaller VA increases
The improvement of kd-tree over uniform grid is clear because Grid does not adapt to the distribution of the data
The improvement over quadtree is because kd-tree use median split while quadtree partition based on space
The improvement is even bigger in Gaussian and Zipfian distributions
Key observation VA of all techniques is higher by 1 order of magnitude because we can avoid a lot of redundant coverage in the case of uploading entire videos.
Construction time in memory is small construction time is not our main concern.
Take away messages:
Uploading keyframes provide much higher VA when compared to uploading entire videos
Good to go with kd-tree because the overhead is small when compared with the benefits