Mapping the Tohoku 2011 Tsunami event with a remote sensing satellite constellation – a reference case for emerging Early Warning System Dissemination Services
This document discusses how tsunami early warning systems could provide information to satellite operators to help speed up crisis mapping after disasters. It describes a project called TRIDEC that aims to integrate tsunami warnings with satellite tasking to allow imaging of affected areas sooner. During the 2011 Tohoku tsunami, satellite imagery through the International Charter helped create maps for rescue efforts. Faster coordination between warnings and satellite tasking could produce maps even sooner to further aid response. Standard messaging formats may help disseminate early warnings to satellite operators for quicker crisis mapping following disasters.
GEO Expert Advisory Group - ESA Thematic Exploitation Platforms - Geohazardsterradue
The second EAG (Expert Advisory Group) meeting was held on February 5th, 2019 in Geneva. Terradue as EAG member was invited to present on solutions supporting the GEO vision for Knowledge Hubs
Geohazards Exploitation Platform (GEP) at EuroGEOSS Workshop 2018terradue
GEP provides large scale processing of Earth Observation data.
Designed in the context of the Geohazards Supersite initiative (GSNL) and the CEOS Disasters Working Group which address a Task of the Disaster Societal Benefit Area of the intergovernmental Group on Earth Observations (GEO).
A model for partnership and community building that is user driven. Started from the International Forum on Satellite EO and Geohazards organised by ESA and GEO in Santorini in 2012 (140+ participants from 20 countries, 70+ organisations incl. international organisations, public institutes, space agencies, universities & private sector).
Big Data from Space outreach presentation describing the "Bring User to Data Scenario" and the Research and Service Support provided for scientific users, research centres, small and medium size enterprises, industry and institutions.
Small Satellites and Earth Observation. The UPC NanoSat programadrianocamps
Small Satellites and Main Applications, with special focus on Earth Observation. The UPC NanoSat program: 3Cat-1 to 3Cat-5/A-B (FFSCAT). Current Trends in Earth Observation Missions with Small Satellites.
Disclaimer: these materials were prepared for Eduacational purposes only.
GEO Expert Advisory Group - ESA Thematic Exploitation Platforms - Geohazardsterradue
The second EAG (Expert Advisory Group) meeting was held on February 5th, 2019 in Geneva. Terradue as EAG member was invited to present on solutions supporting the GEO vision for Knowledge Hubs
Geohazards Exploitation Platform (GEP) at EuroGEOSS Workshop 2018terradue
GEP provides large scale processing of Earth Observation data.
Designed in the context of the Geohazards Supersite initiative (GSNL) and the CEOS Disasters Working Group which address a Task of the Disaster Societal Benefit Area of the intergovernmental Group on Earth Observations (GEO).
A model for partnership and community building that is user driven. Started from the International Forum on Satellite EO and Geohazards organised by ESA and GEO in Santorini in 2012 (140+ participants from 20 countries, 70+ organisations incl. international organisations, public institutes, space agencies, universities & private sector).
Big Data from Space outreach presentation describing the "Bring User to Data Scenario" and the Research and Service Support provided for scientific users, research centres, small and medium size enterprises, industry and institutions.
Small Satellites and Earth Observation. The UPC NanoSat programadrianocamps
Small Satellites and Main Applications, with special focus on Earth Observation. The UPC NanoSat program: 3Cat-1 to 3Cat-5/A-B (FFSCAT). Current Trends in Earth Observation Missions with Small Satellites.
Disclaimer: these materials were prepared for Eduacational purposes only.
THEOS (Thailand Earth Observation System)
THEOS is an Earth observation mission of Thailand, under development at EADS Astrium SAS, Toulouse, France. In July 2004, EADS Astrium SAS signed a contract for delivery of THEOS with GISTDA (Geo-Informatics and Space Technology Development Agency) of Bangkok, Thailand. GISTDA is Thailand's leading national organization (i.e., space agency) in the field of space activities and applications. The Thai Ministry of Science and Technology is funding the program.
The THEOS cooperative agreement includes the production and launch of one optical imaging satellite, as well as the development of the ground segment necessary to operate and control the satellite directly from Thailand. The contract also specifies on-the-job training of Thai engineers as part of the EADS Astrium development team. Also as part of the THEOS program, GISTDA earned the right to receive data from the SPOT-2, 4 and 5 spacecraft of CNES in Thailand, which have many features similar to those of THEOS. The prime objective of THEOS is to provide Thailand with an affordable access to space (i.e., a state-of-the-art Earth observation satellite for the near future), and to spawn through this program's operational experience the country's own capability and infrastructure an indigenous potential for the development of future space missions.
The science objectives call for the provision of:
1) Panchromatic (2 m) and multispectral (15 m) imagery from THEOS observations, and
2) The generation of geo-referenced image products and image processing capabilities for applications in the fields of cartography, land use, agricultural monitoring, forestry management, coastal zone monitoring and flood risk management.
The Thai government has also expressed its intention to offer THEOS data to the disaster mitigation efforts under the International Charter.
GWT 2014: Emergency Conference - 08 I servizi satellitari in near real-time a...Planetek Italia Srl
Geospatial World Tour 2014: Emergency Conference.
Napoli, 28 maggio 2014.
I servizi satellitari in near real-time del Marsec a supporto delle emergenze.
MARSec SpA
Module 10 - Section 4: ICTs for understanding and monitoring the environment ...Richard Labelle
Slide presentations developed to demonstrate how Information and Communication Technologies (ICTs) be used to address climate change, and why ICTs are a crucial part of the solution – i.e. in promoting efficiency, Green Growth & sustainable development, in dealing with climate change and for climate and environmental action. These slide presentations were delivered in February 2011 in Seongnam, near Seoul in Korea.
These presentations were developed and delivered over 2.5 days on the occasion of a Regional Training of Trainers Workshop for upcoming Academy modules on ICT for Disaster Risk Management and Climate Change Abatement. These modules were developed as part of the Academy of ICT Essentials for Government leaders developed by the United Nations (UN) Asia Pacific Centre for ICT Training (APCICT), based in Songdo City, in the Republic of South Korea.
These presentations were developed in 2011, and are somewhat out of date, but most of the principles still apply. Module 10, which has been published, does not include much of the information outlined in these presentations, which are fairly technical. They were developed to address a significant gap in understanding of the technical basis of using ICTs for climate action and because there is a clear bias in development circles against the importance of dealing with climate change mitigation in developing countries. These presentations are an attempt to redress this lack and are published here with this purpose in mind.
The author, Richard Labelle, is presently working on updating these presentations to further highlight the importance of addressing climate change and the important role that technology including ICTs, play in this effort.
Application of Remote Sensing in Civil EngineeringIEI GSC
Presentation cum talk delivered by Dr Anjana Vyas, Dean CEPT University, Ahmedabad during 31st National Convention of Civil Engineering organized by The Institution of Engineers (India) Gujarat State Center, Ahmedabad
Responding to an oil spill requires access and understanding of many types of information. Effective, coordinated operations for the response are based on a shared, common picture of the situation. Interoperability provides shared situational awareness of the crisis and the response activities.
The OGP and IPIECA are conducting a Joint Industry Project to produce a recommended practice for an Oil Spill Response Common Operating Picture (COP) for management of the response. The presentation will provide an overview, plans and status of the OGP/IPEICA project being conducted with support from RDA and OGC.
THEOS (Thailand Earth Observation System)
THEOS is an Earth observation mission of Thailand, under development at EADS Astrium SAS, Toulouse, France. In July 2004, EADS Astrium SAS signed a contract for delivery of THEOS with GISTDA (Geo-Informatics and Space Technology Development Agency) of Bangkok, Thailand. GISTDA is Thailand's leading national organization (i.e., space agency) in the field of space activities and applications. The Thai Ministry of Science and Technology is funding the program.
The THEOS cooperative agreement includes the production and launch of one optical imaging satellite, as well as the development of the ground segment necessary to operate and control the satellite directly from Thailand. The contract also specifies on-the-job training of Thai engineers as part of the EADS Astrium development team. Also as part of the THEOS program, GISTDA earned the right to receive data from the SPOT-2, 4 and 5 spacecraft of CNES in Thailand, which have many features similar to those of THEOS. The prime objective of THEOS is to provide Thailand with an affordable access to space (i.e., a state-of-the-art Earth observation satellite for the near future), and to spawn through this program's operational experience the country's own capability and infrastructure an indigenous potential for the development of future space missions.
The science objectives call for the provision of:
1) Panchromatic (2 m) and multispectral (15 m) imagery from THEOS observations, and
2) The generation of geo-referenced image products and image processing capabilities for applications in the fields of cartography, land use, agricultural monitoring, forestry management, coastal zone monitoring and flood risk management.
The Thai government has also expressed its intention to offer THEOS data to the disaster mitigation efforts under the International Charter.
GWT 2014: Emergency Conference - 08 I servizi satellitari in near real-time a...Planetek Italia Srl
Geospatial World Tour 2014: Emergency Conference.
Napoli, 28 maggio 2014.
I servizi satellitari in near real-time del Marsec a supporto delle emergenze.
MARSec SpA
Module 10 - Section 4: ICTs for understanding and monitoring the environment ...Richard Labelle
Slide presentations developed to demonstrate how Information and Communication Technologies (ICTs) be used to address climate change, and why ICTs are a crucial part of the solution – i.e. in promoting efficiency, Green Growth & sustainable development, in dealing with climate change and for climate and environmental action. These slide presentations were delivered in February 2011 in Seongnam, near Seoul in Korea.
These presentations were developed and delivered over 2.5 days on the occasion of a Regional Training of Trainers Workshop for upcoming Academy modules on ICT for Disaster Risk Management and Climate Change Abatement. These modules were developed as part of the Academy of ICT Essentials for Government leaders developed by the United Nations (UN) Asia Pacific Centre for ICT Training (APCICT), based in Songdo City, in the Republic of South Korea.
These presentations were developed in 2011, and are somewhat out of date, but most of the principles still apply. Module 10, which has been published, does not include much of the information outlined in these presentations, which are fairly technical. They were developed to address a significant gap in understanding of the technical basis of using ICTs for climate action and because there is a clear bias in development circles against the importance of dealing with climate change mitigation in developing countries. These presentations are an attempt to redress this lack and are published here with this purpose in mind.
The author, Richard Labelle, is presently working on updating these presentations to further highlight the importance of addressing climate change and the important role that technology including ICTs, play in this effort.
Application of Remote Sensing in Civil EngineeringIEI GSC
Presentation cum talk delivered by Dr Anjana Vyas, Dean CEPT University, Ahmedabad during 31st National Convention of Civil Engineering organized by The Institution of Engineers (India) Gujarat State Center, Ahmedabad
Application of Remote Sensing in Civil Engineering
Similar to Mapping the Tohoku 2011 Tsunami event with a remote sensing satellite constellation – a reference case for emerging Early Warning System Dissemination Services
Responding to an oil spill requires access and understanding of many types of information. Effective, coordinated operations for the response are based on a shared, common picture of the situation. Interoperability provides shared situational awareness of the crisis and the response activities.
The OGP and IPIECA are conducting a Joint Industry Project to produce a recommended practice for an Oil Spill Response Common Operating Picture (COP) for management of the response. The presentation will provide an overview, plans and status of the OGP/IPEICA project being conducted with support from RDA and OGC.
UAVs are a disruptive technology bringing new geographic data and information to many application domains. UASs are similar to other geographic imagery systems so existing frameworks are applicable. But the diversity of UAVs as platforms along with the diversity of available sensors are presenting challenges in the processing and creation of geospatial products. Efficient processing and dissemination of the data is achieved using software and systems that implement open standards. The challenges identified point to the need for use of existing standards and extending standards. Results from the use of the OGC Sensor Web Enablement set of standards are presented. Next steps in the progress of UAVs and UASs may follow the path of open data, open source and open standards.
Application packaging and systematic processing in earth observation exploita...terradue
An overview of Terradue's solutions supporting Earth Observations (EO) Exploitation Platforms across multiple domains.
Presentation done as part of the Open Geospatial Consortium (OGC) Technical Committee ad-hoc meeting for the setup of a new domain working group on EO Exploitation Platforms.
The Evolution of Disaster Early Warning Systems in the TRIDEC Project Peter Löwe
The TRIDEC project (Collaborative, Complex, and Critical Decision Processes in Evolving Crises) focuses on real-time intelligent information management in the Earth management domain and its long-term applications. It is funded under the European Union’s seventh Framework Programme (FP7). The TRIDEC software framework is applied in two application environments, which include industrial subsurface drilling (ISD) and natural crisis management (NCM).
For each domain, three consecutive demonstrators with extended capabilities are developed and field-tested during the projects lifespan. This article focuses on the technical advances achieved by the light-, mid- and heavyweight NCM demonstrators for Tsunami Early Warning.
Disaster Risk Management ‘enlarged actions’ by Ivan Petiteville, co-chair, CE...Global Risk Forum GRFDavos
Presentation at the Consultion Day event about: Scientific and Technical Platforms / Networks: Achievements and Future Goals during the Global Platform for Disaster Risk Reduction GPDRR 2013 in Geneva
Efficient and thorough data collection and its timely analysis are critical for disaster response and recovery in order to save people's lives during disasters. However, access to comprehensive data in disaster areas and their quick analysis to transform the data to actionable knowledge are challenging. With the popularity and pervasiveness of mobile devices, crowdsourcing data collection and analysis has emerged as an effective and scalable solution. This paper addresses the problem of crowdsourcing mobile videos for disasters by identifying two unique challenges of 1) prioritizing visual data collection and transmission under bandwidth scarcity caused by damaged communication networks and 2) analyzing the acquired data in a timely manner. We introduce a new crowdsourcing framework for acquiring and analyzing the mobile videos utilizing fine granularity spatial metadata of videos for a rapidly changing disaster situation. We also develop an analytical model to quantify the visual awareness of a video based on its metadata and propose the visual awareness maximization problem for acquiring the most relevant data under bandwidth constraints. The collected videos are evenly distributed to off-site analysts to collectively minimize crowdsourcing efforts for analysis. Our simulation results demonstrate the effectiveness and feasibility of the proposed framework.
Links:
http://infolab.usc.edu/DocsDemos/to_ieeebigdata2015.pdf
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=7363814
C1.01: GOOS: an essential collaborative system enabling societal benefit - Jo...Blue Planet Symposium
A sustained ocean observing system forms a basis, along with capacity development, enabling societal benefit from ocean information. The Global Ocean Observing System (GOOS) is driven by societal requirements, including:
- adapting to and mitigating climate variability and change
- preparing for ocean-related hazards and supporting development of the ocean economy, and
- monitoring and promoting stewardship of ocean health.
GOOS is a key contributor to the GEO Blue Planet task. We are a collaborative programme, connecting a community and organizations working on many aspects of a Framework for Ocean Observations: identifying requirements focused on Essential Ocean Variables, coordinating observing networks and monitoring progress towards targets, and connecting to data and information generation activities that create scientific and societal value.
At the global level, these processes are led by three GOOS Panels focused on physics, biogeochemistry, and biology. The panels evaluate the readiness of the observing system, promoting strategic investment by identifying what is essential, and encouraging the development of new capabilities. They work closely with the ocean observing community. A Strategic Mapping is helping to identify how elements integrate into the system. GOOS development projects are evaluating and where necessary will improve and change parts of the sustained ocean observing system.
The combined satellite and in situ observing networks contributing to GOOS have strengthened in recent years.
At the regional level, GOOS Regional Alliances individually focus on local priorities and requirements. Collectively, they work to develop institutional and human capacity to make and benefit from sustained ocean observations.
GOOS both supports and relies on many partners, including other contributors to the GEO Blue Planet, in seeking to sustain present observations, while integrating new essential ocean observations into a sustained observing system.
An Automated Approach of Shoreline Detection Applied to Digital Videos using ...Dwi Putra Asana
Abstract: This study aims to detect a shoreline location and its changes automatically in the temporal resolution.
This approach is implemented on the coastal video monitoring system applications. The proposed method applied
data mining by using two main systems-a training system using classification and shoreline detection systems with
Self-Organizing Map (SOM) and K-Nearest Neighbor (K-NN) algorithms. The training system performs feature
texture extraction using agray-level co-occurrence matrix and the results are stored to classification process. The
detection system has five processing stages: contrast stretching preprocessing and morphological contrast
enhancement, SOM clustering, morphological operations, feature extraction and K-NN classification and detection
shoreline. Preprocessing was used to improve the video image contrast and reliability. SOM algorithm in
segmenting objects in the onshore video images. Morphological operations were applied to eliminate noise on the
objects that were not needed in the spatial domain. The segmentation results of video frames classified by K-NN.
The aim is to provide the class labels on each region segmentation results, namely, sea label, land label and sky
label. The determination of the shoreline is done by scanning the neighboring pixels from the edge of land class
label after binary image transformation. The shoreline change detection was performed by comparing the position of
existing shoreline and shoreline position in the reference video frame. A Receiver Operating Characteristic (ROC)
curve was used to evaluate the performance of shoreline detection systems. The results showed that the combination
of SOM and K-NN was able to detect shoreline and its changes accurately
GPS and its use in vehicle movement study in Earthquake Disaster ManagementMayur Rahangdale
What is GPS?
GPS Segments
Pseudo – Random Numbers (PRN)
Coarse acquisition (C/A) code
P code (Precision or Protected code)
P code (Precision or Protected code)
GPS Trilateration
EARTHQUAKE DISASTER MANAGEMENT
Disaster Management Cycle
ADVANTAGE OF GPS IN DISASTER MANAGEMENT
GPS LIMITATION IN DISASTER MANAGEMENT
HOW DOES GPS PLAY A ROLE IN EARTHQUAKE RESCUE?
Case Study - Great East Japan Earthquake in Ishinomaki City, Japan -11 March 2011.
C1.05: Sustained observations for many users - a perspective from Australia’s...Blue Planet Symposium
Australia is a ‘marine nation’ – an island continent with the third largest ocean territory on the ‘Blue Planet’. Our borders are maritime and we generate massive wealth from marine industries. Most of our population lives in highly urbanised centres on or near the coast, and we are extremely sensitive to ocean-influenced climate and weather, through drought, flood, and tropical cyclones. Our ocean territory contains marine biodiversity of globally significant conservation and tourism value, ranging from the high tropics to Antarctica. These factors combine to establish the need for sustained ocean observing in the Australian context, for many uses and users.
Despite this clear, national need, responsibility for ocean observing and management is fragmented and dispersed. A National Oceans Policy and independent National Oceans Office were established in 1998, but were subsumed into the Federal Environment portfolio by 2005. The Bureau of Meteorology is Australia's national weather agency, and while its role has expanded to encompass climate and water services over the last decade, it is only now beginning to consider an expanded role in marine services. Jurisdiction of the marine environment, including responsibility for marine monitoring, is shared across Federal, State and Territory Governments, across different Departments within those various Governments, and between industrial users and regulators in areas like offshore oil and gas and commercial fishing. It is also significant to note that Australia has no earth observation from space (EOS) capability of its own.
Since 2006, Australia has put in place a national Integrated Marine Observing System (IMOS). Established as a research infrastructure, IMOS routinely operates a wide range of observing equipment, making all of its data openly accessible to the marine and climate science community, other stakeholders and users, and international collaborators. It is integrated from open-ocean to coast, and across physical, chemical and biological ocean variables.
This talk will focus on what has been learnt through the experience of building IMOS as a research infrastructure in a context where sustained ocean observations are needed by many users.
Use of satellite imagery for the generation of an aquaculture atlas : a case ...Blue BRIDGE
Presented by Nicolas Longépé of CLS at the BlueBRIDGE Workshop at SeaTech Week in Brest, France, October 13th, 2016.
http://www.bluebridge-vres.eu/events/join-bluebridge-10th-biennial-sea-tech-week-brest-france
Time, Change and Habits in Geospatial-Temporal Information StandardsGeorge Percivall
Keynote for HIC 2014 – 11th International Conference on Hydroinformatics, New York, USA August 17 – 21, 2014
Time, Change and Habits in Geospatial-Temporal Information Standards
Time and change are fundamental to our scientific understanding of the world. Standards for geospatial-temporal information exist but new needs outstrip current standards. Geospatial-temporal information includes capturing change in features and coverages and modeling the processes that inform change. Key standards for time, calendars, and temporal reference systems are in place. Time series modeling from the WaterML standard is a recent advance of high value to hydrology. The OGC Moving Features standard will establish an encoding format for changes in “rigid” features. Interoperability standards are needed for Coverages with values that change based on observations, analytical expressions, or simulations. Applying a coverage model to time-varying, fluid Earth systems was the topic of the ground breaking GALEON Interoperability Experiment. Standards developments for spatial-temporal process models is progressing with WPS, OpenMI and ESMF - supporting a Model Web concept. A robust framework for sharing geospatial-temporal information is now coming into place based on developments captured in standards by ISO, WMO, ITU, ICSU and OGC - including the newly established OGC Temporal domain working group. The new framework will enable capabilities in expressing and sharing scientific investigations including research on the emergence of forms over time. With these new capabilities we may come to understand Peirce’s observation that over time “all things have a tendency to take habits.”
Similar to Mapping the Tohoku 2011 Tsunami event with a remote sensing satellite constellation – a reference case for emerging Early Warning System Dissemination Services (20)
INTEGRATION OPTIONS FOR PERSISTENT IDENTIFIERS IN OSGEO PROJECT REPOSITORIES:...Peter Löwe
As a contribution to the currently ongoing larger effort to establish Open Science as best practices in academia, this article focuses on the Open Source and Open Access tiers of the Open Science triad and community software projects. The current situation of research software development and the need to recognize it as a significant contribution to science is introduced in relation to Open Science. The adoption of the Open Science paradigms occurs at different speeds and on different levels within the various fields of science and crosscutting software communities. This is paralleled by the emerging of an underlying futuresafe technical infrastructure based on open standards to enable proper recognition for published articles, data, and software. Currently the number of journal publications about research software remains low in comparison to the amount of research code published on various software repositories in the WWW. Because common standards for the citation of software projects (containers) and versions of software are lacking, the FORCE11 group and the CodeMeta project recommending to establish Persistent Identifiers (PIDs), together with suitable metadata setss to reliably cite research software. This approach is compared to the best practices implemented by the OSGeo Foundation for geospatial community software projects. For GRASS GIS, a OSGeo project and one of the oldest geospatial open source community projects, the external requirements for DOI-based software citation are compared with the projects software documentation standards. Based on this status assessment, application scenarios are derived, how OSGeo projects can approach DOI-based software citation, both as a standalone option and also as a means to foster open access journal publications as part of reproducible Open Science.
FOSSGIS 2015: Das audiovisuelle Erbe der OSGeo-ProjektePeter Löwe
Die Menge der audiovisuellen Inhalte hat in den letzten Jahren stark zugenommen und wächst weiterhin stark. Die thematische Palette erstreckt sich dabei von Anleitungsvideos für Softwarekomponenten, über Ergebnis- und Datenvisualisierungen bis hin zur abstrahierten Darstellung der Evolution einzelner Softwareprojekte. Die Fülle dieser Fachinformationen wird noch hauptsächlich in Web2.0 Portalen wie Youtube oder Slideshare ausgetauscht. Daraus ergeben sich Probleme sowohl bezüglich der langfristigen Verfügbarkeit und der Auffindbarkeit anhand geeigneter Metadaten für die Nutzer. Eine zukunftssichere Alternative bietet die Nutzung von innovativen Bibliotheksdiensten, wie dem AV-Portal der Technischen Informationsbibliothek (TIB). Hier werden nicht nur die Metadaten eines Videos indexiert sondern ebenso die gesprochene Sprache, Texteinblendungen und Bildinformationen. Dies führt zu einer erheblich verbesserten Suche nach und in audiovisuellen Ressourcen. Durch die Verbindung eines DOI mit einem Media Fragment Identifier wird die sekundengenaue Zitierfähigkeit der Materialien gewährleistet.
Der Nutzen des Portals wird am Beispiel der erfolgreichen digitalen Erschließung des GRASS GIS Videos des U.S. Army Corps of Engineers Research Laboratory (CERL) aus dem Jahr 1987 demonstriert. Der Inhalt dieses historischen Videos bietet einen Einblick in die Frühphase der GIS Entwicklung. Die Erschließungsgeschichte des Videos seit 2004 ist ein Referenzfall für den aktuellen Stand und das sich abzeichnende Potential audiovisueller Information für die Geoinformatik und speziell den wissenschaftlichen Anwendern und Entwicklern in den OSGeo Projekten.
Unlocking conference videos by DOI/MFID for software project communitiesPeter Löwe
Das TIB AV-Portal der Technischen Informationsbibliothek in Hannover ist eine kunden- und bedarfsorientierte Plattform für hochwertige wissenschaftliche Videos aus dem Bereich Technik und Naturwissenschaften. Automatische Szenenerkennung sowie Text-, Audio- und Bildanalyse werden zur automatischen Anreicherung mit Metadaten genutzt. Die Filme sind mit einem Digital Object Identifier (DOI) und - auf Segmentbasis - mit Media Fragment Identifiern versehen, womit die Filme sekundengenau dereferenziert und zitiert werden können. Praxisnutzen und Potentiale dieser Infrastruktur für die Erschließung und Nachnutzung in Wissenschaft, Lehre und Industrie werden exemplarisch am Beispiel der FOSS4G-Konferenzvideos der Open Source Geospatial Foundation, vorgestellt.
The TIB|AV Portal : OSGeo conference videos as a resource for scientific res...Peter Löwe
This paper reports on new opportunities for research and education in Free and Open Source Geoinformatics as a translational part of Open Science, enabled the growing collection of OSGeo conference video recordings at the German National Library of Science and Technology (TIB). Since 2015, OSGeo conference recordings have been included to the collection sphere of TIB in information sciences. Currently, video content from selected national (FOSSGIS), regional (FOSS4G-NA) and global (FOSS4G) conferences is being actively collected. The annual growth exceeds 100 hours of new content relating to the OSGeo software projects and the OSGeo scientific-technical communities. This is seconded by retrospective acquisition of video material dating from past conferences, going back until 2002 to preserve this content, ensuring both long term availability and access. The audiovisual OSGeo-related content is provided through the TIB|AV Portal, a web-based platform for scientific audiovisual media providing state-of-the art multimedia analysis and retrieval. It implements the requirements by research libraries for reliable long term preservation. Metadata enhancement analysis provides extended search and retrieval options. Digital Object Identifiers (DOI) enable scientific citation of full videos, excerpts and still frames, use in education and also referral in social networks. This library-operated service infrastructure turns the audiovisual OSGeo-related content in a reliable source for science and education.
GIS Day 2015: Geoinformatics, Open Source and Videos - a library perspectivePeter Löwe
Digital audiovisual content has become an important communication channel in Science. The TIB|AV-Portal for audiovisual scientific-technical information meets the requirements to preserve such content and to provide innovative services for search and retrieval. Quality checked audiovisual content from Open Source Geoinformatics communities is constantly being acquired for the portal as a part of TIB's mission to preserve relevant content in applied computer sciences for science, industry, and the general public.
GRASS GIS, Star Trek and old Video Tape – a reference case on audiovisual pre...Peter Löwe
This presentation showcases new options for the preservation of audiovisual content in the OSGeo communities beyond the established software repositories or Youtube. Audiovisual content related to OSGeo projects such as training videos and screencasts can be preserved by advanced multimedia archiving and retrival services which are currently developed by the library community. This is demonstrated by the reference case of a newly discovered high resolution version of the GRASS GIS 1987 promotional video which made available from into the AV-portal of the German National Library of Science and Technology (TIB). The portal allows for extended search capabilities based on enhanced metadata derived by automated video analysis. This is a reference case for future preservation activities regarding semantic-enhanced Web2.0 content from OSGeo projects.
3D-printing with GRASS GIS – a work in progress in report FOSS4G 2014Peter Löwe
As the amount of scientific data continues to grow, researchers need new tools to help them visualize complex data. Immersive data-visualisations are helpful, yet fail to provide tactile feedback and sensory feedback on spatial orientation, as provided from tangible objects.
The production of a tangible representation of a scientific data set is one step in a line of scientific thinking, leading from the physical world into scientific reasoning and back: The process starts with a physical observation, or from a data stream generated by an environmental sensor. This data stream is turned into a geo-referenced data set. This data is turned into a volume representation which is converted into command sequences for the printing device, leading to the creation of a 3D printout via additive manufacturing (“3D-printing”). As a last, but crucial step, this new object has to be documented and linked to the associated metadata, and curated in long term repositories to preserve its scientific meaning and context.
This presentation showcases a reference workflow to produce tangible 3D data-prints based on Free and Open Source Software (FOSS), using both GRASS GIS and Paraview. The workflow was successfully validated in various application scenarios using a RapMan printer to create 3D specimens of elevation models, geological underground models, ice penetrating radar soundings for planetology, and space time stacks for Tsunami model quality assessment.
Tectonic Storytelling with Open Source and Digital Object Identifiers - a cas...Peter Löwe
The communication of advances in research to the common public for both education and decision making is an important aspect of scientific work. An even more crucial task is to gain recognition within the scientific community,
which is judged by impact factor and citation counts. Recently, the latter concepts have been extended from
textual publications to include data and software publications.
This paper presents a case study for science communication and data citation. For this, tectonic models, Free and Open Source Software (FOSS), best practices for data citation and a multimedia online-portal for scientific content
are combined. This approach creates mutual benefits for the stakeholders: Target audiences receive information on
the latest research results, while the use of Digital Object Identifiers (DOI) increases the recognition and citation of
underlying scientific data. This creates favourable conditions for every researcher as DOI names ensure citeability and long term availability of scientific research.
In the developed application, the FOSS tool for tectonic modelling GPlates is used to visualise and manipulate
plate-tectonic reconstructions and associated data through geological time. These capabilities are augmented by the Science on a Halfsphere project (SoaH) with a robust and intuitive visualisation hardware environment.
The tectonic models used for science communication are provided by the AGH University of Science and Technology.
They focus on the Silurian to Early Carboniferous evolution of Central Europe (Bohemian Massif) and were
interpreted for the area of the Geopark Bergstraße Odenwald based on the GPlates/SoaH hardware- and software stack.
As scientific story-telling is volatile by nature, recordings are a natural means of preservation for further use, reference and analysis. For this, the upcoming portal for audiovisual media of the German National Library of Science and Technology TIB is expected to become a critical service infrastructure. It allows complex search queries, including metadata such as DOI and media fragment identifiers (MFI), thereby linking data citation and science
communication.
Data Science: History repeated? – The heritage of the Free and Open Source GI...Peter Löwe
Data Science is described as the process of knowledge extraction from large data sets by means of scientific
methods. The discipline draws heavily from techniques and theories from many fields, which are jointly used to
furthermore develop information retrieval on structured or unstructured very large datasets. While the term Data
Science was already coined in 1960, the current perception of this field places is still in the first section of the hype cycle according to Gartner, being well en route from the technology trigger stage to the peak of inflated
expectations.
In our view the future development of Data Science could benefit from the analysis of experiences from
related evolutionary processes. One predecessor is the area of Geographic Information Systems (GIS). The
intrinsic scope of GIS is the integration and storage of spatial information from often heterogeneous sources, data
analysis, sharing of reconstructed or aggregated results in visual form or via data transfer. GIS is successfully
applied to process and analyse spatially referenced content in a wide and still expanding range of science
areas, spanning from human and social sciences like archeology, politics and architecture to environmental and
geoscientific applications, even including planetology.
This paper presents proven patterns for innovation and organisation derived from the evolution of GIS,
which can be ported to Data Science. Within the GIS landscape, three strategic interacting tiers can be denoted: i) Standardisation, ii) applications based on closed-source software, without the option of access to and analysis of the implemented algorithms, and iii) Free and Open Source Software (FOSS) based on freely accessible program code enabling analysis, education and ,improvement by everyone. This paper focuses on patterns gained from the synthesis of three decades of FOSS development. We identified best-practices which evolved from long term FOSS projects, describe the role of community-driven global umbrella organisations such as OSGeo, as well as the standardization of innovative services. The main driver is the acknowledgement of a meritocratic attitude.
These patterns follow evolutionary processes of establishing and maintaining a web-based democratic culture
spawning new kinds of communication and projects. This culture transcends the established compartmentation and
stratification of science by creating mutual benefits for the participants, irrespective of their respective research
interest and standing. Adopting these best practices will enable
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
Kubernetes & AI - Beauty and the Beast !?! @KCD Istanbul 2024Tobias Schneck
As AI technology is pushing into IT I was wondering myself, as an “infrastructure container kubernetes guy”, how get this fancy AI technology get managed from an infrastructure operational view? Is it possible to apply our lovely cloud native principals as well? What benefit’s both technologies could bring to each other?
Let me take this questions and provide you a short journey through existing deployment models and use cases for AI software. On practical examples, we discuss what cloud/on-premise strategy we may need for applying it to our own infrastructure to get it to work from an enterprise perspective. I want to give an overview about infrastructure requirements and technologies, what could be beneficial or limiting your AI use cases in an enterprise environment. An interactive Demo will give you some insides, what approaches I got already working for real.
Neuro-symbolic is not enough, we need neuro-*semantic*Frank van Harmelen
Neuro-symbolic (NeSy) AI is on the rise. However, simply machine learning on just any symbolic structure is not sufficient to really harvest the gains of NeSy. These will only be gained when the symbolic structures have an actual semantics. I give an operational definition of semantics as “predictable inference”.
All of this illustrated with link prediction over knowledge graphs, but the argument is general.
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mapping the Tohoku 2011 Tsunami event with a remote sensing satellite constellation – a reference case for emerging Early Warning System Dissemination Services
1. ISOPE-2013 Anchorage Conference
The 23rd International Ocean and Polar Engineering Conference
Anchorage, Alaska, USA, June 30−July 5, 2013: www.isope.org; www.isope2013.org
Mapping the Tohoku 2011 Tsunami event
with a remote sensing satellite constellation
– a reference case for emerging Early Warning System
Dissemination Services
Peter Löwe, Joachim Wächter
Centre for GeoinformationTechnology (CeGIT), GFZ
German Research Centre for Geosciences
Potsdam, Germany
2. ISOPE-2013
Anchorage
ISOPE-2013
Anchorage
TRIDEC
• New technologies for real time‐
intelligent information management
in collaborative, complex critical
decision processes
• In TRIDEC new developments in
Information and Communication
Technology (ICT) are used to
extend existing platforms with a
component-based technology
framework.
• Demonstration in two scenarios:
Tsunami Early Warning Systems
(Natural Crisis Management)
and Drilling Operations.
2
3. ISOPE-2013
Anchorage
The potential of improved satellite crisis mapping
•During a Tsunami early warning event, TRIDEC Natural Crisis
Management (NCM) systems provide crucial information on when
and where coastlines will be affected.
•This critical information can be provided to the operators of
satellite remote sensing systems for follow up actions.
TRIDEC NCM GUI
4. ISOPE-2013
Anchorage
Remote Sensing for Disaster Mitigation
• 1999: International Charter "Space and Major
Disasters“ founded by ESA and CNES.
• Approach: „An authorized user can request the
mobilization of the space and associated ground
resources of the member agencies to obtain data and
information on a disaster occurrence.“
• An Emergency On-Call Officer prepares an archive
(„before“) and acquisition („after“) plan.
• Charter-Members handle data acquisition and delivery
on an emergency basis.
5. ISOPE-2013
Anchorage
Tohoku 2011 Earthquake and Tsunami
• In the wake of the Tohoku Tsunami
Disaster of March 11 2011, the
International Charter for Space
and Major Disasters was activated
to coordinate both the imaging
campaigns and the creation of crisis
maps.
• The affected areas were imaged by
satellite-based remote sensing
sensors.
• Crisis map products were used by
Search and Rescue to save lifes.
9. ISOPE-2013
Anchorage
Benefits from a EO dissemination channel
• Preparations for satellite imaging can begin before the tsunami
devastates an Area of Interest,
• reducing the time between tsunami landfall and first satellite
image take,
• speeding up production of crisis maps,
• enabling earlier and better coordinated response by
Search and Rescue (SAR),
• potenially allowing for before/after coverage.
12. ISOPE-2013
Anchorage
Evolution of Earth Observation Systems
Core parameters for satellite-based remote sensing:
• Spatial/spectral resolution („Pixel size“ of sensor)
• Temporal resolution (Revisit rate: „Number of satellites“)
•Single satellite
• Pace for image acquisition planning: Weeks / Days
•Constellations of several satellites in one orbit plane
• Pace: Day(s) / Hours
•Multiple Constellations in multiple orbit planes (upcoming)
• Pace: Hours / Minutes
13. ISOPE-2013
Anchorage
Workflow Integration: Status
•Currently, Tsunami Early Warning only considers bullets 2 and 3.
•Tsunami Early Warning must provide Tsunami Information (bullet 1) to
start the image acquisition process (orange)
Satellite
Image
Data
Save
lives
MappingTaskingOrderingTsunami
Warning
Imaging
Satellite
Image
Map
1 2 3
14. ISOPE-2013
Anchorage
Earth Observation: Planning and Tasking
• Operators of remote sensing satellites operate a planning and
tasking/commandeering cycle to control image acquisition.
• Planning: Arrange and prioritize image orders by
• Area of Interest (AOI)
• Time of Interest (TOI)
• Optical Sensors: Cloud Forecast Constraints
• Other Factors: Urgency, on-board memory availability, power,
etc.
• Tasking: Setting of an imaging schedule.
• Commandeering: Upload of imaging schedule to sensors + execution.
15. ISOPE-2013
Anchorage
Case Study: Rapid Eye Constellation
• Commercially operated constellation of five
satellites
• Charter Member
• Disaster Mapping since 2009
Summer 2013; Flooding in Central
Europe
16. ISOPE-2013
Anchorage
Example: The RapidEye Satellite Constellation
• Constellation operational since February 2009
• Five identical optical remote sensing satellites
• ~ 630km above ground
• ~ 90 minutes orbit period
• Sun-synchronous (overpass 11:30 a.m. local time)
Images: RapidEye
17. ISOPE-2013
Anchorage
RapidEye: Optical Imaging Sensors
•Five spectral bands: red / green / blue / red-edge / near-infrared
•Resolution of data products: 5m*5m pixel.
•Swath width: 77 km
•Max. swath length: 1200 km
Egypt 2010 Image: RapidEye AG
Example: Anchorage
18. ISOPE-2013
Anchorage
Example: Order Planning/Tasking/Execution
The timeline reflects the RapidEye “Two plannings per day” scenario.
10h – 25h
Daily Deadlines
Order turnaround time
Source: Hoja et al.: Optimised Near-Real Time Data Acquisition and
Pre- processing of Satellite Data for Disaster Related Rapid Mapping :
PFG2010/6,429-438
23. ISOPE-2013
Anchorage
Use of Volunteered Geographic Information by
the EO Community
• Gulf of Mexico Oilspill 2010
• Louisiana Bucket Brigade
• Ushahidi Open Source Social
Mapping
Before
Ushahidi/Bucket Brigade: Oil
residue found on beach
Confirmation by EO
26. ISOPE-2013
Anchorage
XML-based Message Format Candidates:
CAP and EDXL
• Common Alert Protocoll
(CAP): Data format for
exchanging public warnings
and emergencies between
alerting technologies.
• Flexible geographic targeting
using latitude/longitude
“boxes”;
• Facility for digital images,
audio, and video.
• Emergency Data Exchange
Language (EDXL) - a suite of
messaging standards,
advanced by the OASIS
Emergency Management
Technical Committee,
EDXL spatial tags
27. ISOPE-2013
Anchorage
Outline of an automated heads-up process
• Derivation of affected coastline segments from Tsunami-
simulations and run-up modelling.
• Prioritization by estimated time of arrival and severity
estimates.
• Mapping of affected areas to EO tile grids.
• Tasking of EO constellations.
28. ISOPE-2013
Anchorage
Conclusion: EO Operators Perspective
Benefits from dedicated EO warning services:
•Preparation for short term crisis mapping ahead of time
•Image acquisition as early as possible
•Larger data stock
•Increased business opportunities
29. ISOPE-2013
Anchorage
Conclusion: Tsunami Early Warning Systems
(TEWS)
•The earlier crisis maps become available after a Tsunami,
the more lives can be saved.
•Satellite-based crisis mapping for large areas is a valuable tool
for disaster and crisis management.
•TEWS dissemination components can provide customized
message formats for EO communities including AOI/ETA .
•A suitable dissemination channel for EO will speed up
Crisis Mapping and will help to save lives.
Satellite
Image
Data
Save
lives
MappingTaskingOrderingTsunami
Warning
Imaging
Satellite
Image
Map
[Wikipedia CAP] The Common Alerting Protocol (CAP) is an XML -based data format for exchanging public warnings and emergencies between alerting technologies . CAP allows a warning message to be consistently disseminated simultaneously over many warning systems to many applications. CAP increases warning effectiveness and simplifies the task of activating a warning for responsible officials. Individuals can receive standardized alerts from many sources and configure their applications to process and respond to the alerts as desired. Alerts from the Department of Homeland Security , the Department of the Interior's United States Geological Survey, and the Department of Commerce's National Oceanic and Atmospheric Administration (NOAA), and state and local government agencies can all be received in the same format, by the same application. That application can, for example, sound different alarms based on the information received. By normalizing alert data across threats, jurisdictions, and warning systems, CAP also can be used to detect trends and patterns in warning activity, such as trends that might indicate an undetected hazard or hostile act. From a procedural perspective, CAP reinforces a research-based template for effective warning message content and structure. The CAP data structure is backward-compatible with existing alert formats including the Specific Area Message Encoding (SAME) used in Weatheradio and the broadcast Emergency Alert System as well as new technology such as the Commercial Mobile Alert System (CMAS), while adding capabilities including: Flexible geographic targeting using latitude/longitude “boxes” and other geospatial representations in three dimensions; Multilingual and multi-audience messaging; Phased and delayed effective times and expirations; Enhanced message update and cancellation features; Template support for framing complete and effective warning messages; Digital encryption and signature capability; Facility for digital images, audio, and video. [Wikipedia EDXL] EDXL is advanced by the OASIS Emergency Management Technical Committee, [2] a group that was formed in 2003 and remains open to participation from organizations, agencies, and individuals from around the world. EDXL is based on detailed requirements and draft specifications provided to OASIS by emergency practitioners, with support from the Emergency Interoperability Consortium, [3] through a project sponsored by the United States Department of Homeland Security’s Disaster Management E-Gov Initiative. EDXL-DE was approved as an OASIS Standard in 2006; EDXL-RM and –HAVE were approved as OASIS Standards in 2008. Implementation of EDXL is promoted by the OASIS Emergency Management Adoption Committee, which was formed in 2009. Background The Disaster Management eGov Initiative of the Department of Homeland Security (DHS) determined in 2004 to launch a project to develop interagency emergency data communications standards. It called together a group of national emergency response practitioner leaders and sought their guidance on requirements for such standards. In June, 2004 the first such meeting identified the need for a common distribution element for all emergency messages. Subsequent meetings of a Standards Working Group developed detailed requirements and a draft specification for such a distribution element (DE). During the same period the DM Initiative was forming a partnership with industry members of the Emergency Interoperability Consortium (EIC) to cooperate in the development of emergency standards. EIC had been a leading sponsor of the Common Alerting Protocol (CAP). Both organizations desired to develop an expanded family of data formats for exchanging operational information beyond warning. EIC members participated in the development of the DE, and in the broader design of the design of a process for the development of additional standards. This was named Emergency Data Exchange Language (EDXL). The goal of the EDXL project is to facilitate emergency information sharing and data exchange across the local, state, tribal, national and non-governmental organizations of different professions that provide emergency response and management services. EDXL will accomplish this goal by focusing on the standardization of specific messages (messaging interfaces) to facilitate emergency communication and coordination particularly when more than one profession is involved. It is not just an "emergency management" domain exercise. It is a national effort including a diverse and representative group of local, state and federal emergency response organizations and professionals, following a multi-step process. Just as a data-focused effort targets shared data elements, the EDXL process looks for shared message needs, which are common across a broad number of organizations. The objective is to rapidly deliver implementable standard messages, in an incremental fashion, directly to emergency response agencies in the trenches, providing seamless communication and coordination supporting each particular process. The effort first addresses the most urgent needs and proceeds to subsequent message sets in a prioritized fashion. The goal is to incrementally develop and deliver standards. EDXL is intended as a suite of emergency data message types including resource queries and requests, situation status, message routing instructions and the like, needed in the context of cross-disciplinary, cross-jurisdictional communications related to emergency response. The priorities and requirements are created by the DM EDXL Standards Working Group (SWG) which is a formalized group of emergency response practitioners, technical experts, and industry. The draft DE specification was trialed by a number of EIC members starting in October, 2004. In November, 2004, EIC formally submitted the draft to the OASIS Emergency Management Technical Committee for standardization. EDXL-RM (Resource Messaging) Overview EDXL-RM (Resource Message) [5] OASIS Standard, which describes a suite of standard messages for sharing data among information systems that coordinate requests for emergency equipment, supplies, and people. Purpose The primary purpose of the Emergency Data Exchange Language Resource Messaging (EDXL-RM) Specification is to provide a set of standard formats for XML emergency response messages. These Resource Messages are specifically designed as payloads of Emergency Data Exchange Language Distribution Element- (EDXL-DE)-routed messages. Together EDXL-DE and EDXL-RM are intended to expedite all activities associated with resources needed to respond and adapt to emergency incidents. The Distribution Element may be thought of as a "container". It provides the information to route "payload" message sets (such as Alerts or Resource Messages), by including key routing information such as distribution type, geography, incident, and sender/recipient IDs. The Resource Message is constrained to the set of Resource Message Types contained in this specification. The Resource Message is intended to be the payload or one of the payloads of the Distribution Element which contains it. History Disaster Management (DM) is a communications program in the Department of Homeland Security’s (DHS) Office for Interoperability and Compatibility (OIC) and managed by the Science and Technology (S&T) Directorate. The program was initiated as one of the President’s e-government initiatives. DM’s mission is to serve as the program within the Federal Government to help local, tribal, state, and federal public safety and emergency response agencies improve public safety response through more effective and efficient interoperable data sharing. The DHS DM program sponsors a Practitioner Steering Group (PSG). The DM Practitioner Steering Group (PSG) governance was formalized following publication of the EDXL Distribution Element. It plays a key role in the direction, prioritization, definition, and execution of the DHS-DM program. The group is composed of representatives of major emergency response associations, setting priorities and providing recommendations regarding messaging standards development as well as the other facets of the DM program. The PSG specified messaging standards-based systems interoperability as the top priority for the DHS Disaster Management program. The EDXL Resource Messaging Specification effort was identified as the top priority standard by this group following the EDXL-DE. The requirements and specification effort was initiated by this group in partnership with industry members of the Emergency Interoperability Consortium (EIC) in a Standards Working Group (SWG). That group developed a draft specification which was submitted to the OASIS Emergency Management Technical Committee to begin work on this EDXL-RM specification. The process remained the same as with the EDXL-DE specification with the exception that the Technical Committee requested that the initial candidate specification submitted by the expert group be recast as a formal Requirements Document according to a template that the Technical Committee provided to the expert group. The candidate specification was then resubmitted along with this requested requirements document.