Third meeting of the International Network of Soil Information Institutions (INSII), 31 October - 1 November 2017, FAO headquarters, Rome. Composed by institutions working on soil information in FAO member countries INSII provides expertise, and facilitates soil data/information sharing. Presentation by Rik van den Bosch, Soil Data Facility Chair, International Soil Reference and Information Centre (ISRIC).
Pillar 4 Implementation Plan (P4IP) and current soil information developmentsExternalEvents
Third meeting of the International Network of Soil Information Institutions (INSII), 31 October - 1 November 2017, FAO headquarters, Rome. Composed by institutions working on soil information in FAO member countries INSII provides expertise, and facilitates soil data/information sharing. Presentation by Neil McKenzie, Chair of the Pillar 4 of the Global Soil Partnership.
Application and integration of Python in the geodata management of a local go...shj69
The document discusses the use of Python in managing geospatial data for a local government office in Basel-Stadt, Switzerland. The office manages geospatial data infrastructure and applications, develops GIS services, and maintains key geospatial datasets. Python is used to synchronize attribute data between applications and geospatial databases, and to migrate interfaces like the Bluebox interface for exchanging data between geospatial and cadastral systems from Java to Python. Python tools like SQLAlchemy are used for communicating with databases.
The document describes a new GIS fiber optics software developed by Worldcall GIS & Software team. The software creates linked databases for fiber optic network infrastructure management, information gathering, and mapping capabilities. It integrates fiber data with GIS software to provide geographic views of the network. This allows more efficient network operation and management compared to the original software, saving the company millions of costs. The remainder of the document shows screenshots of the key database summaries and linkages between the GIS and database components.
The document outlines the goals and timeline of a project to develop a system to help digitize and interpret archaeological findings. The project aims to 1) develop tools to document archaeological findings using mobile devices, 2) digitize existing catalogues of artifacts, 3) provide on-site support for interpreting findings, 4) archive all collected data through web-accessible repositories, and 5) ensure long-term access and reuse of open data. It will involve developing technologies for digitizing paper records, populating a centralized database, enabling similarity searches of artifacts, and testing mobile and web-based tools with users over its 2018-2019 timeline.
This document discusses RIPEstat and RIPE Atlas, which are open data and measurement platforms operated by RIPE NCC. It provides an overview of the data sources and APIs available on RIPEstat, as well as the global network of probes and types of measurements that can be performed using RIPE Atlas. Examples of how RIPEstat and RIPE Atlas are used are also presented.
Third meeting of the International Network of Soil Information Institutions (INSII), 31 October - 1 November 2017, FAO headquarters, Rome. Composed by institutions working on soil information in FAO member countries INSII provides expertise, and facilitates soil data/information sharing. Presentation by Rik van den Bosch, Soil Data Facility Chair, International Soil Reference and Information Centre (ISRIC).
Pillar 4 Implementation Plan (P4IP) and current soil information developmentsExternalEvents
Third meeting of the International Network of Soil Information Institutions (INSII), 31 October - 1 November 2017, FAO headquarters, Rome. Composed by institutions working on soil information in FAO member countries INSII provides expertise, and facilitates soil data/information sharing. Presentation by Neil McKenzie, Chair of the Pillar 4 of the Global Soil Partnership.
Application and integration of Python in the geodata management of a local go...shj69
The document discusses the use of Python in managing geospatial data for a local government office in Basel-Stadt, Switzerland. The office manages geospatial data infrastructure and applications, develops GIS services, and maintains key geospatial datasets. Python is used to synchronize attribute data between applications and geospatial databases, and to migrate interfaces like the Bluebox interface for exchanging data between geospatial and cadastral systems from Java to Python. Python tools like SQLAlchemy are used for communicating with databases.
The document describes a new GIS fiber optics software developed by Worldcall GIS & Software team. The software creates linked databases for fiber optic network infrastructure management, information gathering, and mapping capabilities. It integrates fiber data with GIS software to provide geographic views of the network. This allows more efficient network operation and management compared to the original software, saving the company millions of costs. The remainder of the document shows screenshots of the key database summaries and linkages between the GIS and database components.
The document outlines the goals and timeline of a project to develop a system to help digitize and interpret archaeological findings. The project aims to 1) develop tools to document archaeological findings using mobile devices, 2) digitize existing catalogues of artifacts, 3) provide on-site support for interpreting findings, 4) archive all collected data through web-accessible repositories, and 5) ensure long-term access and reuse of open data. It will involve developing technologies for digitizing paper records, populating a centralized database, enabling similarity searches of artifacts, and testing mobile and web-based tools with users over its 2018-2019 timeline.
This document discusses RIPEstat and RIPE Atlas, which are open data and measurement platforms operated by RIPE NCC. It provides an overview of the data sources and APIs available on RIPEstat, as well as the global network of probes and types of measurements that can be performed using RIPE Atlas. Examples of how RIPEstat and RIPE Atlas are used are also presented.
The document summarizes a project conducted by Edinburgh Council to survey dropped kerbs using ESRI's Collector application. The survey gathered data on dropped kerbs, raised tables, guardrails, and residential crossovers across the city. While the Collector worked well for gathering the data, some issues arose around labeling, syncing large amounts of data, and downloading data with geometry errors. The collected data will now be added to the Roads Services asset management system and used to prioritize accessibility improvements through smaller streetworks projects and larger capital renewal schemes.
Item 2. ASP work from December 2016 to May 2018: JapanExternalEvents
The representative from Japan reported on the country's activities related to the Global Soil Partnership pillars from December 2016 to May 2018. For pillar 1 and 2, Japan utilized soil information to promote conservation agriculture and environmental subsidies, though no specific commitments were made. Under pillar 3, Japan expanded research networks, prepared a national soil organic carbon map, and conducted research on heavy metal contamination. Pillar 4 and 5 activities included developing a national soil inventory, preparing a soil organic carbon map, reviewing laboratory standards, correlating soil classifications, and upgrading a soil inventory mobile application.
The document describes a workshop on creating, expanding, and exploiting linked open statistical data. It discusses data cubes on the web and the RDF Data Cube Vocabulary for representing statistical data. The OpenCube toolkit is presented, including an OLAP browser for performing operations on integrated linked data cubes and a map view. The toolkit was deployed using Flemish government statistical data. The browser allows selecting dimensions and measures, drilling down, and expanding cubes by merging compatible sources. Challenges in combining disparate cubes were also discussed.
This document discusses how Georgia Power integrated their land records document management system (LIMS) with their GIS system. It describes their old separate LIMS and GIS systems, the challenges with those, and their process of developing a new integrated LIMSGIS system. The new system combines GIS searching and mapping capabilities with document searching and retrieval through their Onbase document management system. This allows users to search for land records both spatially using GIS maps as well as through keywords to access associated documents. The integration project took two years to complete and has provided benefits like improved user experience and faster response times to research requests.
GSP Pillar 4 Implementation Plan (GLOSIS) - Yusuf YiginiExternalEvents
This document discusses progress and plans for developing components of the Global Soil Information System (GLOSIS). It outlines work to create SoilSTAT for monitoring global soil resources, develop soil profile databases in two tiers, create a global polygon soil map, update the Harmonized World Soil Database, produce fine-resolution global soil grids, and continue the GSP Capacity Development Programme to build soil data management capacities. Key issues discussed include resource needs, technical specifications, coordination with other initiatives, and priority-setting among various products and activities.
This is a class presentation on MapServer...
The facts and figures explained here are somebody else's work. The author has just accumulated and presented them.
“Cadastral Maps for Socio-Economic Data Visualization and Integration for Lan...irjes
The impact of mining and mineral extraction activities can be significant on the surrounding land,
water and air bodies, in any operational area. The environmental degradation ranges from localized surface and
ground water contamination to the damaging effects of airborne pollutants on the regional ecosystem; which
need the properly designed geospatial database. The monitoring of these environmental impacts requires a userfriendly
and cost effective method to quantify the land cover changes over large time periods. Now-a-days, it
has become compulsory to use the remote sensing techniques for regular monitoring of these environmental
hazards in-and-around the mining areas using cadastral map. This paper provides a case study on the use of
geospatial techniques for environmental monitoring in the mining areas.
The document discusses Romania's National Spatial Data Infrastructure and the implementation of its geospatial data portal. It outlines the legal framework for the National Infrastructure for Spatial Information (INIS) and the roles of the INIS Council stakeholders. It describes challenges in coordinating the stakeholders, technical issues, and the status of implementing the Romanian GeoPortal with basemap and orthophoto data. Future phases are planned to further develop data models, access services, and online licensing.
European Spatial Data Infrastructure - INSPIRE and beyond inspireeu
This document provides an overview of the European Spatial Data Infrastructure INSPIRE and discusses related activities. It describes the key components and implementing rules of INSPIRE, outlines the implementation roadmap, and reviews the scope of spatial data themes covered under INSPIRE annexes. It also discusses the INSPIRE Maintenance and Implementation Framework, two Joint Research Centre led actions under the Interoperability Solutions for European Public Administrations program (ARE3NA and EULF), and indicates potential areas of interaction between INGeoCloudS and INSPIRE/ISA initiatives.
National Spatial Data Infrastructure (NSDI)Atiqa khan
The document lists important questions for the National Spatial Data Infrastructure (NSDI) exam on June 28th, 2014. It asks about the key components of an SDI, including policies, people, data, and technology. It defines metadata as data that describes other data and service resources for access and use, with an emphasis on accurate labeling. Finally, it states that the main objective of the NSDI is the development of spatial databases and transfer of data between government institutions and the public/private sectors based on user needs.
Spatial data infrastructure in the cloud, 2011Moullet
The document discusses spatial data infrastructures (SDIs) and the potential advantages and disadvantages of using cloud computing for SDIs. Some key points:
- An SDI is a framework of spatial data, metadata, users and tools that allow for efficient and flexible use of spatial data.
- Cloud computing provides on-demand access to configurable computing resources over a network. There are different service and deployment models.
- Potential advantages of cloud computing for SDIs include scalability, pay-as-you-go costs, and not needing dedicated servers. However, organizations need to keep control of important components and ensure security and privacy.
- Switzerland's Federal Office of Topography uses an infrastructure
Building Spatial Data Infrastructures for Spatial Planning in Africa: Lagos e...Samuel Dekolo
Lagos is the fastest growing Megacity in Sub-Saharan Africa, with its population estimated to double in the first quarter of this century; it is expected to be the third largest urban agglomerations in the world. This growth is not without challenges, as the city is grappling with myriads of urban management problems. City planners lack the most important ingredient of land use management, which is Information. In spite of huge investment on spatial data infrastructures at the national and state levels of government, most land use planners at both state and local government level agencies are ignorant of existing geospatial technology portals and unlock the full potentials of information and communication technologies. A statewide survey of the spatial data infrastructures of the city’s urban and land use management ministry and agencies proves its pathetic state, thereby creating information gap void between urban development and intelligent management. The result is has led to a sporadic growth of slums and unplanned settlements which now accounts for over 60% of the city. To avoid an impasse, it is necessary to review the level of geospatial technologies used at the local level and recommend formidable means of integration in the decision making process. This paper examines the level of geospatial technologies and Spatial Data Infrastructure use in spatial planning agencies and barriers to implementation in the 20 local governments of Lagos State and suggests the way forward.
András Sipos (Budapest City Archives; HU) presenting the e-proceedings of the ICARUS workshop “Cartography and Cadastral Maps: Visions from the past, for a vision of our future” - held in Pisa, October 2013 - on 28.04.2015 at the international conference "Archival Cooperation and Community Building in the Digital Age" within the panel “Building from scratch: written heritage online" at Břevnov Archabbey in Prague (CZ).
Spatial Data Infrastructure Best Practices with GeoNodeSebastian Benthall
The document discusses GeoNode, an open source software project that aims to build a next generation Spatial Data Infrastructure (SDI) based on best practices from modern web services. Some key points:
- GeoNode aims to make uploading, sharing, and working with spatial data as easy as blogging by reducing barriers to participation.
- It aligns incentives for contribution and use through features like user profiles, reputation systems, and ability to style and share maps.
- The software implements open standards and APIs to make data accessible and allow other applications to integrate.
Maps can be classified based on scale into small scale maps and large scale maps. Small scale maps show important information over larger areas like world maps and country maps. Large scale maps provide detailed information for small areas like village maps and cadastral maps which show land boundaries and ownership details. Topographical maps also fall under large scale maps and provide terrain and elevation details.
Topographic maps use contour lines to represent elevation and slope of land. Contour lines connect points of equal elevation and never cross. Closer lines indicate steeper slopes while widely spaced lines show more gentle slopes. Index contours are bold lines labeled with the elevation. A benchmark is a point of known exact elevation marked as B.M. on maps. Map scale relates distances on a map to actual distances on land and can be ratio, graphical, or verbal.
Introduction to aerial photography and photogrammetry.pptsrinivas2036
Aerial photography and photogrammetry are techniques used in remote sensing. Aerial photography involves taking photographs from aircraft and has been used since the 1850s. Photogrammetry uses photographs to measure and obtain spatial information about the objects and terrain photographed. It allows for the creation of topographic maps, cadastral maps, and large-scale construction plans more quickly and economically than traditional ground-based surveying. While aerial photography and photogrammetry provide advantages over field surveys, some on-site control and verification is still needed.
The document summarizes a project conducted by Edinburgh Council to survey dropped kerbs using ESRI's Collector application. The survey gathered data on dropped kerbs, raised tables, guardrails, and residential crossovers across the city. While the Collector worked well for gathering the data, some issues arose around labeling, syncing large amounts of data, and downloading data with geometry errors. The collected data will now be added to the Roads Services asset management system and used to prioritize accessibility improvements through smaller streetworks projects and larger capital renewal schemes.
Item 2. ASP work from December 2016 to May 2018: JapanExternalEvents
The representative from Japan reported on the country's activities related to the Global Soil Partnership pillars from December 2016 to May 2018. For pillar 1 and 2, Japan utilized soil information to promote conservation agriculture and environmental subsidies, though no specific commitments were made. Under pillar 3, Japan expanded research networks, prepared a national soil organic carbon map, and conducted research on heavy metal contamination. Pillar 4 and 5 activities included developing a national soil inventory, preparing a soil organic carbon map, reviewing laboratory standards, correlating soil classifications, and upgrading a soil inventory mobile application.
The document describes a workshop on creating, expanding, and exploiting linked open statistical data. It discusses data cubes on the web and the RDF Data Cube Vocabulary for representing statistical data. The OpenCube toolkit is presented, including an OLAP browser for performing operations on integrated linked data cubes and a map view. The toolkit was deployed using Flemish government statistical data. The browser allows selecting dimensions and measures, drilling down, and expanding cubes by merging compatible sources. Challenges in combining disparate cubes were also discussed.
This document discusses how Georgia Power integrated their land records document management system (LIMS) with their GIS system. It describes their old separate LIMS and GIS systems, the challenges with those, and their process of developing a new integrated LIMSGIS system. The new system combines GIS searching and mapping capabilities with document searching and retrieval through their Onbase document management system. This allows users to search for land records both spatially using GIS maps as well as through keywords to access associated documents. The integration project took two years to complete and has provided benefits like improved user experience and faster response times to research requests.
GSP Pillar 4 Implementation Plan (GLOSIS) - Yusuf YiginiExternalEvents
This document discusses progress and plans for developing components of the Global Soil Information System (GLOSIS). It outlines work to create SoilSTAT for monitoring global soil resources, develop soil profile databases in two tiers, create a global polygon soil map, update the Harmonized World Soil Database, produce fine-resolution global soil grids, and continue the GSP Capacity Development Programme to build soil data management capacities. Key issues discussed include resource needs, technical specifications, coordination with other initiatives, and priority-setting among various products and activities.
This is a class presentation on MapServer...
The facts and figures explained here are somebody else's work. The author has just accumulated and presented them.
“Cadastral Maps for Socio-Economic Data Visualization and Integration for Lan...irjes
The impact of mining and mineral extraction activities can be significant on the surrounding land,
water and air bodies, in any operational area. The environmental degradation ranges from localized surface and
ground water contamination to the damaging effects of airborne pollutants on the regional ecosystem; which
need the properly designed geospatial database. The monitoring of these environmental impacts requires a userfriendly
and cost effective method to quantify the land cover changes over large time periods. Now-a-days, it
has become compulsory to use the remote sensing techniques for regular monitoring of these environmental
hazards in-and-around the mining areas using cadastral map. This paper provides a case study on the use of
geospatial techniques for environmental monitoring in the mining areas.
The document discusses Romania's National Spatial Data Infrastructure and the implementation of its geospatial data portal. It outlines the legal framework for the National Infrastructure for Spatial Information (INIS) and the roles of the INIS Council stakeholders. It describes challenges in coordinating the stakeholders, technical issues, and the status of implementing the Romanian GeoPortal with basemap and orthophoto data. Future phases are planned to further develop data models, access services, and online licensing.
European Spatial Data Infrastructure - INSPIRE and beyond inspireeu
This document provides an overview of the European Spatial Data Infrastructure INSPIRE and discusses related activities. It describes the key components and implementing rules of INSPIRE, outlines the implementation roadmap, and reviews the scope of spatial data themes covered under INSPIRE annexes. It also discusses the INSPIRE Maintenance and Implementation Framework, two Joint Research Centre led actions under the Interoperability Solutions for European Public Administrations program (ARE3NA and EULF), and indicates potential areas of interaction between INGeoCloudS and INSPIRE/ISA initiatives.
National Spatial Data Infrastructure (NSDI)Atiqa khan
The document lists important questions for the National Spatial Data Infrastructure (NSDI) exam on June 28th, 2014. It asks about the key components of an SDI, including policies, people, data, and technology. It defines metadata as data that describes other data and service resources for access and use, with an emphasis on accurate labeling. Finally, it states that the main objective of the NSDI is the development of spatial databases and transfer of data between government institutions and the public/private sectors based on user needs.
Spatial data infrastructure in the cloud, 2011Moullet
The document discusses spatial data infrastructures (SDIs) and the potential advantages and disadvantages of using cloud computing for SDIs. Some key points:
- An SDI is a framework of spatial data, metadata, users and tools that allow for efficient and flexible use of spatial data.
- Cloud computing provides on-demand access to configurable computing resources over a network. There are different service and deployment models.
- Potential advantages of cloud computing for SDIs include scalability, pay-as-you-go costs, and not needing dedicated servers. However, organizations need to keep control of important components and ensure security and privacy.
- Switzerland's Federal Office of Topography uses an infrastructure
Building Spatial Data Infrastructures for Spatial Planning in Africa: Lagos e...Samuel Dekolo
Lagos is the fastest growing Megacity in Sub-Saharan Africa, with its population estimated to double in the first quarter of this century; it is expected to be the third largest urban agglomerations in the world. This growth is not without challenges, as the city is grappling with myriads of urban management problems. City planners lack the most important ingredient of land use management, which is Information. In spite of huge investment on spatial data infrastructures at the national and state levels of government, most land use planners at both state and local government level agencies are ignorant of existing geospatial technology portals and unlock the full potentials of information and communication technologies. A statewide survey of the spatial data infrastructures of the city’s urban and land use management ministry and agencies proves its pathetic state, thereby creating information gap void between urban development and intelligent management. The result is has led to a sporadic growth of slums and unplanned settlements which now accounts for over 60% of the city. To avoid an impasse, it is necessary to review the level of geospatial technologies used at the local level and recommend formidable means of integration in the decision making process. This paper examines the level of geospatial technologies and Spatial Data Infrastructure use in spatial planning agencies and barriers to implementation in the 20 local governments of Lagos State and suggests the way forward.
András Sipos (Budapest City Archives; HU) presenting the e-proceedings of the ICARUS workshop “Cartography and Cadastral Maps: Visions from the past, for a vision of our future” - held in Pisa, October 2013 - on 28.04.2015 at the international conference "Archival Cooperation and Community Building in the Digital Age" within the panel “Building from scratch: written heritage online" at Břevnov Archabbey in Prague (CZ).
Spatial Data Infrastructure Best Practices with GeoNodeSebastian Benthall
The document discusses GeoNode, an open source software project that aims to build a next generation Spatial Data Infrastructure (SDI) based on best practices from modern web services. Some key points:
- GeoNode aims to make uploading, sharing, and working with spatial data as easy as blogging by reducing barriers to participation.
- It aligns incentives for contribution and use through features like user profiles, reputation systems, and ability to style and share maps.
- The software implements open standards and APIs to make data accessible and allow other applications to integrate.
Maps can be classified based on scale into small scale maps and large scale maps. Small scale maps show important information over larger areas like world maps and country maps. Large scale maps provide detailed information for small areas like village maps and cadastral maps which show land boundaries and ownership details. Topographical maps also fall under large scale maps and provide terrain and elevation details.
Topographic maps use contour lines to represent elevation and slope of land. Contour lines connect points of equal elevation and never cross. Closer lines indicate steeper slopes while widely spaced lines show more gentle slopes. Index contours are bold lines labeled with the elevation. A benchmark is a point of known exact elevation marked as B.M. on maps. Map scale relates distances on a map to actual distances on land and can be ratio, graphical, or verbal.
Introduction to aerial photography and photogrammetry.pptsrinivas2036
Aerial photography and photogrammetry are techniques used in remote sensing. Aerial photography involves taking photographs from aircraft and has been used since the 1850s. Photogrammetry uses photographs to measure and obtain spatial information about the objects and terrain photographed. It allows for the creation of topographic maps, cadastral maps, and large-scale construction plans more quickly and economically than traditional ground-based surveying. While aerial photography and photogrammetry provide advantages over field surveys, some on-site control and verification is still needed.
1. Levelling is used to determine relative heights and elevations of points and establish points at required elevations. It involves using instruments like levels and staffs.
2. There are different types of levels (dumpy, tilting, wye, automatic) and staffs (self-reading, target). Precise levelling is done to establish permanent benchmarks.
3. Adjustments must be made to level instruments during setup and permanently. Methods like differential, profile and cross levelling are used depending on the task. Reciprocal levelling involves backsight-foresight exchange to check for errors.
This document provides an overview of surveying concepts and techniques. It discusses:
1) The definitions, classifications, instruments, and methods used in surveying like chain surveying, compass surveying, plane table surveying, and total station surveying.
2) The objectives of surveying which include preparing maps, plans and transferring details to mark locations on the ground for engineering projects.
3) The primary divisions of surveying into plain surveying which ignores curvature of the earth, and geodetic surveying which accounts for curvature over large areas.
4) Fundamental surveying principles like working from the whole to parts, and locating new points using two measurements from fixed references.
Land use planning refers to allocating land resources for different uses consistent with development goals. It involves classifying land, determining suitable land uses, and regulating development through tools like zoning maps and land use plans. The objectives are to promote efficient land use, reconcile conflicts, and guide sustainable development patterns. Land use planning organizations at the national, regional, and local levels coordinate to classify land and guide land allocation and conversion nationally and within their jurisdictions.
Compass surveying involves measuring directions of survey lines using a magnetic compass and measuring lengths using a chain or tape. It is used when the area is large, undulating and has many details. In compass surveying, a series of connected lines are established through traversing. The magnetic bearing of each line is measured using a prismatic compass or surveyor's compass, and the distance is measured using a chain. Compass surveying is recommended for large and undulating areas without suspected magnetic interference. The key principles are measuring bearings using a compass and distances using a chain to establish connected lines through traversing without requiring triangulation.
Towards an INSPIREd e-reporting & INSPIRE priority datasets in SlovakiaMartin Tuchyna
SKI contribution for the Workshop "Priority list of datasets for eReporting" (https://inspire.ec.europa.eu/events/conferences/inspire_2017/submissions/222.html ) held during the INSPIRE 2017 conference (http://inspire.ec.europa.eu/conference2017).
This is an updated version of my ‘Introducing INSPIRE’ presentation.
The slides cover: the what/why/how of INSPIRE, the UK Location Information Infrastructure, and what you need to do to comply.
Both the slides and associated notes are available (click the "notes on slide x" at the top of the comments section of the page)
GLOSIS vision | GSP Soil Data Facility, ISRIC - Bas KempenExternalEvents
This document provides an overview of the Global Soil Information System (GloSIS) being implemented from 2017-2020. GloSIS aims to create a federated system that brings together soil data from various national and institutional sources according to common standards. It will include a domain model, data exchange protocols, reference nodes for hosting data, and a discovery hub. A key goal is also establishing national soil information systems through the CountrySIS program to encourage more sources to participate. The implementation is proceeding through developing technical specifications, piloting the system, and providing support and guidelines to establish decentralized nodes and increase global soil data interoperability and sharing. Challenges include limited funding and voluntary contributions as well as engaging more countries
The document outlines Rik van den Bosch's work plan for developing the Soil Data Facility (SDF) from 2018-2020. Key aspects of the plan include developing technical specifications for the Global Soil Information System (GloSIS) and its data products from 2017-2018, building GloSIS infrastructure and populating data products from 2018-2020. It also discusses specifications for GloSIS point databases and grids, a vision document outlining GloSIS, engaging data providers, and connecting providers and users. The document proposes roles for various organizations in developing GloSIS, with SDF leading technical development, FAO hosting and developing the discovery hub, and all organizations working together.
1) Postgres and PostGIS have been used at EDINA for over 8 years to power major geospatial services like Digimap.
2) It is used for data storage, mapping, spatial indexing, querying, and data downloads. Postgres allows EDINA to handle large amounts of geospatial data and large user bases.
3) EDINA finds Postgres reliable, performant, scalable, and standards-compliant with good support tools. It will continue being the core database for EDINA's geoservices.
Promote the Good of the People of the United Kingdom by Maintaining Monetary ...DataWorks Summit
The Bank of England is the central Bank of the United Kingdom, established in 1694. Representatives from the Bank’s Data Analytics & Modelling team will discuss the Bank of England's journey to delivering a Big Data capability and how the Hortonworks HDP platform is helping us deliver on our mission statement of “promote the good of the people of the United Kingdom by maintaining monetary and financial stability". We will explore the challenges we've faced, how we have overcome some of these and those that remain to be conquered. We will also present our strategy for the Bank’s future Big Data platform as we look to scale up further in the coming years.
We will focus in particular on our first successful ‘Big Data’ production system. This exists in response to the financial crises of 2008 and the subsequent push to make the derivative markets safer by reducing systemic risk. In Europe this was delivered through the European Market Infrastructure Regulation (EMIR). We will explain the Bank of England’s role in monitoring UK entities within this important market and describe the significant challenges facing our team in building a data analytics platform to facilitate this
Speakers
Nick Vaughan, Domain SME - Data Analytics & Modelling
Bank of England
Adrian Waddy, Technical Lead
Bank of England
BigDataEurope 1st SC5 Workshop, Project Teleios & LEO, by M. Koubarakis, Univ...BigData_Europe
The TELEIOS and LEO projects developed techniques for managing and analyzing large volumes of linked open Earth observation data. Key technologies included the Strabon spatial RDF store, MonetDB SciQL for scientific queries, GeoTriples for transforming data to RDF, and Sextant for visualization. Applications included semantic catalogs, wildfire monitoring, burn scar mapping, and precision farming. Lessons indicated open EO data will grow significantly and integrating it with linked data can support many applications, requiring scientific database and semantic web technologies.
The document discusses the key components and concepts of a National Spatial Data Infrastructure (NSDI). An NSDI aims to integrate distributed geospatial data through partnerships between different levels of government and private organizations. It establishes standards, frameworks and metadata to facilitate discovery and sharing of geospatial data. Central to an NSDI is a clearinghouse that allows users to search metadata from distributed servers according to common protocols. When properly implemented through the coordination of stakeholders, an NSDI can help reduce data duplication, lower costs and make critical spatial information more accessible.
Next Generation Open Data Platforms | AWS Public Sector Summit 2016Amazon Web Services
The document discusses the evolution and future of open data platforms like GeoPlatform.gov, which implements goals of the National Spatial Data Infrastructure Strategic Plan. Key points include:
- GeoPlatform.gov provides a consistent portal for open geospatial data and services, and shared hosting of these resources.
- Recent improvements include new map viewers, analytics dashboards, and tools for data discovery, sharing, and collaboration.
- Future plans include expanding data registries, monitoring service quality, and providing more open APIs to developers.
- The overall aim is to advance open data sharing and make national geospatial resources more accessible, interoperable, and usable.
This document summarizes a presentation given at the Miami Valley Planning & Zoning Workshop on GIS in the Miami Valley region. It discusses the results of a 2015 GIS needs assessment survey, including what data and software are currently used. It also outlines several GIS projects undertaken by MVRPC, including a recreational asset inventory map, updated regional data layers, and new online mapping applications. Lastly, it discusses future GIS training opportunities and plans to continue providing GIS services and expertise to member organizations.
This document summarizes a presentation given at the Miami Valley Planning & Zoning Workshop on GIS in the Miami Valley region. It discusses the results of a 2015 GIS needs assessment survey, including what data and software are currently used. It also outlines several GIS projects undertaken by MVRPC, including a recreational asset inventory map, updated regional data layers, and new online mapping applications. MVRPC aims to continue providing GIS services, training, and collaborative regional mapping applications to share data and expertise with local jurisdictions.
Progress in the implementation of the endorsed Plans of Actions: Pillar 4 – ...FAO
The document outlines progress on Pillar 4 of a Global Soil Information System plan of action. It discusses the establishment of an International Network of Soil Information Institutions to develop and maintain a Global Soil Spatial Data Infrastructure and Global Soil Information System. Near term deliverables include launching a SoilSTAT concept for monitoring soils, developing web platforms and data standards, and convening working groups to begin producing global soil maps, profiles and grids. Execution of the Pillar 4 Implementation Plan is seen as crucial to generating the next State of the World's Soil Resources report and monitoring soils for sustainable development goals.
This document summarizes a new cloud-based platform for analyzing large satellite Earth observation (EO) data. Key points:
- Traditional desktop tools cannot handle the growing size of EO data, requiring a shift to cloud-native analytics using big data tools.
- The demonstrated platform brings algorithms to distributed data using open-source tools like Kubernetes, Docker, and Jupyter notebooks running on Google Cloud.
- It provides scalable infrastructure for processing and analyzing petabytes of EO data through pipelines, supports open data standards, and ensures data sovereignty with Canadian storage.
This document outlines the planning process for a GIS analysis project to identify suitable sites for a new wastewater treatment plant. It discusses why many GIS projects fail, the types and phases of GIS projects. It then details the steps involved in a GIS analysis project including: defining the objective, creating the project database by identifying and preparing data, analyzing the data using spatial modeling functions, and presenting the results. For the wastewater treatment plant project, the analysis will identify parcels that meet criteria like being in a low-lying area near water and not residential, and the results will be shown on a presentation map.
Open Land Use - the Current Status and Steps Forwardplan4all
The document discusses the Open Land Use Map initiative which aims to create a detailed and freely available land use map of Europe. It summarizes the current status of the Open Land Use Map, which integrates heterogeneous land use data sources into a harmonized and INSPIRE compliant dataset. It also discusses plans to create an Open Land Use Map for Africa using satellite imagery and machine learning techniques to classify land use at the municipal level across Africa.
Driving Business Innovation: Latest Generative AI Advancements & Success StorySafe Software
Are you ready to revolutionize how you handle data? Join us for a webinar where we’ll bring you up to speed with the latest advancements in Generative AI technology and discover how leveraging FME with tools from giants like Google Gemini, Amazon, and Microsoft OpenAI can supercharge your workflow efficiency.
During the hour, we’ll take you through:
Guest Speaker Segment with Hannah Barrington: Dive into the world of dynamic real estate marketing with Hannah, the Marketing Manager at Workspace Group. Hear firsthand how their team generates engaging descriptions for thousands of office units by integrating diverse data sources—from PDF floorplans to web pages—using FME transformers, like OpenAIVisionConnector and AnthropicVisionConnector. This use case will show you how GenAI can streamline content creation for marketing across the board.
Ollama Use Case: Learn how Scenario Specialist Dmitri Bagh has utilized Ollama within FME to input data, create custom models, and enhance security protocols. This segment will include demos to illustrate the full capabilities of FME in AI-driven processes.
Custom AI Models: Discover how to leverage FME to build personalized AI models using your data. Whether it’s populating a model with local data for added security or integrating public AI tools, find out how FME facilitates a versatile and secure approach to AI.
We’ll wrap up with a live Q&A session where you can engage with our experts on your specific use cases, and learn more about optimizing your data workflows with AI.
This webinar is ideal for professionals seeking to harness the power of AI within their data management systems while ensuring high levels of customization and security. Whether you're a novice or an expert, gain actionable insights and strategies to elevate your data processes. Join us to see how FME and AI can revolutionize how you work with data!
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
The Zero-ETL Approach: Enhancing Data Agility and InsightSafe Software
In the ever-evolving landscape of data management, Zero-ETL is an approach that is reshaping how businesses handle and integrate their data. This webinar explores Zero-ETL, a paradigm shift from the traditional Extract, Transform, Load (ETL) process, offering a more streamlined, efficient, and real-time data integration method.
We will begin with an introduction to the concept of Zero-ETL, including how it allows direct access to data in its native environment and real-time data transformation, providing up-to-date information with significantly reduced data redundancy.
Next, we'll take you through several demonstrations showing how Zero-ETL can deliver real-time data and enable the free movement of data between systems. We will also discuss the various tools that support all aspects of Zero-ETL, providing attendees with an understanding of how they can adopt this innovative approach in their organizations.
Lastly, the session will conclude with an interactive Q&A segment, allowing participants to gain deeper insights into how Zero-ETL can be tailored to their specific business needs and how they can get started today.
Join us to discover how Zero-ETL can elevate your organization's data strategy.
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
Following the popularity of “Cloud Revolution: Exploring the New Wave of Serverless Spatial Data,” we’re thrilled to announce this much-anticipated encore webinar.
In this sequel, we’ll dive deeper into the Cloud-Native realm by uncovering practical applications and FME support for these new formats, including COGs, COPC, FlatGeoBuf, GeoParquet, STAC, and ZARR.
Building on the foundation laid by industry leaders Michelle Roby of Radiant Earth and Chris Holmes of Planet in the first webinar, this second part offers an in-depth look at the real-world application and behind-the-scenes dynamics of these cutting-edge formats. We will spotlight specific use-cases and workflows, showcasing their efficiency and relevance in practical scenarios.
Discover the vast possibilities each format holds, highlighted through detailed discussions and demonstrations. Our expert speakers will dissect the key aspects and provide critical takeaways for effective use, ensuring attendees leave with a thorough understanding of how to apply these formats in their own projects.
Elevate your understanding of how FME supports these cutting-edge technologies, enhancing your ability to manage, share, and analyze spatial data. Whether you’re building on knowledge from our initial session or are new to the serverless spatial data landscape, this webinar is your gateway to mastering cloud-native formats in your workflows.
Cloud Frontiers: A Deep Dive into Serverless Spatial Data and FMESafe Software
Following the popularity of "Cloud Revolution: Exploring the New Wave of Serverless Spatial Data," we're thrilled to announce this much-anticipated encore webinar.
In this sequel, we'll dive deeper into the Cloud-Native realm by uncovering practical applications and FME support for these new formats, including COGs, COPC, FlatGeoBuf, GeoParquet, STAC, and ZARR.
Building on the foundation laid by industry leaders Michelle Roby of Radiant Earth and Chris Holmes of Planet in the first webinar, this second part offers an in-depth look at the real-world application and behind-the-scenes dynamics of these cutting-edge formats. We will spotlight specific use-cases and workflows, showcasing their efficiency and relevance in practical scenarios.
Discover the vast possibilities each format holds, highlighted through detailed discussions and demonstrations. Our expert speakers will dissect the key aspects and provide critical takeaways for effective use, ensuring attendees leave with a thorough understanding of how to apply these formats in their own projects.
Elevate your understanding of how FME supports these cutting-edge technologies, enhancing your ability to manage, share, and analyze spatial data. Whether you're building on knowledge from our initial session or are new to the serverless spatial data landscape, this webinar is your gateway to mastering cloud-native formats in your workflows.
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
Imagine a world where information flows as swiftly as thought itself, making decision-making as fluid as the data driving it. Every moment is critical, and the right tools can significantly boost your organization’s performance. The power of real-time data automation through FME can turn this vision into reality.
Aimed at professionals eager to leverage real-time data for enhanced decision-making and efficiency, this webinar will cover the essentials of real-time data and its significance. We’ll explore:
FME’s role in real-time event processing, from data intake and analysis to transformation and reporting
An overview of leveraging streams vs. automations
FME’s impact across various industries highlighted by real-life case studies
Live demonstrations on setting up FME workflows for real-time data
Practical advice on getting started, best practices, and tips for effective implementation
Join us to enhance your skills in real-time data automation with FME, and take your operational capabilities to the next level.
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationSafe Software
Imagine a world where information flows as swiftly as thought itself, making decision-making as fluid as the data driving it. Every moment is critical, and the right tools can significantly boost your organization's performance. The power of real-time data automation through FME can turn this vision into reality.
Aimed at professionals eager to leverage real-time data for enhanced decision-making and efficiency, this webinar will cover the essentials of real-time data and its significance. We'll explore:
FME's role in real-time event processing, from data intake and analysis to transformation and reporting
An overview of leveraging streams vs. automations
FME's impact across various industries highlighted by real-life case studies
Live demonstrations on setting up FME workflows for real-time data
Practical advice on getting started, best practices, and tips for effective implementation
Join us to enhance your skills in real-time data automation with FME, and take your operational capabilities to the next level.
Beyond Boundaries: Leveraging No-Code Solutions for Industry InnovationSafe Software
Hiring and retaining software development talent is next to impossible for AEC firms and other industries alike.
Join us and guest speakers from HOK, a leader in the AEC industry, as they share their success in navigating the tight talent market through the use of no-code solutions and FME.
Discover how HOK approached the process of building a custom tool to automate the creation of projects and user management for Trimble Connect and ProjectSight.
Using a mix of traditional and no-code in FME, our guest speakers will reveal how the team bridged the resource gap and used the available talent pool, producing the mission-critical web app “Trajectory”.
They will also dive into details, illustrating first-hand how JSON data was used as a “glue” between two development groups.
Learn how embracing FME as a no-code solution can unlock potential within your teams, foster collaboration, and drive efficiency.
Powering Real-Time Decisions with Continuous Data StreamsSafe Software
In an era where making swift, data-driven decisions can set industry leaders apart, understanding the world of data streaming and stream processing is crucial. During this webinar, we'll explore:
Stream Processing Overview: Dive into what stream processing entails and the value it brings organizations.
Stream vs. Batch Processing: Learn the key differences and benefits of stream processing compared to traditional batch processing, highlighting the efficiency of real-time data handling.
Mastering Data Volumes: Discover strategies for effectively managing both high and low volume data streams, ensuring optimal performance.
Boosting Operational Excellence: Explore how adopting data streaming can enhance your organization's operational workflows and productivity.
Spatial Data's Role in Streams: Understand the importance of spatial data in stream processing for more informed decision-making.
Interactive Demos: Watch practical demos, from dynamic geofencing to group-based processing.
Plus, we’ll show you how you can do it without coding! Register now to take the first step towards more informed, timely, and precise decision-making for your organization.
The Critical Role of Spatial Data in Today's Data EcosystemSafe Software
In today's data-driven landscape, integrating spatial data is becoming increasingly crucial for organizations aiming to harness the full potential of their data. Spatial data offers unique insights based on location, making it a fundamental component for addressing various challenges across different sectors, including urban planning, environmental sustainability, public health, and logistics.
Our webinar delves into the indispensable role of spatial data in data management and analysis. We'll showcase how omitting spatial data from your data strategy not only weakens your data infrastructure, but also limits the depth of your insights. Through real-world case studies, we'll highlight the transformative impact of spatial data, demonstrating its ability to uncover complex patterns, trends, and relationships.
Join us for this introductory-level webinar as we explore the critical importance of spatial data integration in driving strategic decision-making processes. By the end of the webinar, you'll gain a renewed perspective on how spatial data is essential for confronting and overcoming challenges across various domains.
Cloud Revolution: Exploring the New Wave of Serverless Spatial DataSafe Software
Once in a while, there really is something new under the sun. The rise of cloud-hosted data has fueled innovation in spatial data storage, enabling a brand new serverless architectural approach to spatial data sharing. Join us in our upcoming webinar to learn all about these new ways to organize your data, and leverage data shared by others. Explore the potential of Cloud Native Geospatial Formats in your workflows with FME, as we introduce five new formats: COGs, COPC, FlatGeoBuf, GeoParquet, STAC and ZARR.
Learn from industry experts Michelle Roby from Radiant Earth and Chris Holmes from Planet about these cloud-native geospatial data formats and how they can make data easier to manage, share, and analyze. To get us started, they’ll explain the goals of the Cloud-Native Geospatial Foundation and provide overviews of cloud-native technologies including the Cloud-Optimized GeoTIFF (COG), SpatioTemporal Asset Catalogs (STAC), and GeoParquet.
Following this, our seasoned FME team will guide you through practical demonstrations, showcasing how to leverage each format to its fullest potential. Learn strategic approaches for seamless integration and transition, along with valuable tips to enhance performance using these formats in FME.
Discover how these formats are reshaping geospatial data handling and how you can seamlessly integrate them into your FME workflows and harness the explosion of cloud-hosted data.
Igniting Next Level Productivity with AI-Infused Data Integration WorkflowsSafe Software
Learn where FME meets AI in this upcoming webinar to offer you incredible time savings. This webinar is tailored to ignite imaginations and offer solutions to your data integration challenges. As the new digital era sets sail on the winds of AI, the tangibility of its integration in our daily schema is unfolding.
Segment 1, titled “AI: The Good, the Bad and the FME” by Darren Fergus of Locus, navigates through the realms of AI, scrutinizing its pervasive impact while underscoring the symbiotic potential of FME and AI. Join in an engaging demonstration as FME and ChatGPT collaboratively orchestrate a PowerPoint narrative, epitomizing the alliance of AI with human ingenuity.
In Segment 2, “Integrating GeoAI Models in FME” by Dennis Wilhelm and Dr. Christopher Britsch of con terra GmbH, the spotlight veers towards operationalizing AI in our daily tasks through FME. A practical approach to embedding GeoAI Models into FME Workspaces is unveiled, showcasing the ease of incorporating AI-driven methodologies into your FME workflows, skyrocketing productivity levels.
To follow, Segment 3, "Unleash generative AI on your terms!" by Oliver Morris of Avineon-Tensing. While the prospects of Generative AI are thrilling, security and IT reservations, especially with 'phone home' tools, are genuine concerns. However, with open-source tools, you can locally harness large language models. In this demo, we'll unravel the magic of local AI deployment and its seamless integration into an FME workspace.
Bonus! Dmitri will join us for a fourth segment to tie us off, showcasing what he has been up to this week, including using OpenAI API for texturing in FME, amoung other projects.
Join us to explore the synergy of FME and AI: opening portals to a realm of revolutionized productivity and enriched user experiences.
The Zero-ETL Approach: Enhancing Data Agility and InsightSafe Software
In the ever-evolving landscape of data management, Zero-ETL is an approach that is reshaping how businesses handle and integrate their data. This webinar explores Zero-ETL, a paradigm shift from the traditional Extract, Transform, Load (ETL) process, offering a more streamlined, efficient, and real-time data integration method.
We will begin with an introduction to the concept of Zero-ETL, including how it allows direct access to data in its native environment and real-time data transformation, providing up-to-date information with significantly reduced data redundancy.
Next, we'll take you through several demonstrations showing how Zero-ETL can deliver real-time data and enable the free movement of data between systems. We will also discuss the various tools that support all aspects of Zero-ETL, providing attendees with an understanding of how they can adopt this innovative approach in their organizations.
Lastly, the session will conclude with an interactive Q&A segment, allowing participants to gain deeper insights into how Zero-ETL can be tailored to their specific business needs and how they can get started today.
Join us to discover how Zero-ETL can elevate your organization's data strategy.
Mastering MicroStation DGN: How to Integrate CAD and GISSafe Software
Dive deep into the world of CAD-GIS integration with our expert-led webinar. Discover how to seamlessly transfer data between Bentley MicroStation and leading GIS platforms, such as Esri ArcGIS. This session goes beyond mere CAD/GIS conversion, showcasing techniques to precisely transform MicroStation elements including cells, text, lines, and symbology. We’ll walk you through tags versus item types, and understanding how to leverage both. You’ll also learn how to reproject to any coordinate system. Finally, explore cutting-edge automated methods for managing database links, and delve into innovative strategies for enabling self-serve data collection and validation services.
Join us to overcome the common hurdles in CAD and GIS integration and enhance the efficiency of your workflows. This session is perfect for professionals, both new to FME and seasoned users, seeking to streamline their processes and leverage the full potential of their CAD and GIS systems.
Geospatial Synergy: Amplifying Efficiency with FME & EsriSafe Software
Dive deep into the world of geospatial data management and transformation in our upcoming webinar focusing on the powerful integration of FME and Esri technologies. This insightful session comprises two compelling segments aimed at enhancing your geospatial workflows, while minimizing operational hurdles.
In the first segment, guest speaker Jan Roggisch from Locus unveils how Auckland Council triumphed over the challenges of handling large, frequent data updates on ArcGIS Online using FME. Discover the journey from manual data handling to an automated, streamlined process that reduced server downtime from minutes to seconds: setting a new standard for local government organizations.
The second segment, led by James Botterill from 1Spatial, unveils the magic of incorporating ArcPy into your FME workflows. Delve into real-world scenarios where ArcGIS geoprocessing is harmoniously orchestrated within FME using the PythonCaller. Gain insights into raster-vector data conversion, spatial analysis, and a host of practical tips and tricks that empower you to leverage the combined capabilities of FME and Esri for efficient data manipulation and conversion.
Join us to explore the remarkable possibilities that open up when FME and Esri technologies converge – enhancing your ability to manage and transform geospatial data with unprecedented efficiency.
Introducing the New FME Community Webinar - Feb 21, 2024 (2).pdfSafe Software
Join us at Safe Software as we unveil the exciting new FME Community platform.
Picture yourself entering a vibrant, interconnected world, where every click brings you closer to a fellow FME enthusiast, a new idea, or a solution that could revolutionize your workflow.
Since its inception, the FME Community has been a dynamic hub for knowledge sharing, where thousands of users converge to exchange insights, engage in stimulating discussions, and collaboratively solve challenges. Now, envision this community reimagined - retaining the features you know and love, but infused with new, cutting-edge functionalities designed to make your experience even more enriching and effortless. The Community is also planned to soon act as a central hub for all FME community acticity across the web.
This webinar is your personal tour through this enhanced FME Community landscape. Whether you're an experienced user familiar with every nook and cranny of the old platform, or you're setting foot in this community for the first time, our webinar will ensure you navigate the new terrain with ease and confidence. Discover how to maximize your engagement, tap into the wealth of resources available, and contribute to the growing tapestry of FME innovation.
Join us in celebrating the future of FME collaboration, where your next breakthrough idea, insightful article, or spirited discussion awaits. Don't miss this opportunity to be a part of the evolution of the FME Community!
Breaking Barriers & Leveraging the Latest Developments in AI TechnologySafe Software
Explore how to best leverage the latest of AI technology in our upcoming webinar, where we delve into advancements and trends in the field since our previous AI webinars in 2023. Join us for a session filled with fresh insights and practical knowledge. We're stitching together the final threads of this presentation as we speak, keeping pace with AI's breakneck speed. Expect a session brimming with the freshest insights, releases and breakthroughs in AI – right up to the minute! A spotlight of this session is set to include Dmitri Bagh’s exploration of innovative AI integrations with FME, ranging from generating 3D features for augmented reality using Dall-E, to enhancing urban planning with orthoimagery completion, and showcasing the power of AI in workspace analysis and geoart creation.
Whether you're new to AI or an experienced practitioner, this webinar is tailored to keep you at the forefront of AI innovation. Get ready for a session that is as informative as it is inspiring, equipping you with the tools to excel in the dynamic world of artificial intelligence.
Best Practices to Navigating Data and Application Integration for the Enterpr...Safe Software
Navigating the complexities of managing vast enterprise data across multiple systems can be challenging. This webinar is your guide to navigating and simplifying enterprise integration.
As a technology leader, you may grapple with legacy systems, shadow IT, and budget constraints. Data and personnel silos often impede technological progress. FME champions integrating superior business systems to bolster your organization's digital strength – efficiently and affordably, using your current team and accessible services.
Join us and partner guest speakers from Seamless in an engaging session exploring the essential roles of data and systems in modern enterprises. We'll provide insights on achieving high-quality data management, establishing strong governance, and enabling teams to manage their data effectively. Delve into strategies for ensuring high-quality data and building robust governance structures, with tips and tricks along the way.
This webinar features real-life case studies demonstrating success in diverse industries. Learn cutting-edge strategies for data governance and system integration. Don't miss this opportunity to gain valuable insights and best practices for transforming your data governance and system integration processes.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
UiPath Test Automation using UiPath Test Suite series, part 6DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 6. In this session, we will cover Test Automation with generative AI and Open AI.
UiPath Test Automation with generative AI and Open AI webinar offers an in-depth exploration of leveraging cutting-edge technologies for test automation within the UiPath platform. Attendees will delve into the integration of generative AI, a test automation solution, with Open AI advanced natural language processing capabilities.
Throughout the session, participants will discover how this synergy empowers testers to automate repetitive tasks, enhance testing accuracy, and expedite the software testing life cycle. Topics covered include the seamless integration process, practical use cases, and the benefits of harnessing AI-driven automation for UiPath testing initiatives. By attending this webinar, testers, and automation professionals can gain valuable insights into harnessing the power of AI to optimize their test automation workflows within the UiPath ecosystem, ultimately driving efficiency and quality in software development processes.
What will you get from this session?
1. Insights into integrating generative AI.
2. Understanding how this integration enhances test automation within the UiPath platform
3. Practical demonstrations
4. Exploration of real-world use cases illustrating the benefits of AI-driven test automation for UiPath
Topics covered:
What is generative AI
Test Automation with generative AI and Open AI.
UiPath integration with generative AI
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
UiPath Test Automation using UiPath Test Suite series, part 5DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 5. In this session, we will cover CI/CD with devops.
Topics covered:
CI/CD with in UiPath
End-to-end overview of CI/CD pipeline with Azure devops
Speaker:
Lyndsey Byblow, Test Suite Sales Engineer @ UiPath, Inc.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Large Language Model (LLM) and it’s Geospatial Applications
The Loading and Maintenance of Cadastral Data in the Spatial Data Infrastructure (SDI)
1. The Loading and
Maintenance of Cadastral
Data in Manitoba’s Spatial
Data Infrastructure (SDI)
Bob Bruce & Tony Viveiros
FME World Tour 2014
April 11
2. Presentation Outline
• Background to the SDI Project
• Description of the Cadastral Datasets
• Implementation of Table and Feature
Class to house the data
• Detailed description of Workbench files
and their operation
• Next steps
2
3. What is a SDI?
The technologies,
policies and people
necessary to enable the
use of spatially
referenced data through
all levels of government,
stakeholders and the
public.
SDI Vision:
Manitoba’s source
for authoritative
geospatial data.
3
4. Manitoba SDI Objectives
1.To build (design and implement) a SDI for the
Province of Manitoba that:
– stores, manages and disseminates Manitoba’s
geospatial data across government and to the public
– is scalable to meet anticipated future needs
2.To provide a SDI Governance model that
clarifies the tasks and policies required for
SDI success.
4
9. The Cadastral Datasets
• Accurately computed cadastral datasets
in Winnipeg
• Accurately computed cadastral datasets
in Manitoba
• Manitoba Reference Grid
• Northern Quarter Sections
9
10. Cadastral Datasets in Winnipeg
• Maps of areas of Winnipeg on the MLI• Area 6 of the Winnipeg Cadastral
Mapping
10
11. Cadastral Datasets in Manitoba
• In the MLI we
currently start by
accessing data
from a map of
the province to
go to subareas
• By clicking on a
subarea we go
to a map that
allows us to
select and
download data
from a dataset
11
13. Northern Quarter Section Grid
A computed
theoretical quarter
section grid of the
unsurveyed north
13
14. Maintaining a Seamless Cadastral
Polygon Layer
• Each of the cadastral polygon fabrics have
been fitted together and are ready to be
combined into one polygon fabric
• These fabrics are combined into a single
GeoDatabase polygon feature class
• The individual datasets are tracked with a
database table 14
15. Cadastral Jobs Table
A record of which cadastral datasets are
currently in the cadastral polygon feature
class
15
16. Workbench Files
• One workbench file to ‘decide’ whether
an input dataset is a new one or a
replacement one and run the processes
to populate the feature class and job
table
• Once workbench file (called by the first
one) to actually insert a dataset into the
polygon feature class
16
23. • Develop processes to import the
Northern ‘theoretical’ quarter section
grids
• Develop a workbench file that uses the
Directory and File Pathnames reader to
scan folders for datasets and call the
workbench files
Further Improvements
23
SDI Vision: Manitoba’s source for authoritative geospatial data.
SDI Operational Environment.
Provide ArcGIS Desktop in a CITRIX environment. This will ease deployment issues, offer 64bit computing benefits, and improve performance for users with limited bandwidth. ArcGIS Desktop creates a high amount of network traffic when working with data over the network, this causes poor performance when using network data on a slow connection. CITRIX will improve this on this problem because it is installed in the same data centre where the data is located making it as close as possible to the data over the network. The CITRIX client connection is optimized to provide a quality screen display on the client computer. Printing is done on the client computer to a local printer/plotter.
A SQL Server Enterprise Geodatabase (ArcSDE technology) will be available for Operational GIS usage. Each department will have its own database and security will be in place to give users a working area. All Esri vector data can be moved into the Enterprise Geodatabase. Sandbox testing will determine limitations. Working in this environment offers features that most users have not previously had access to in an Enterprise environment – multiuser editing, user security model, long transactions, and versioning. The data in the database will be backed up regularly. The design is scalable so that more server capacity can be added to meet growing demand.
Network file storage will be allocated for the centralization of GIS data. The expectation is that remaining operational vector data on local drives or existing network storage which is not immediately going into the Enterprise Geodatabase will be moved to this new location.
The new network and database storage will be available to CITRIX on a fast network link and will also be available to existing ArcGIS Desktop users over their provided network connections.
Imagery Storage: On bottom of screen a storage component for imagery and LIDAR will be added to provide storage for the SDI as well as desktop users. File based raster/lidar data will be stored here. Where possible, mosaics will be created that will be used by the SDI imagery services as well as on the desktops. This exists outside of a particular zone because it will be used by all.
Spatial Data Warehouse (SDW) center of slide. Provides source of authoritative data released by owners for use by other internal GOM users. SDW contains an Enterprise Geodatabase to store/serve published data and also contains an ArcGIS Server to provide this data through spatial web services. The spatial web services will be used by the internal web based dataviewer as well as GIS desktop clients. Metadata will be provided in this environment to provide additional information about the spatial datasets.
Internal GIS users can access the information using the GIS on their desktops or CITRIX. Non GIS users can access the data using a the internal webbased dataviewer.
Publishing environment is similar to SDW but located outside of the internal PDN to be accessible from the Internet.
External GIS users can use the spatial webservices from the PUB environment in their GIS, non GIS users can use the external webbased viewer to consume the spatial web services. External users will not be able to access the enterprise geodatabase directly as the internal GIS users can access the SDW. Internal users can access all aspects of the PUB environment. Metadata will be provided in the PUB environment.
The SDW and PUB ArcGIS Servers will publish imagery web services from source data stored in the Imagery storage at the bottom of slide. This maximizes usage by allowing the data to be used by all components of the SDI as well as desktop clients, and minimizes storage requirements by only having to store it once.
Extract Transform and Load (ETL) processes will be in place to move data from one environment to another.
Operational is considered a working area for data production. Select datasets in the Operational will be identified to be hosted in the SDW. Once data is ready to be moved into the SDW, ETL scripts will be developed to ETL the data into the SDW. The process can be one-time, scheduled frequency, or on demand. The data flowing from Operational to the SDW can be as is or can be manipulated in content and/or structure before it gets loaded into the SDW. Once it is loaded into the SDW it will be readonly visible to GOM. Spatial webservices will be configured to display the information as mapservices, featureservices, WMS/WFS...
The ETL process from SDW into PUB is very similar.
Winnipeg is subdivided into 342 individual areas each maintained separately with a total of 256,900 polygons
Manitoba outside of Winnipeg currently has 323 individual areas with a total of 290,500 polygons
The Manitoba Reference Grid is the original grid of quarter section mapping in the surveyed area of Manitoba combined and then subdivided into 9 areas. These tiles are maintained by making corrections as gross errors are discovered and areas are removed from them as areas of accurate cadastral mapping are created. The edges of these tiles are adjusted to fit the areas of accurate cadastral mapping that are added to the mapping of Manitoba.
The Northern Quarter Sections are calculated using the few positions of known surveyed points and by fitting the design dimensions of the township system to their positions and extending the design dimensions outward from the locations of known surveyed points
Winnipeg is subdivided into 342 individual areas each maintained separately with a total of 256,900 polygons
Winnipeg is now covered by mapping areas which are individually maintained as new survey plans are registered with The Property Registry
These graphics show how tiles are accessed via the MLI (Manitoba Land Initiative) by first selecting an area of the city from the overall map then by selecting a tile from a subarea
Manitoba outside of Winnipeg currently has 323 individual areas with a total of 290,500 polygons. Each of these individual areas is maintained separately as new survey plans are registered
We select a subarea of the province from the overall index map. From a subarea we can select individual datasets. The table below the map shows some details of the selected dataset including the version number
This fabric is a rough approximation of the township system and it is used to provide a map of this system to external users. It is not an ideal mapping product but it is the only one that we have. At various times gross errors are discovered in this grid and are repaired and the revised fabric is made available via the MLI.
The fabric was subdivided into 9 tiles in order to make it more manageable and it is maintained and made available via these tiles and also as a combined product. This fabric is also fitted to the areas of the accurate cadastral mapping that were described in the previous slide and portions of the reference grid are removed when new cadastral mapping is produced and ‘holes’ in this grid appear in these areas. Tiles in this fabric have version numbers and when they are revised the version numbers are incremented.
The graphic on the slide shows the page on the MLI where the fabric is made available to the public.
The product is a computed, topologically structured, theoretical quarter section grid of the unsurveyed north. Coordinates are in the NAD83 UTM projection.
The Northern Quarter Sections are calculated using the few positions of known surveyed points and by fitting the design dimensions of the township system to their positions and extending the design dimensions outward from the locations of known surveyed points
This grid is fitted to the Reference Grid and any accurate cadastral datasets that abut it
The sole exception to the fit of the polygon fabrics is the City of Winnipeg cadastral mapping. It was produced and still is maintained in the June 1990 computation of the NAD83 coordinates. Since that time several more accurate sets of coordinates for survey control have been produced and the coordinate set used to produce the City of Winnipeg is relatively inaccurate compared to the others. This means that the coordinates along the boundary between the City of Winnipeg cadastral mapping and the surrounding Manitoba accurate cadastral mapping are incompatible and may produce overlaps. How this situation will be resolved is still not determined
A single enterprise GeoDatabase polygon feature class provides the ability to combine all of the polygon fabrics and makes it available to all users within and external to Manitoba
This is the database table that records the current versions of the individual datasets contained in the cadastral polygon feature class. This is used to track the current version of the datasets and it will also be available to users to query the other attributes of the datasets such as, the latest survey plan numbers mapped in this version and the number of parcels that are contained in the dataset.
The FME is used to populate and update this table using the attribute information that is produced during the cadastral mapping operations.
The first workbench file checks the dataset version that is contained in the input data and looks for this dataset being present and for the version being a newer one than the one that is already present. This file calls the second workbench file to put the input dataset into the feature class if it is a new dataset or it removes the polygons for the dataset if there is a newer version of the file to be put into the feature class and then calls the workbench file to insert the new data.
The second workbench file is used to fill in fields that are not present in the input and performs checks and repairs on coordinates in the data (more on that later)
This operation has two main areas:
- the decision of which kind of operation is being done (inset of a new or update of existing dataset) and the insert of a new
dataset – this area is in the purple bookmark area
- the update of an existing dataset – this area is in the green bookmark area
The transformers circled in red are the WorkspaceRunner instances that run the second workspace to actually insert the new cadastral polygons into the feature class
This is a screen capture from running the process to update a City of Winnipeg cadastral dataset. In this case the data flows through the steps to process updates and City of Winnipeg data.
This process takes three input parameters: which area of Manitoba the accurate cadastral data is in (Winnipeg or Manitoba), the ACCESS database that the jobs tables are stored in, and the input SHAPE file.
This has three readers, one SHAPE file reader for the cadastral polygons and a reader for the Manitoba jobs table and the Winnipeg jobs table
This part of the workbench file starts with sampling one SHAPE file polygon, it then separates the input data into the streams for the City of Winnipeg and Manitoba because the job information for each of these types is contained in different database tables, then it goes on to use the FeatureMerger transformer to select the record for the job from the appropriate database table and attach all of the database field data to the sampled polygon. Next the Joiner transformer is used to attempt to get a record for this cadastral dataset from the SDI SDE jobs table. If a version of the dataset is already present in the feature class then the Tester transformer sends the processing down below to the update area of the process otherwise it proceeds to the right to the insertion of a new dataset.
The last part of the insertion of a new dataset into the GeoDatabase feature class involves calling the workbench file to do this task using the WorkspaceRunner transformer, if the run was successful then a success message is issued to the log and the cadastral jobs table is updated with the job information using the writer set up for that process.
The screen capture on the right side shows the configuration of this writer with the Advanced – Writer Mode value set to INSERT to allow data to be inserted into the table
This is the area of the workbench file where polygon datasets are input into the feature class as updates. If a previous version (a lower version number) of a dataset is in the feature class then that version is deleted and the new version is inserted.
First the version number of the job in the cadastral jobs table is tested against the version number contained in the sample polygon taken from the input dataset to see if it is an lower version number using the Tester transformer. If this test fails then an error message is generated and the process terminates. If a higher version number of the dataset is found then attributes are created to run SQL statements against the polygon feature class to remove the previous versions of the job. With these SQL attributes set up the workflow branches into the SDE writer that will use them to remove the previous version of the job. This next screen capture shows the SDE writer set up with the the Advanced – Writer Mode value set to DELETE to do the deletion.
The other branch proceeds with the insertion of a new dataset into the GeoDatabase feature class by calling the workbench file to do this task using the WorkspaceRunner transformer, if the run was successful then a success message is issued to the log and the workflow proceeds to make attributes to run SQL statements to update the cadastral jobs table. The next graphic shows the AttributeCreator parameters to make these attributes. Then the workflow proceeds to the SDE writer to perform the update to write the new attribute values to the cadastral job table. This writer is set up in the same way that was shown in the previous slide.
This was developed as a separate workbench file in order to allow tasks that were common to the insertion of cadastral polygons to be only written in one place. There are many fields that are added to the feature class to provide a textual explanation of the attribute where there was just a coded value in the attributes. <CLICK> Here is an example of one of the many calls to the AttributeValueMapper transformer that creates and populates an attribute based on the code contained in another attribute.
During the testing stage the FME was reporting topological problems with self intersecting polygons when I was importing one dataset that I knew had no problems with it. When I contacted SAFE Software support they said that « The problem is most likely due to the coordinate rounding that occurs when features are written to SDE. The coordinate shift due to the rounding can cause small self-intersections in the data”. They suggested using the ArcSDEGridSnapper transformer to simulate the coordinate conversion and the GeometryValidator transformer to fix the coordinate errors. Inserting these transformers and setting their parameters to the values that matched the SDE feature class allowed the geometrical errors to be fixed and the polygons to be inserted into the GeoDatabase feature class
The workbench files to import and update the accurate cadastral mapping and the Manitoba Reference Grid tiles have been tested and they are ready for operation. All nine of the Reference Grid tiles have been imported, four datasets within Winnipeg have been imported and updated and five datasets of the accurate cadastral mapping have been imported and updated. This graphic shows the polygon feature class displayed along with the provincial boundary dataset.