Austrian cadastral map: Difference between revisions
No edit summary |
No edit summary |
||
Line 92: | Line 92: | ||
== Deliverables == | == Deliverables == | ||
Our deliverables consist into a folder containing all data from the 1848 cadaster used for the training our models and the Jupyter Notebook used for testing and geometries extraction. The training folder contains cropped images from the Austrian cadastral map and the corresponding masks, edges only, classes only and both. It contains also the txt files for the edge and classes colors and the csv files giving the path to the training images on the iccluster server. These are linked via the config files containing the parameters of the models. | |||
The Notebook is taken from the 1808 pipeline and adapted to our model. | |||
== Evaluation == | == Evaluation == |
Revision as of 13:29, 10 December 2020
Abstract
Many Venetian maps and cadasters were digitized in the Venice Time Machine project. Cadastral maps offer a detailed representation of properties in a specific area. As a part of the Time Machine, a pipeline was created to extract geometries from the 1808 Napoleonian cadastral map of Venice. The goal of this project is to use this pipeline to extract geometries from the 1848 Austrian cadaster and adapt it as necessary. The extracted geometries could then be used to compare the different shapes of the city of Venice in 1808, 1848 and eventually today.
Planning
Project steps
1. First overview
- Install the pipeline used on the 1808 cadaster and train a few models on it (done)
- Use the models trained to predict geometries on the Austrian cadaster (done)
2. Training on the Austrian cadaster
- Georeference the Austrian cadaster
- Prepare training data from the Austrian cadaster (make masks)
- Adapt model to the Austrian cadaster
- Train models on the prepared data
- Evaluate models
3. Optimizing of post-processing
- Understand the post-processing done on the probability maps
- Research similar solutions
- Try to optimize post-processing
4. Extensions (if time)
- Compare new 1848 geometries to 1808 ones
- Make some statistics on the similarities/differences
Timetable
Timeframe | Model | Post-processing |
---|---|---|
Weeks 8-10 | Finish georeferencing | Examine post-processing |
Week 10-11 | Prepare training data from Austrian cadaster, adapt model | |
Weeks 11-13 | Train and test models on Austrian cadaster | Research and optimize post-processing |
Week 13 | Combine and evaluate results | If time, compare with 1808 geometries |
Week 14 | Final Project presentation |
Methods
- Georeferencing
While working with maps and cadaster, an important task is the georeferencing of the images. Indeed, the geometries resulting from processing of the digitized cadastral map have to be comparable to others and placed correctly on a world map. To achieve this, the original cadaster images are georeferenced using QGIS, a geographic information system application. Common points between the image and a current map, such as OpenStreet Map, are identified and used by the georeferencement tool of QGIS to create a georeferenced tif document. This ensures that the geometries created from this document can be placed correctly on a world map.
- DHSegment-torch
Historical document segmentation has been an issue in Digital Humanities for a number of years, due to the diversity of these documents. DHSegment is a method that uses a generic CNN-architecture that can be used for multiple different processing tasks.
This method consists of two steps. The first step takes the images the type of documents to be processed and the masks associated as input to train a Fully Convolutional Neural Network. When given a new image corresponding to the same type of document, the network will output a map of label probabilities associated with each pixel. The second step is post-processing. It takes the probabilities map and using standard image processing techniques, transforms it to an output depending on the task.
In our project, we had to adapt the DHSegment-torch methods used on the 1808 cadaster to the 1848 one. Reusing the exact models was not conclusive as colour conventions were not the same between the cadasters. We had to make new masks ourselves on the 1848 cadaster images using GIMP. The masks use colours corresponding to the class labels that we wanted the model to recognise; edge, street, water, courtyard, building and church, see Fig. 1. These masks and the associated images were then given to the network to train different models. One model was trained to recognise only the edges, another only the classes and the last could recognise both.
The same post-processing as for the 1808 cadaster was used on the probabilities maps given as outputs of our models. First, a watershed is done using the local minima taken from the predicted edge map as markers. Then, cv2.findContours, a function often used to find the shapes in image processing, is used to create the geometries from the watershed. The geometries are assigned the majoritarian class and transformed into a GeoJSON file that can be opened in QGIS. Fig. 2 shows an example of geometries extracted on a sample image and opened in QGIS.
Motivation
DHSegment is now widely used for very diverse historical documents segmentation tasks. The goal of this project was to assess if the DHSegment pipeline used to extract geometries from the 1808 Napoleonian cadaster could be reused for other cadasters using different standards.
To test this, the 1848 Austrian cadastral map was used. After training models on the existing images and masks from the 1808 cadaster, we tested them on samples from the 1848 cadaster.
Deliverables
Our deliverables consist into a folder containing all data from the 1848 cadaster used for the training our models and the Jupyter Notebook used for testing and geometries extraction. The training folder contains cropped images from the Austrian cadastral map and the corresponding masks, edges only, classes only and both. It contains also the txt files for the edge and classes colors and the csv files giving the path to the training images on the iccluster server. These are linked via the config files containing the parameters of the models. The Notebook is taken from the 1808 pipeline and adapted to our model.