Skip to content

Latest commit

 

History

History
196 lines (141 loc) · 10.2 KB

README.md

File metadata and controls

196 lines (141 loc) · 10.2 KB

Map Gardening

A collection of scripts by Alan McConchie for the analysis of "map gardening" in OpenStreetMap

Installation

Requires a database of the format produced by the importer component of the osm-history-renderer. Most of the following steps are just to get the osm-history-renderer running.

I originally ran these tools on OSX, but more recently ported them to run on Amazon Web Services (AWS). For more instructions about setting up an AWS instance, follow AWS.md first.

Get an osm history dump by downloading history-YYYY-MM_DD.osm.pbf from here: http://planet.openstreetmap.org/planet/experimental/

curl http://planet.openstreetmap.org/pbf/full-history/history-latest.osm.pbf -o history-2015-01-26.osm.pbf

Alias this from .osm.pbf to .osh.pbf

sudo apt-get install git
git clone https://github.com/almccon/osmium
git clone https://github.com/almccon/osm-history-splitter
git clone https://github.com/almccon/osm-history-renderer
sudo apt-get update
sudo apt-get install libboost-dev zlib1g-dev libshp-dev libgd2-xpm-dev libgdal1-dev libexpat1-dev libgeos++-dev libsparsehash-dev libv8-dev libicu-dev libprotobuf-dev protobuf-compiler doxygen libboost-test-dev libosmpbf-dev clang make
cd osmium
sudo make install
cd ..
cd osm-history-splitter
sudo make install

Now, do the splitting. This seems to require a machine with a lot of memory. The extracted files will end up in the current directory. On EC2 copy them to a mounted disc so they can be accessed by a smaller instance.

./osm-history-splitter --hardcut /mnt/ebs/history-2014-12-22.osh.pbf osm-history-splitter.config
cd osm-history-renderer
sudo apt-get install libproj-dev postgresql postgresql-client postgresql-9.3-postgis-2.1 pgadmin3 postgresql-contrib
cd importer # inside the osm-history-renderer directory
make

Now we create our database and import the extract. Generally, instead of osm-history-render I use osm-history-render-place as the database name, and keep each extract in a separate database. Then make sure this database name is saved in MapGardening/__init__.py so the scripts know which place corresponds to which database.

sudo -u postgres createuser $USER --superuser -P # when prompted, use the same password stored in __init__.py

createdb osm-history-render
psql -d osm-history-render -c "create extension postgis;"
psql -d osm-history-render -c "create extension hstore;"

cd importer # inside the osm-history-renderer/importer/ directory
./osm-history-importer --dsn "dbname='osm-history-render'" /mnt/ebs/losangeles.osh.pbf

(currently this fails for me after closing the polygon table)

Message:

ERROR:  data type timestamp without time zone has no default operator class for access method "gist"
HINT:  You must specify an operator class for the index or define a default operator class for the data type.
terminate called after throwing an instance of 'std::runtime_error'
  what():  command failed
Aborted (core dumped)

Despite the core dump, this process seems to result in a usable database.

Or to initialize several at the same time: (test this for escapes, and for crashing at the importer step)

cd importer # inside the osm-history-renderer directory
for place in amsterdam auckland barcelona bayarea berlin boston buenosaires cairo chicago crimea cyprus douala haiti istanbul jakarta jerusalem kathmandu lasvegas london losangeles manchester mexicocity miami minsk montevideo montreal moscow mumbai nairobi newyork quebec paris rio santiago seattle seoul sydney tirana tokyo toronto vancouver yaounde
do
  createdb osm-history-render-$place
  psql -d osm-history-render-$place -c "create extension postgis;"
  psql -d osm-history-render-$place -c "create extension hstore;"
  ./osm-history-importer --dsn "dbname='osm-history-render-$place'" /mnt/ebs/$place.osh.pbf || true
  psql -d osm-history-render-$place -c "create index hist_point_idx on hist_point using gist(geom);"
  psql -d osm-history-render-$place -c "VACUUM ANALYZE hist_point;"
done

Doing the analyses

The python scripts have one more dependency:

sudo apt-get install python-psycopg2

Then initialize the analysis table:

./initialize_nodetable.py -p place

When the databases have been loaded, run the mapgardening analysis scripts.

  • ./proximity_check.py -p $place
  • ./raster_stats.py # runs for all places by default (currently doesn't work)
  • ./user_analysis.py # runs for all places by default

We also need to run a modified version of @mvexel's UserStats script that sums up the user activity by month. git clone https://github.com/almccon/OSMQualityMetrics, and follow the instructions there to make sure that osmium is also installed.

cd OSMQualityMetrics
for place in amsterdam auckland barcelona bayarea berlin boston buenosaires cairo chicago crimea cyprus douala haiti istanbul jakarta jerusalem kathmandu lasvegas london losangeles manchester mexicocity miami minsk montevideo montreal moscow mumbai nairobi newyork quebec paris rio santiago seattle seoul sydney tirana tokyo toronto vancouver yaounde
do
echo $place
../osmium/osmjs/osmjs -j UserStats.js -l array /mnt/ebs_large/$place.osh.pbf
mv userstats.csv userstats_$place.csv
done

Finally move those userstats_*.csv files into the userstats/ directory in the mapgardening/ repo. Also make sure that the output_userstatsbydate_*_raster_1000m.tsv files are in the same folder. Then we combine the analysis outputs into a form that the visualization tool can understand:

  • ./combine_stats.py

(modify MapGardening/__init__.py for configuration)

Add a password to the ubuntu account on postgres. Log into psql: \password ubuntu

Raster debugging

The primary output of the analysis is the suite of tsv files generated by the user_analysis.py script. However, some useful information is also stored in the rasters that are used for the analysis. It can be interesting to dump these rasters out to GeoTIFFs and look at them.

These loops will dump every raster file for every listed place. (If you ran the raster analysis more than once, you might get more than one raster as a result. This snippet only dumps rasters that were generated at 1000 m resolution.

for place in amsterdam auckland barcelona bayarea berlin boston buenosaires cairo chicago crimea cyprus douala haiti istanbul jakarta jerusalem kathmandu lasvegas london losangeles manchester mexicocity miami minsk montevideo montreal moscow mumbai nairobi newyork quebec paris rio santiago seattle seoul sydney tirana tokyo toronto vancouver yaounde; 
do
for resolution in 1000;
do
for table in `psql -t -P pager=off -d osm-history-render-$place -c "select tablename from blankspot_manager where resolution = $resolution;" 2>/dev/null | sed s/blankspots_raster_${resolution}_//`; 
do 
echo $place $table $resolution; psql -d osm-history-render-$place -c "COPY (SELECT encode(ST_AsTIFF(rast, 'LZW'),'hex') AS rasttiff FROM raster_$table) TO '/tmp/raster_${resolution}_${place}_$table.hex'" && xxd -p -r /tmp/raster_${resolution}_${place}_$table.hex > /tmp/raster_${resolution}_${place}_$table.tiff; 
done;
done;
done

Then to download these to a local machine for viewing in a GIS, simply do this (replacing the IP address with the address of your current AWS instance)

scp -i ~/.ssh/osmhistory.pem 'ubuntu@52.12.120.7:/tmp/*.tiff' .

Each raster contains three bands. Band 1 contains the number of nodes found in each cell. Band 2 contains the date of the earliest node in each cell, as an integer of the format YYYYMMDD. Band 3 contains the integer user id of the first editor in each cell.

For Tilemill mapping and debugging

git clone https://github.com/mapbox/tilemill
cd tilemill
sudo add-apt-repository ppa:developmentseed/mapbox
sudo apt-get install npm
sudo apt-get update
sudo apt-get install tilemill libmapnik nodejs
npm install

Read some of this: https://www.mapbox.com/tilemill/docs/linux-install/

Then:

nodejs /usr/share/tilemill/index.js export every-line-ever output.png

What is map gardening?

For now, let me explain by offering the abstract from my presentation at the 2013 Association of American Geographers conference:

While the volume of Volunteered Geographic Information (VGI) continues its rapid growth, many collaborative mapping projects such as OpenStreetMap (OSM) are entering a mature phase, where the primary activity shifts away from the initial creation of data to its ongoing maintenance. As OSM approaches "completeness" in some geographic areas, there remain the less glamourous--but no less important--tasks of adding attribute data and periodically updating the map to reflect changes on the ground.

Most research to date does not differentiate between various types of interaction with VGI datasets, thereby leaving many questions unanswered. For example: are the users who initially contribute data the same users who maintain it later? Do the types of tasks undertaken by users change over the duration of their engagement with a project? Are initial contributors more or less likely to have local knowledge of a place than those who maintain that data later?

In this presentation we propose a typology of modes of interaction with VGI datasets, specifically distinguishing between the acts of adding new data and editing existing data. We adapt the term "wiki gardening" from research on Wikipedia, and propose the term "map gardening" to describe volunteer-led maintenance of VGI data. We apply this concept to the OpenStreetMap dataset, presenting a quantitative analysis of the extent and frequency of map gardening, and making comparisons across different spatial and temporal scales and between individual users. We argue that this typology offers a new insight for understanding the evolution and social dynamics of crowdsourced geospatial datasets.

This github repository includes some of the scripts I am using to perform this quantitative analysis. At this point they may not be complete or functional, but in time I hope to post all my code here.

For more information, see this blog post, or watch this video from the 2013 State of the Map US Conference