BBD is a library for helping teams perform political campaign data analysis.
BBD is available on PyPI:
python -m pip install bbd
For a running start, you can copy the examples folder to your machine and run the example python scripts to map median household income per census tract in Larimer county, Colorado (among other things).
Follow along with this walkthrough to learn more detailed infomation about using this library and it's submodules.
Module |
---|
Census |
GIS |
Geocoder |
First, find the data you want to visualize. This is often easiest through the census API, and the next bit of text will read a bit like a "How to Get Census Data" tutorial.
For our working example, we'll use median household income (which is coded in the census data as as "DP03_0062E").
We can simply downloaded the census data like so:
>>> from bbd import census
>>> data = census.get_acs(
>>> geography=census.Geography.CD,
>>> variables="group(DP03),NAME",
>>> year=2018,
>>> state="co",
>>> dataset=census.DataSets.ACS5_PROFILE,
>>> )
https://api.census.gov/data/2018/acs/acs1/profile?get=group(DP03),NAME&for=congressional%20district:*&in=state:08
For more information on how to come up with your own census API requests I'd highly recommend the first 6 minutes of this video. Or, you could simply start browsing census data sets. For reference, the following table describes the basic elements of the API call used to get this working example's data.
URL Part | Meaning |
---|---|
get=group(DP03) | Data columns include those in group DP03 (economics) |
,NAME | Include name of each entity, just a nicety |
for=congr...district:* | One row for each congressional district |
in=state:08 | Only get rows for state 08 (Colorado) |
This census data is stored more or less as a big table in json format:
[
["NAME", "GEO_ID", "DP03_0001E", ...] # Headers
["Congressional District 1", "5001600US0801", "693303", ...] # Data
["Congressional District 2", "5001600US0802", "672976", ...] # Data
["...", "...", "...", ...] # Data
]
When you request it from the API with get_acs
, it is automatically converted to a format that is plottable and join-able with shapefiles. If you downloaded the file manually and want to convert it yourself, you can use census.load_json_file
.
>>> print(data)
{
"NAME": ["Congressional District 1", "Congressional District 2", ...],
"GEO_ID": ["5001600US0801", "5001600US0802", ...]
"DP03_0001E": ["693303", "672976", ...]
}
Luckily, the census provides shapefiles for pretty much every GEO_ID
you can find in the census API. You can automatically download them like this:
>>> shapefile_path = census.get_shapefile(
>>> geography=census.Geography.CD,
>>> state="co",
>>> year=2018,
>>> cache=True,
>>> )
To get these files manually, just head to this website and select the relevent geography/location.
I happen to know that the GEO_ID in the shapefile is stored under the key "GEOID", and is only 4 digits long. There are reasons for this described here, but all we need to do is make an entry for our census data that matches the shapefile.
>>> data["GEOID"] = [geoid[-4:] for geoid in data["GEO_ID"]]
To determine how many digits you need for your own shapefile, I'd recommend just taking a quick look with PyShp.
python -m pip install pyshp
>>> import shapefile
>>>
>>> # What is the name of the GEO ID field?
>>> shp = shapefile.Reader(shapefile_path)
>>> shp.fields
[
["GEOID", "C", 4, 0], # Cool! Name of GEO ID field is "GEOID"
["ALAND", "N", 14, 0],
["AWATER", "N", 14, 0],
[...],
]
>>>
>>> # Now we need to see how many digits long the GEO ID is.
>>> for record in shp.records():
>>> print(record["GEOID"])
0801
0803
0805
...
>>> # Looks like 4 digits it is!
You can also of course simply view the property table in your favorite GIS software (e.g. the free QGIS).
The only thing better than a map is a map with color. You know the saying. The column to color by must be numeric.
>>> # Note: the name of this entry will be used as the text caption on the colormap
>>> data["Median Household Income"] = [float(x) for x in data["DP03_0062E"]]
We also want our map to have nice interactive tooltips so that you can see information about each shape by hovering over it with your mouse! To visualize the income, I'll just add comma separators and a $ sign.
>>> data["Median HouseHold Income (pretty format)"] = [
>>> "${:,.2f}".format(x) for x in data["Median Household Income"]
>>> ]
Finally, we can create the leaflet map and save it to our machine.
>>> geojson_map = gis.make_map(
>>> shapefile_path,
>>> data,
>>> join_on="GEOID", # Common field between shapefile and our census data
>>> color_by="Median Household Income", # Color shapes based on this data field
>>> save_to="path/to/save/map.html", # Save leaflet map here!
>>> )
Simply open path/to/save/map.html
in your favorite browser to view your map!
The geocoder module allows us to use OpenStreetMap's Nominatim api to geocode large batches of addresses slowly but for free!
Warning! Per Nominatim's Terms of Usage, this process may not be run in parallel.
If we have Addresses from VAN or another source that we want to map, we can use the geocoder submodule to get the Latitude, Longitude pair for each address.
Data exported to VAN will usually be in a tab-separated value text file containing columns with address components. As such, geocoder.LocationsGeocoder
can take in a DataFrame or list of dict's with address components. Also, geocoder.LocationsGeocoder
will accept lists of string addresses, but it will perform less effectively than using address components.
>>> from bbd import geocoder
>>> import pandas as pd
>>>
>>> # data can be pd.DataFrame, list of dicts w/ address components, or list of str addresses
>>> data = pd.read_csv("examples/data/example_addresses.tsv", sep = "\t")
>>> # Set a valid email to pass to Nominatim as user_agent.
>>> valid_email = 'replace_with_your_email@address.com'
>>> # Path to the save file for your geocoded results (Will be in tab-separated value format)
>>> path = r"path/to/save/file.tsv"
>>>
>>> # Make the geolocator object with data, a valid email, and a save path.
>>> geolocator = geocoder.LocationsGeocoder(data, valid_email, path)
>>> # Finally run the geolocator to begin the process.
>>> geolocator.run()
The Components of an address are:
Address Component | Alias |
---|---|
Street | Address |
City | |
County | |
State | |
Postalcode | Zip, Zip5, Postal |
Country* |
*Country is set to United States by default.
Say we are geocoding addresses all of which are in Volusia County, Florida but do not explicitly define the county. Some of these addresses are in Orange City, without setting the default county the Nominatim parser may incorrectly interpret the address as in Orange County and provide the wrong coordinates (real example).
>>> defaults = {
>>> 'county':'Volusia',
>>> 'state':'Florida',
>>> 'country':'United States'
>>> }
>>> # Defaults only loaded in if data is structured with Address Components.
>>> geolocator = geocoder.LocationsGeocoder(data, valid_email, path, defaults)
>>> geolocator.run()
Since only one address can be geocoded per second, this process is very long running for large datasets (n=100,000 is roughly 24hrs runtime). In order to allow for data verification before commiting the full runtime, LocationsGeocoder allows for running only a subset of data through the batch_size and num_batches arguments.
>>> # Setting batch_size & num_batches allows you to run a subset of data first before
>>> # committing to the long run time of the full dataset.
>>> geolocator = geocoder.LocationsGeocoder(data, valid_email, path, batch_size = 300)
>>> geolocator.run(num_batches = 2)
Developed by Bluebonnet Data
To develop bbd, first clone the repository, then set up a virtual environment locally
git clone https://github.com/bluebonnet-data/bbd
virtualenv venv
source venv/bin/active
Then install the local package:
pip install -e .
To ensure pytest is referencing your local package installation, you may also need to follow the steps here: https://stackoverflow.com/questions/35045038/how-do-i-use-pytest-with-virtualenv