Skip to content

A dataset on unclassified training activities for non-US security forces arranged and funded by the United States Department of State and Department of Defence between 2001 and 2021. Sources, scraping, cleaning and publishing toolset included. Search and explore the data at: https://trainingdata.securityforcemonitor.org

License

security-force-monitor/fmtrpt_data

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

17 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Foreign Military and Training Report: Parsing project

A dataset on unclassified training activities for non-US security forces arranged and funded by the United States Department of State and Department of Defence between 2001 and 2021. Sources, scraping, cleaning and publishing toolset included.

Access the data at: https://trainingdata.securityforcemonitor.org

1. Overview

At Security Force Monitor we create data on organizational structure, command personnel, geographical footprint of police, military and other security forces. We connect this to information about human rights abuses, assisting human rights researchers, litigators and investigative journalists in their work to hold security forces to account. All our data is published on WhoWasInCommand.com.

A key question we have is whether specific units and personnel implicated in human rights abuses have received training or assistance from the United States government. Whether the training or assistance was given prior to or after the unit's implication in a human rights abuse, it raises important questions about the effectiveness of the implementation of the "Leahy laws", which are a vetting and due diligence processes in place to prevent this from happening.

These questions can in part be answered by looking at the joint Department of State and Department of Defence report, "Foreign Military Training and DoD Engagement Activities of Interest". Released annually since 2000 (the 2000 report covered the U.S. government's fiscal year 1999-2000) this important, statutorily-mandated report shows in great detail how the US has spent much of its training and assistance budget, and with what aims. Generally, the US Department of State will release these reports as PDFs. They can be found at the following locations online:

Training activities that were completed in each fiscal year are recorded in Volume I: Section IV of the report, mostly in tabular form. For most years this data includes the name of the training course, the trainee unit, their country, the exact start and end date of the training, and so on. We include more detail on this below.

The value of the data is high, but its accessibility is low because of the way the data are published. To establish, for example, every training a single student unit had received since 1999, an analyst would have to manually search over 80 different PDFs.

This project addresses this challenge by providing clean, accurate and standardized machine-readable copy of all the data contained in the reports. We also aim to provide a simple and effective way for anyone to search the data and retrieve it in a standard, easy to use format. Importantly, we also show how the data were created and processed, keeping a direct link between the original report and the data. We also wanted to create a way to quickly update and analyse the dataset with each new release from the Department of State.

There is an existing dataset created by the team at the excellent Security Assistance Monitor (SAM). The platform tthey have created gives a great way to explore the dataset, setting out spending and delivery trends really clearly. For our needs, there are two limitations to this dataset and platform. First, though SAM have included the fiscal year in which a training was carried out, they have not included the data on start date and end date of every training, which are critical to our specific research aims. Second, at the time we initiated this project in 2019 the detailed dataset ended at FY 2015-2016 and we were unclear whether SAM had future plans to revisit and extend the data. Unfortunately, it was easier to start from scratch than to attempt to extend this existing work, hence this project. The ideal future situation, of course, is that the DoD/DoS would release this important data in a machine readable form in the first place so no parsing was necessary. We live in hope!

Longer term, our aim is to extend this dataset to include non-US security assistance and training, including that provided bilaterally by governments, and by international organizations like the United Nations and the European Union. We also intend to integrate this dataset with the information on our WhoWasinCommand.com platform, which will give our community even greater insight into who has provided support to the security forces they are researching.

2. Data and data model

The final dataset is comprised of the fields described in the table below. Fields prefixed with qa are "quality assurance" fields, included to assist with auditing the data. The fieldnames were updated in August 2021, and training:id:admin was created to create a stable way to reference a specific training or assistance intervention in the dataset.

Field Description Example of use
training:id:admin Unique code for each training item in the dataset, which enables referencing "6077cea5-e571-4fa9-b2aa-f0ead4d34760"
training:source Unique code used by SFM to identify the source of the data "048fb2d9-6651-4ba0-b36a-a526539f4cfd"
qa:training_group file containing the first, raw scrape of data SFM made; use for audit "2001_2002_Africa_fmtrpt.tsv"
training:country Country of training recipient "Cote d'Ivoire"
training:country_iso_3166_1 ISO-3166-1 standard code for country of training recipient "CIV"
training:program DoD/DoS financing program East Asia and Pacific Region, Thailand, International Military Education and Training (IMET), FY01
training:course_title Name of training course "HOT SHIP XFER 180 FT PAC"
training:delivery_unit US security force unit(s) involved in delivery of training "1/10 SFG, 7 SOS"
training:recipient_unit Name of recipient(s) of training "SPECIAL SERVICES GROUP NAVY, ST-3 PLT F, SBU-12 RIB DET N, SDVT-1 TU BAHRAINI NAVY"
qa:training_start_date Start date of training as stated in source (mostly m-dd-yy) "1/16/01"
training:start_date Start date of training, in ISO 8601 (YYYY-MM-DD) "2001-01-16"
qa:training_end_date End date of training (mostly m-dd-yy) "1/31/01"
training:end_date End date of training as stated in source, in ISO 8601 (YYYY-MM-DD) "2001-01-31"
training:location Place where the training was held "LACKLAND AFB, TX 78236"
training:quantity Number of trainees "30"
training:total_cost Total cost of that specific training, in USD "8000"
qa:training_source_url Specific PDF or web-page from which the row is extracted. Included as a helper. "https://2009-2017.state.gov/documents/organization/10967.pdf"
training:status:admin Whether the row has been hand checked against the source "Not checked against source; verify accuracy before use"
qa:training_date_first_seen Date that row of data was first made available publicly (YYYY-MM-DD, or parts thereof) "2002-03"
qa:training_date_scraped Date that row of data was obtained and scraped by Security Force Monitor to create this dataset (YYYY-MM-DD) "2019-07-17"

3. Data sources and coverage

In the table below, we provide an overview of the available source materials, where to get them, what data are contained within them, and whether we have included the material in the dataset. In summary, we have included everything from the 2001-2002 onwards. With the exception of a four year streak where data in the US unit field was not included in the data, the field coverage is consistent year on year.

Source (Vol. 1 - Training Activities) Included in dataset? Country DOD/DOS Program Course Title US Unit Student Unit Start date End date Location Quantity Total Cost
FY 2020-2021 PDF: All, A, EAP,E, NE, SCA, WH Yes
FY 2019-2020 PDF: All Yes
FY 2018-2019 PDF: All Yes
FY 2017-2018 PDF: All Yes
FY 2016-2017 PDF: All Yes
FY 2015-2016 PDF: All Yes
FY 2014-2015 PDF: All, A, EAP,E, NE, SCA, WH Yes
FY 2013-2014 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2012-2013 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2011-2012 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2010-2011 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2009-2010 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2008-2009 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2007-2008 PDF: All, A, EAP, E, NE, SCA, WH Yes
FY 2006-2007 HTML: All, A, EAP, E, NE, SCA, WH Yes
FY 2005-2006 PDF: All, A, EAP, E, NE, SA, WH Yes
FY 2004-2005 PDF: All, A, EAP, E, NE, SA, WH Yes
FY 2003-2004 PDF: All, A, EAP, E, NE, NIS SA, WH Yes
FY 2002-2003 PDF: All, A, EAP, E, NE, NIS, SA, WH Yes
FY 2001-2002 PDF: All, A, EAP, E, NE, NIS, SA, WH Yes
FY 2000-2001 PDF: All, A + EAP, E + NE + NIS, SCA + WH No
FY 1999-2000* HTML: All, A 1 2 3, EAP 1 2, E 1 2, NE 1 2 NIS 1, SA 1, WH 1 2 3 3 4 No

Key:

A: Africa

EAP: East Asia and Pacific

E: Europe

NE: Near East

NIS: Newly Independent States

SCA: South Central Asia

SA: South Asia

WH: Western Hemisphere

* Unavailable on official website, available in Internet Archive

4. Notes on specific reports

2020-2021 report

Report home page: https://www.state.gov/reports/foreign-military-training-and-dod-engagement-activities-of-interest-2020-2021/

Date from the 2020-2021 report are included in the dataset. The 2020-2021 report is published on the State Department website in separate PDFs corresponding to the regional groupings commonly in use in this publication.

2019-2020 report

Report home page: https://www.state.gov/reports/foreign-military-training-and-dod-engagement-activities-of-interest-2019-2020/

Data from the 2019-2020 report are included in the dataset. The 2019-2020 report is published on the State Department website as a single PDF file.

Page 101 of the PDF initally published in August 2021 is corrupted and contains an empty table where the conclusion of the year's report of training activities in Angola ends and those in Benin begins. We have removed both Angola and Benin from the dataset until this is fixed.

2018-2019 report

Report home page: https://www.state.gov/reports/foreign-military-training-and-dod-engagement-activities-of-interest-2018-2019/

Data from the 2018-2019 report are included in the dataset. The 2019-2020 report is published on the State Department website as a single PDF file.

2017-2018 report

Report home page: https://www.state.gov/reports/foreign-military-training-and-dod-engagement-activities-of-interest-2017-2018/

Data from the 2017-2018 report are included in the dataset. The 2017-2018 report is available in the 2009-2017 archives of the State Department website, and is published as a single PDF file.

2016-2017 report

Report home page: https://www.state.gov/reports/foreign-military-training-and-dod-engagement-activities-of-interest-2016-2017/

Data from the 2016-2017 report are included in the dataset. The 2016-2017 report is available in the 2009-2017 archives of the State Department website, and is published as a single PDF file. This PDF is in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent.

2015-2016 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2016/index.htm

Data from the 2015-2016 report are included in the dataset. The 2015-2016 report is available in the 2009-2017 archives of the State Department website, and is published as a single PDF file. This PDF is in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent.

2014-2015 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2015/index.htm

Data from the 2014-2015 report are included in the dataset. The 2014-2015 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent.

2013-2014 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2014/index.htm

Data from the 2013-2014 report are included in the dataset. The 2013-2014 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent.

2012-2013 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2013/index.htm

Data from the 2012-2013 report are included in the dataset. The 2012-2013 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent.

2011-2012 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2012/index.htm

Data from the 2011-2012 report are included in the dataset. The 2011-2012 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent. The name of the US unit along with another value called US qty (the number of personnel involved perhaps?) are included in the same field; it only happens in this report, and we have not extracted US qty into the dataset.

2010-2011 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2011/index.htm

Data from the 2010-2011 report are included in the dataset. The 2010-2011 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent. The name of the US unit involved in delivering the activities is not recorded in the 2010-2011 report; we have noted this in the relevant rows.

2009-2010 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2010/index.htm

Data from the 2009-2010 report are included in the dataset. The 2009-2010 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent. The name of the US unit involved in delivering the activities is not recorded in the 2009-2010 report; we have noted this in the relevant rows.

2008-2009 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2009/index.htm

Data from the 2008-2009 report are included in the dataset. The 2006-2009 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are in the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent. The name of the US unit involved in delivering the activities is not recorded in the 2008-2009 report; we have noted this in the relevant rows.

2007-2008 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2008/index.htm

Data from the 2007-2008 report are included in the dataset. The 2007-2008 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. These PDFs are the first of the format that remains until at least the publication of the 2017-2018 report, which at the time of writing is the most recent. The name of the US unit involved in delivering the activities is not recorded in the 2007-2008 report; we have noted this in the relevant rows.

2006-2007 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2007/index.htm

Data from the 2006-2007 report are included in the dataset. The 2006-2007 report is available in the 2009-2017 archives of the State Department website, and is published only as a set of HTML pages. As a result, we wrote a separate parser for this report.

2005-2006 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2006/index.htm

Data from the 2005-2006 report are included in the dataset. The 2005-2006 report is available in the 2009-2017 archives of the State Department websites, and is published as a set of PDF files. The PDF structure is similar to the 2001-2002 report, but this is the last year that this structure is used; the 2006-2007 structure is very different.

2004-2005 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2005/index.htm

Data from the 2004-2005 report are included in the dataset. The 2004-2005 report is available in the 2009-2017 archives of the State Department websites, and is published as a set of PDF files. The PDF structure is similar to the 2001-2002 report.

2003-2004 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2004/index.htm

Data from the 2003-2004 report are included in the daaset. The 2003-2004 report is available in the 2009-2017 archives of the State Department websites, and is published as a set of PDF files. The PDF structure is similar to the 2001-2002 report.

2002-2003 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2003/index.htm

Data from the 2002-2003 report are included in the dataset. The 2002-2003 report is available in the 2009-2017 archives of the State Department website, and is published as a set of PDF files. The PDF structure is similar to the 2001-2002 report.

2001-2002 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2002/index.htm

Data from the 2001-2002 report are included in the dataset. The 2001-2002 report is available in the 2009-2017 archives of the State Department website, and is available as a set of PDF files. There are minor differences in formatting between the different files that constitute the 2001-2002 report, and sometimes rows of data are split across pages causing occasional parser errors that need correcting. Further, because of the PDF structure for this particular report, the training quantity is sometimes included inside the parser output for the location. This can be fixed in bulk at a later data cleaning stage. This PDF structure persists, albeit with a number of small structural and aesthetic changes, until 2005-2006.

2000-2001 report

Report home page: https://2009-2017.state.gov/t/pm/rls/rpt/fmtrpt/2001/index.htm

Data from the 2000-2001 report is not included in the dataset. The 2001-2001 report is available in the 2009-2017 archives of the State Department website, and is available as a PDF. The 2000-2001 report does not contain the name of the US unit and the student unit, the location, start date or end date of the training. Accordingly, we did not parse out this report for inclusion in the dataset. Should we later decide to parse them, note that The PDFs are formatted into two columns which makes parsing trickier than for later releases.

1999-2000 report

Report home page (Internet Archive only): https://web.archive.org/web/20170322193536/https://2009-2017.state.gov/www/global/arms/fmtrain/toc.html

Data from the 1999-2000 report is not included in the dataset.The 1999-2000 report is available in the 2009-2017 archives of the State Department website, and is only avaailable from an official source in HTML format. This report does not contain the name of the US unit and the student unit, the location, start date or end date of the training. Accordingly, we did not parse out this report for inclusion in the dataset.

5. Technical

Approach to extraction and cleaning

All the PDFs contain tabular data, with some variances in content structure and appearance. All the PDFs have been generated directly from another program - likely MS Word - rather than scanned in and processed with OCR.

We sought to extract this data in a replicable, programmatic way rather than cut-and-paste our way through the tables once. The reports are mostly in a tabular structure, ordered by country of the recipient units, and the program through which the Department of State or Department of Defence made the expenditure. The available data fields are mostly the same across the 19 publications released to date. The key differences are in the formatting and layout: there are five different formats across the publication set, but two formats cover the majority of the publications. There are, however, small variations even in the same formats, which presents some challenges to creating an automated extraction process (see below on report structures).

We have taken the below approach to extracting and parsing data from these reports:

  • Establish the page ranges in a PDF or group of PDFs that contain different sections of data (e.g. Africa, Europe) and use that to structure the overall parsing process and its outputs.
  • Convert the PDF files to very verbose XML that pulls out every table item or paragraph into a separate row containing detailed paragraph positioning and font-sizing.
  • Filter out anything that isn't in the tables (including, sadly, page numbering), and use the left-right positioning and font-size to determine what the content is. For example, in the 2017-2018 report, text with a <left> value of 27 and <font> type of 1 is consistently going to contain the country name. In most PDFs, particularly the older ones, this value will vary though the content will be the same.
  • Use the linear document structure, which is naturally ordered into Country, Program, Training item and values (e.g. Course name, US unit) as the basis of a nested XML structure.
  • Lint the XML output to show errors in the structure, and massage it until the linter stops flagging errors. This is mostly a matter of figuring out the logic of opening and closing tags, and finding where the particular PDF might break it. The lint will usually show the line numbers where the problem occurs, and the sort of error, giving us enough to fix it.
  • Use eXtensible Stylesheet Language Transformations (XSLT) to deduplicate training items. In some PDFs, the value for a particular row will go on over one or more columns, meaning that we have orphans. The process so far will capture this and correctly flag what the value is (e.g. location, or student unit). However, we want to merge these items so we can later export them into a flat format. A technique called "Muenchian Grouping" enables us to perform this deduplication at the correct level in the nested structure.
  • Flatten the XML and export the data in a flat, tab-separated format.
  • Post-process to standardize countries, dates and fix other issues that emerged during parsing and may be common to different chunks of data. For example, it makes sense to standardize dates and countries in bulk at this later stage. The variety of ways that a training location can be spelled in 20 years of data is huge, so it makes sense to reveal this and fix them all at once in bulk.

The toolset is created in bash:

  • Basic tools: sed, awk, grep, tr, perl
  • Additional tools: ghostscript, pdftohtml, xmlstartlet, xmllint (and csvkit and xsv to write some helper scripts)

The approach is crude but (at least so far) effective as it dealt with a range of problems thrown up by the formatting of the PDFs. There are better of course better ways to approach this problem, but for now it will do!

Project structure and operations

The scrapers/ folder contains the source files, tools and steps used for each annual publication. In group_and_clean/ there are tools to pull together the raw data extracted from each report, aggregate it, and then clean it up. The final, aggregated and cleaned data is contained in the data/ folder. The publish/ folder contains scripts to convert the data into a sqlite database, which can be published to Heroku using a tool called datasette.

.
├── LICENSE.md
├── README.md
├── data
│   ├── final_fmtrpt_all_20190716.tsv
│   ├── final_fmtrpt_all_20191216.tsv
│   ├── final_fmtrpt_all_20210806.tsv
│   └── final_fmtrpt_all_20220325.tsv
├── group_and_clean
│   ├── 0_get_data_files
│   ├── 1_pre_stack_clean
│   ├── 2_stack_em_up
│   ├── 3_big_clean
│   ├── 4_training_uuids
│   ├── 5_prepub_filters
│   └── 6_append_to_prior
├── publish
│   ├── input
│   └── src
└── scrapers
    ├── fmtrpt_1999_2000
    ├── fmtrpt_2000_2001
    ├── fmtrpt_2001_2002
    ├── fmtrpt_2002_2003
    ├── fmtrpt_2003_2004
    ├── fmtrpt_2004_2005
    ├── fmtrpt_2005_2006
    ├── fmtrpt_2006_2007
    ├── fmtrpt_2007_2008
    ├── fmtrpt_2008_2009
    ├── fmtrpt_2009_2010
    ├── fmtrpt_2010_2011
    ├── fmtrpt_2011_2012
    ├── fmtrpt_2012_2013
    ├── fmtrpt_2013_2014
    ├── fmtrpt_2014_2015
    ├── fmtrpt_2015_2016
    ├── fmtrpt_2016_2017
    ├── fmtrpt_2017_2018
    ├── fmtrpt_2018_2019
    ├── fmtrpt_2019_2020
    └── fmtrpt_2020_2021

35 directories, 6 files

Running the scrapers and getting the data out of the PDFs

Each annual report has its own scraper. In more detail, here's what the completed scraper process looks like for the report of the Fiscal Year 2019-2020:

.
├── input
│   └── FMT_Volume-I_FY2019_2020.pdf
├── notes
│   └── notes_on_2016_2017_fmtrpt.md
├── output
│   ├── 0_pdf_slice
│   │   ├── 2019_2020_Africa_fmtrpt.pdf
│   │   ├── 2019_2020_East_Asia_and_Pacific_fmtrpt.pdf
│   │   ├── 2019_2020_Europe_fmtrpt.pdf
│   │   ├── 2019_2020_Near_East_fmtrpt.pdf
│   │   ├── 2019_2020_South_Central_Asia_fmtrpt.pdf
│   │   └── 2019_2020_Western_Hemisphere_fmtrpt.pdf
│   ├── 1_pdf_xml
│   │   ├── 2019_2020_Africa_fmtrpt.xml
│   │   ├── 2019_2020_East_Asia_and_Pacific_fmtrpt.xml
│   │   ├── 2019_2020_Europe_fmtrpt.xml
│   │   ├── 2019_2020_Near_East_fmtrpt.xml
│   │   ├── 2019_2020_South_Central_Asia_fmtrpt.xml
│   │   └── 2019_2020_Western_Hemisphere_fmtrpt.xml
│   ├── 2_xml_refine
│   │   ├── 2019_2020_Africa_fmtrpt_raw.xml
│   │   ├── 2019_2020_East_Asia_and_Pacific_fmtrpt_raw.xml
│   │   ├── 2019_2020_Europe_fmtrpt_raw.xml
│   │   ├── 2019_2020_Near_East_fmtrpt_raw.xml
│   │   ├── 2019_2020_South_Central_Asia_fmtrpt_raw.xml
│   │   └── 2019_2020_Western_Hemisphere_fmtrpt_raw.xml
│   ├── 3_xml_errors
│   │   ├── errors_2019_2020_Africa_fmtrpt.xml
│   │   ├── errors_2019_2020_East_Asia_and_Pacific_fmtrpt.xml
│   │   ├── errors_2019_2020_Europe_fmtrpt.xml
│   │   ├── errors_2019_2020_Near_East_fmtrpt.xml
│   │   ├── errors_2019_2020_South_Central_Asia_fmtrpt.xml
│   │   └── errors_2019_2020_Western_Hemisphere_fmtrpt.xml
│   ├── 4_xml_dedup
│   │   ├── 2019_2020_Africa_fmtrpt_dedup.xml
│   │   ├── 2019_2020_East_Asia_and_Pacific_fmtrpt_dedup.xml
│   │   ├── 2019_2020_Europe_fmtrpt_dedup.xml
│   │   ├── 2019_2020_Near_East_fmtrpt_dedup.xml
│   │   ├── 2019_2020_South_Central_Asia_fmtrpt_dedup.xml
│   │   └── 2019_2020_Western_Hemisphere_fmtrpt_dedup.xml
│   └── 5_xml_tsv
│       ├── 2019_2020_Africa_fmtrpt.tsv
│       ├── 2019_2020_East_Asia_and_Pacific_fmtrpt.tsv
│       ├── 2019_2020_Europe_fmtrpt.tsv
│       ├── 2019_2020_Near_East_fmtrpt.tsv
│       ├── 2019_2020_South_Central_Asia_fmtrpt.tsv
│       └── 2019_2020_Western_Hemisphere_fmtrpt.tsv
└── src
    ├── analyse_xml.sh
    ├── deduplicate_training_items.xsl
    ├── extractor_2019_2020_all_sections.sh
    └── fmtrpt_fy_2019_2020_sections

10 directories, 42 files

The scraping and parsing script is src/extractor_2019_2020_all_sections.sh, which operates on input/FMT_Volume-I_FY2019_2020.pdf. The input PDF is a set of large tables that spread across multiple pages. These tables also have various headings indicating the region and country of training, and the US security assistance program that has financed the training.

The scraper knows what type of data is in each row and column by looking at the combinations of left and font values. For example, in an XML rendering of the PDF, a typical line looks like this:

<text top="103" left="27" width="278" height="10" font="2"><b>FHA - Fiscal Year 2019 Department of Defense Training </b></text>

All lines in the XML where the left value is "27" and the font is "2" will contain data about the program that finances the training. For example:

$ grep "left=\"27\".*font=\"2\"" output/1_pdf_xml/2019_2020_Africa_fmtrpt.xml|head -n5

<text top="103" left="27" width="278" height="10" font="2"><b>FHA - Fiscal Year 2019 Department of Defense Training </b></text>
<text top="232" left="27" width="275" height="10" font="2"><b>IMET-1 - Fiscal Year 2019 Department of State Training </b></text>
<text top="561" left="27" width="292" height="10" font="2"><b>IMET-D&amp;R - Fiscal Year 2019 Department of State Training </b></text>
<text top="809" left="27" width="264" height="10" font="2"><b>PKO - Fiscal Year 2019 Department of State Training </b></text>
<text top="166" left="27" width="277" height="10" font="2"><b>IMET-X - Fiscal Year 2019 Department of State Training </b></text>

Other values like top and width change, but left and font are contant or have small variations that we we can easily detect and add to the scraper. This works for other values held in the tables, such as the name of the training course, and the US and student units. Run the helper script src/analyze_xml.sh on any of the XML files in output/1_pdf_xml to get the left and font combinations for the column headings of that file, which then get plugged into the early part of the script. You will stll need to keep a look out for small variations in the position of values in each column, as the helper script just looks at column headings. These are often best detected by looking for blanks in the final tsv output: blank values for columns like training:us_unit and training:location are found where the row is a "Fiscal Year Program Total". If you find a blank that is not in these rows, check the raw XML for that value and see whether the left position is different than the one assigned that attribute in the extractor script.

The script then parses these values out and surrounds them with a tag like <training_name> or <program>, building a sort-of-XML, which we then bring into compliance by fixing missing end and start tags.

The file src/fmtrpt_fy_2019_2020_sections tells the scraper which file and page range to associate with data on which region (e.g. "Near East").

It works through a sequence of steps, the output of each is contained in output/:

  • output/0_pdf_slice/ contains PDFs of each region's data, based on the page numbering contained in src/fmtrpt_fy_2019_2020_sections.
  • output/1_pdf_xml/ contains a verbose XML version of the PDF. Use src/analyse_xml.sh on these files to find out the relevant positional values for the different columns of data, which we then plug into the scraper.
  • output/2_xml_refine/ contains a simpler, cleaner XML of just the training items.
  • output/3_xml_errors/ contains the output of XML linting. When this passes, the script will move to the next steps.
  • output/4_xml_dedup/ contains the output of an XSLT operation (contained in src/deduplicate_training_items.xsl)
  • output/5_xml_tsv/ converts the clean, deduplicated XML to a flat, tab-separated, final output.

We've structured it this way to make the production process very transparent. It's crude, but it works on these PDFs!

Grouping and cleaning up the scraper outputs

After running all the scraper scripts and getting .tsv files into the output/5_xml_tsv folders of each scraper, the next step is to aggregate them. The toolset and process to do this is contained in group_and_clean/:

.
├── 0_get_data_files
├── 1_pre_stack_clean
├── 2_stack_em_up
├── 3_big_clean
├── 4_training_uuids
├── 5_prepub_filters
└── 6_append_to_prior

7 directories, 0 files

Starting at group_and_clean/0_get_data_files/ the outputs of each step form the input of the next. The relevant script is in the src directory in each step:

  • 0_get_data_files/ pull all the .tsv files from every output/5_xml_tsv/ in the directory tree. You can also adapt this to pull out files specific to a fiscal year.
  • 1_pre_stack_clean/ escapes all speech marks in all the files (which fixes a quoting/sniffing problem for the next step).
  • 2_stack_em_up/ aggregates all the .tsv files into a single .tsv.
  • 3_big_clean/ moves the aggregated .tsv through a number of transformations (cleaning, standardizing, filtering) to produce a superior output.
  • 4_training_uuids mints a UUID for each row of data, which we can use as a reference for that specific training intervention. This step in the process should only be run on newly acquired data.
  • 5_prepub_filters is where we can remove any problematic data before publishing. For example, the 2019-2020 data has a problem with the Angola data, so we remove it until that is fixed.
  • 6_append_to_prior is where we append the newly acquired and cleaned data to the whole dataset. One of the inputs here must be the final output of the last data update, which is stored in hte data/ folder.

The file dropped in group_and_clean/6_apprend_to_prior/output can be copied to data/ as the final outcome of the scraping and cleaning process.

Publishing the data

publish
├── input
│   └── final_fmtrpt_all_20210806.tsv
└── src
    ├── createdb.sh
    ├── deploy.sh
    ├── metadata.yaml
    └── state-department-data.db # Removed in git repo, as it's quite large

2 directories, 5 files

Convert the data to sqlite (we used sqlite-utils), and use datasette (docs) to publish the data online in a useful and highly functional way. We include the full text search and indexing functions outlined in the documentation. We have used Heroku to serve the data, routing to a subdomain of https://securityforcemonitor.org.

The scripts to do this are contained in the publish/ directory. You'll need to set up heroku locally in order to use them. Note that heroku may not allow you to overwrite an existing application - each time you deploy, it must be to a new application name, after which the CNAME record for trainingdata.securityforcemonitor.org will need updating with a new DNS target.

A note on variations in report structures

Since the first publication in 2000 the report has changed structure and content a number of times. Parsers like the ones we have written use the font size and positioning values to determine what field a particular piece of text should be placed in. These values change in each report format, but the changes can easily be seen inside the XML and changes made to the parser per those values.

The current format ("E"), is the easiest to parse and was implemented for the FY2007-2008 report, and has remained broadly stable since then. Format "C", used between FY2000-2001 and FY 2005-2006 is more tricky.

Report Pattern Format
FY 2020-2021 E PDF
FY 2019-2020 E PDF
FY 2018-2019 E PDF
FY 2017-2018 E PDF
FY 2016-2017 E PDF
FY 2015-2016 E PDF
FY 2014-2015 E PDF
FY 2013-2014 E PDF
FY 2012-2013 E PDF
FY 2011-2012 E PDF
FY 2010-2011 E PDF
FY 2009-2010 E PDF
FY 2008-2009 E PDF
FY 2007-2008 E PDF
FY 2006-2007 D HTML
FY 2005-2006 C PDF
FY 2004-2005 C PDF
FY 2003-2004 C PDF
FY 2002-2003 C PDF
FY 2001-2002 C PDF
FY 2000-2001 B PDF
FY 1999-2000 A HTML

About

A dataset on unclassified training activities for non-US security forces arranged and funded by the United States Department of State and Department of Defence between 2001 and 2021. Sources, scraping, cleaning and publishing toolset included. Search and explore the data at: https://trainingdata.securityforcemonitor.org

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published