Skip to content

Network Visualizer for the 'Geschichten aus der Geschichte' Podcast

License

Notifications You must be signed in to change notification settings

Dr-Lego/gag-network

Repository files navigation

Geschichten aus der Geschichte Network Visualizer

This project aims to visualize the intricate web of connections within the German podcast "Geschichten aus der Geschichte" (Stories from History). It creates a network visualization based on the podcast's episode list on Wikipedia, showcasing how the Wikipedia articles of the topics mentioned in the podcast are interconnected.

Project Overview

The GAG Network Visualizer scrapes data from Wikipedia, processes it, and generates a interactive network graph. It reveals interesting connections between historical topics discussed in the podcast episodes.

Key features:

  • Scrapes podcast episode data and related Wikipedia articles
  • Processes and analyzes links between articles
  • Generates a visual network of interconnected topics
  • Provides an interactive web interface to explore the network

Tech Stack

  • Python
  • Web scraping: requests, BeautifulSoup, and Selenium
  • Data processing: pandas and numpy
  • Text parsing: wikitextparser
  • Multiprocessing
  • SQLite database
  • SvelteKit as JavaScript frontend framework
  • Network visualization with vis.js

Performance Optimization

Significant effort has been put into optimizing the performance of this project as processing complete Wikipedia dumps:

  • Wikipedia index file is converted to an sqlite database at the beginning to make it searchable more easily.
  • Utilizes multiprocessing for parallel execution of tasks.
  • Uses database indexing and optimized SQL queries to speed up data retrieval.
  • Pre-loads and compresses network data to reduce initial loading times for the visualization.

Prerequisites

  • Chrome browser for Selenium WebDriver (or change it in build.py)
  • Python packages: pip install -r requirements.txt
  • Node.js tools: sudo npm install terser roadroller -g
  • Initialize frontend: cd frontend && npm install

Downloading Wikipedia Dumps

  1. Visit dumps.wikimedia.org/dewiki and ddumps.wikimedia.org/enwiki
  2. Choose the latest timestamp directory
  3. Download the two topmost files: *-pages-articles-multistream.xml.bz2 and *-pages-articles-multistream-index.txt.bz2

Setting Up Environment Variables

You need to set several environment variables that define the user agent, database path, and paths to the Wikipedia dump files.

GAG_USER_AGENT="GAG-Network (your-email@example.com)"
GAG_DATABASE="/path/to/database"
GAG_WIKIDUMP_DE="/path/to/german/wikipedia-dump.xml.bz2"
GAG_WIKIDUMP_DE_INDEX="/path/to/german/wikipedia-dump-index.txt.bz2" 
GAG_WIKIDUMP_EN="/path/to/english/wikipedia-dump.xml.bz2"
GAG_WIKIDUMP_EN_INDEX="/path/to/english/wikipedia-dump-index.txt.bz2" 

Running the Project

The project's main script (main.py) orchestrates the entire process of data collection, processing, and visualization creation. You can run it in different modes depending on your needs.

Usage

You can run main.py in three different modes:

  1. Full Process

    To run the entire process (refresh data and create save):

    python main.py
    

    This will execute all steps in order:

    • Create the database
    • Transform icons
    • Load the network
    • Build the frontend
  2. Refresh Data Only

    To only refresh the data without creating a new save:

    python main.py --data
    

    This will:

    • Create the database
    • Transform icons
  3. Create Save Only

    To create a new preloaded network save without refreshing data:

    python main.py --preload
    

    This will:

    • Load the network
    • Build the frontend

Output

After running main.py, you will have:

  • A SQLite database with scraped and processed data
  • JavaScript files containing network data and metadata
  • A preloaded network save for faster initial loading

Notes

  • The full process can be time-consuming, especially when processing complete Wikipedia dumps. Ensure you have sufficient computational resources available.
  • If you encounter any issues, check the console output for error messages and ensure all prerequisites are correctly set up.
  • The --data option is useful when you want to update the underlying data without regenerating the visualization.
  • The --preload option is helpful when you've made changes to the visualization code but don't need to refresh the underlying data.

After running the script, you can view the visualization by opening build/index.html in a web browser. An up-to-date build can also be found at Dr-Lego/gag-network-build.