Skip to content

explore deep learning models generated automatically from relational database catalog info

Notifications You must be signed in to change notification settings

ryanmark1867/dl_auto_catalog

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 

Repository files navigation

dl_auto_catalog

  • repo for relational db scraping related to Manning book Deep Learning with Structured Data https://www.manning.com/books/deep-learning-with-structured-data
  • the code in this repo extracts table metadata from the catalog of a Postgres database so it can be used to automatically train deep learning models with combinations of data in the database

Directory structure

  • data - processed datasets and pickle files for intermediate datasets
  • models - saved trained models
  • notebooks - code
  • pipelines - pickled pipeline files

To exercise the code

  1. Install Postgres https://www.postgresql.org/download/ including pgadmin
  2. follow instructions to create sample db: https://www.postgresqltutorial.com/postgresql-sample-database/
  3. once you have created the sample db, update the config file notebooks/scrape_db_catalog_config.yml to ensure that the user, host, port and database settings match the credentials for your database
  4. run notebooks/scrape_db_catalog.py. This module will:
  • prompt you to enter the Postgres password corresponding with the user you specified in the config file
  • connect to the database using the credentials you specified in the config file and the password you supplied interactively
  • run a query to get details about the columns of every table in the specified schema
  • save the results of the query in a dataframe that gets persisted as a pickle file

Background

About

explore deep learning models generated automatically from relational database catalog info

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages