Skip to content

on demand databases deployment, varuios kinds, adding more as i use them!

Notifications You must be signed in to change notification settings

CanaanGM/databases-infrastructure

Repository files navigation

databases folder that houses all sorts of types'o bases, created like this to make thew host system clean without running servers all the time.

using docker compose untill i learn a better way~!

  • TODO:

  • split this up into a readme for each db

Apps to view the databases:

How to connect ?:

a docker images management interface

connect to : localhost:900 and go from there


Redis

redis-cli -a < password u sat in docker-compose >

  • dotnet:
  "CacheSettings": {
    "ConnectionString": "localhost:6379"
  }

Sql server:

this is handy ~! "Server=localhost;Database=Master;User Id=sa;Password=PASSWORD; dotnet 7, u need to trust the connection "Server=localhost;Database=Master;User Id=sa;Password=PASSWORD; ;Trust Server Certificate=true;


mySql:

"server=localhost:3306;user id=USERNAME;password=PASSWORD;persistsecurityinfo=True"


Postgres:

dotnet: "Server=127.0.0.1;Port=5433;Database=DB <Can be Empty>;User Id=User; Password=Password;"

psycopg2 (python):

    dialect+driver://username:password@host:port/database
    postgresql+psycopg2://admin:admin1234@localhost/fastlms

connect to it while running in docker/docker-compose containers

docker exec -it <CONTAINER NAME> psql -U <USERNAME>

in the container: mongo -> mongosh .

connection string:

mongodb://[username:password@]host1[:port1][,...hostN[:portN]][/[defaultauthdb][?options]]
mongodb://mongodbadmin:admin@localhost:27017   << with auth
mongodb://localhost:27017/                     << without auth
or
docker exec -it < container name/id > /bin/bash

Vault warden (local host)

connection should be "http:localhost:< PORT>"

DataLake local-dev "LocalAzureConnection": "AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;DefaultEndpointsProtocol=http;BlobEndpoint=http://< CONTAINER NAME>:10000/devstoreaccount1;QueueEndpoint=http://< CONTAINER NAME>:10001/devstoreaccount1;TableEndpoint=http://< CONTAINER NAME>:10002/devstoreaccount1;"


Airflow

  • to be added lator

Kafka

  • normal connection thru localhost:9092

  • if u wanna connect thru offsetexplorer do this:

    • in the advanced tap add to bootstrap folders:
broker:9092, localhost:9092

create a user thru WSL

    echo -e "AIRFLOW_UID=$(id -u)" > .env

init thre db

    docker-compose up airflow-init

run

    docker-compose up airflow-init

???

https://airflow.apache.org/docs/apache-airflow/stable/start/docker.html#setting-the-right-airflow-user


RabbitMq

  • username and password if not specified are guest/guest.
  • can be accessed @ localhost:15672 - which is the management dashboard

useful links:

  1. https://docs.microsoft.com/en-us/sql/linux/sql-server-linux-configure-mssql-conf?view=sql-server-ver15#datadir
  2. https://hub.docker.com/r/aerospike/aerospike-server
  3. https://hub.docker.com/_/aerospike
  4. https://docs.aerospike.com/server/operations/install/docker-desktop

for future me: the db files are in g drive <3

About

on demand databases deployment, varuios kinds, adding more as i use them!

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages