{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":637670891,"defaultBranch":"main","name":"pg-bulk-ingest","ownerLogin":"uktrade","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-05-08T07:00:51.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/15107524?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1717163740.0","currentOid":""},"activityList":{"items":[{"before":"b72453464673897eeb0ee2fd83ab98bf89c873bb","after":null,"ref":"refs/heads/docs/correct-dataset-uri","pushedAt":"2024-05-31T13:55:40.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"}},{"before":"a8648f6d9f1e0d1789f42d2e704409099ee89fb5","after":"672b3c2f3795dc836e18a7a7ce71b90b4b077126","ref":"refs/heads/main","pushedAt":"2024-05-31T13:55:39.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"Merge pull request #220 from uktrade/docs/correct-dataset-uri\n\ndocs: use correct URI format for dataset-aware scheduling","shortMessageHtmlLink":"Merge pull request #220 from uktrade/docs/correct-dataset-uri"}},{"before":null,"after":"b72453464673897eeb0ee2fd83ab98bf89c873bb","ref":"refs/heads/docs/correct-dataset-uri","pushedAt":"2024-05-31T13:54:37.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"docs: use correct URI format for dataset-aware scheduling\n\nThe URI for postgresql:// datasets now seems to have a stricter format:\n\npostgresql://host/database/schema/table\n\nBut it does allow the database to be empty, so going for that to keep it\nshorter, and still applicable to most cases that run a single database on the\nhost.","shortMessageHtmlLink":"docs: use correct URI format for dataset-aware scheduling"}},{"before":"a927012057738095ca83697b5c74980d72214ee5","after":null,"ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-30T08:13:32.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"}},{"before":"fd74b2ab7cb45865c7154d99cb01488e14ad381e","after":"a8648f6d9f1e0d1789f42d2e704409099ee89fb5","ref":"refs/heads/main","pushedAt":"2024-05-30T08:13:31.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"Merge pull request #219 from uktrade/feat/adding-login-during-ingest\n\n feat: logging during ingest","shortMessageHtmlLink":"Merge pull request #219 from uktrade/feat/adding-login-during-ingest"}},{"before":"556a57cc6d5cb096d07ee8b6170a3526bceedfdc","after":"a927012057738095ca83697b5c74980d72214ee5","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-30T08:10:30.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"feat: logging during ingest\n\nIn some cases people think pg-bulk-ingest is either doing nothing during\ningest, or loading all the data into memory. Adding in a bit more logging to\nmake it clearer data is going into the database, including interleaved with\nfetching data from source (due to the magic of iterables/generators).\n\nThe complex logic on when to log is an attempt to make the logging suitable for\nboth small and large numbers of rows without per ingest configuration, and\nkeeping the noise down in the logs. It logs every 100 rows initially, going up\nto logging every million rows later in the ingest. For example it could log:\n\nIngesting from source into the database...\nIngested 100 rows...\nIngested 200 rows...\nIngested 300 rows...\nIngested 400 rows...\nIngested 500 rows...\nIngested 600 rows...\nIngested 700 rows...\nIngested 800 rows...\nIngested 900 rows...\nIngested 1000 rows...\nIngested 2000 rows...\nIngested 3000 rows...\nIngested 4000 rows...\nIngested 5000 rows...\nIngested 6000 rows...\nIngested 7000 rows...\nIngested 8000 rows...\nIngested 9000 rows...\nIngested 10000 rows...\nIngested 20000 rows...\nIngested 30000 rows...\nIngested 40000 rows...\nIngested 50000 rows...\nIngested 60000 rows...\nIngested 70000 rows...\nIngested 80000 rows...\nIngested 90000 rows...\nIngested 100000 rows...\nIngested 200000 rows...\nIngested 300000 rows...\nIngested 400000 rows...\nIngested 500000 rows...\nIngested 600000 rows...\nIngested 700000 rows...\nIngested 800000 rows...\nIngested 900000 rows...\nIngested 1000000 rows...\nIngested 2000000 rows...\nIngested 3000000 rows...\nIngested 4000000 rows...\nIngested 5000000 rows...\nIngested 6000000 rows...\nIngested 7000000 rows...\nIngested 8000000 rows...\nIngested 9000000 rows...\nIngested 10000000 rows...\nIngested 11000000 rows...\nIngested 12000000 rows...\nIngested 12312312 rows in total\n\nCo-authored-by: Tash Boyse <57753415+nboyse@users.noreply.github.com>\nCo-authored-by: Michal Charemza \nCo-authored-by: Mohizur Khan \nCo-authored-by: Josh Wong <166488409+joshwong-cs@users.noreply.github.com>","shortMessageHtmlLink":"feat: logging during ingest"}},{"before":"92e9ab7b347becd9c664c03a43301c3457f4315d","after":"556a57cc6d5cb096d07ee8b6170a3526bceedfdc","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-30T08:07:02.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"feat: logging during ingest\n\nIn some cases people think pg-bulk-ingest is either doing nothing during\ningest, or loading all the data into memory. Adding in a bit more logging to\nmake it clearer data is going into the database, including interleaved with\nfetching data from source (due to the magic of iterables/generators).\n\nThe complex logic on when to log is an attempt to make the logging suitable for\nboth small and large numbers of rows without per ingest configuration, and\nkeeping the noise down in the logs. It logs every 100 rows initially, going up\nto logging every million rows later in the ingest. For example it could log:\n\nIngesting from source into the database...\nIngested 100 rows...\nIngested 200 rows...\nIngested 300 rows...\nIngested 400 rows...\nIngested 500 rows...\nIngested 600 rows...\nIngested 700 rows...\nIngested 800 rows...\nIngested 900 rows...\nIngested 1000 rows...\nIngested 2000 rows...\nIngested 3000 rows...\nIngested 4000 rows...\nIngested 5000 rows...\nIngested 6000 rows...\nIngested 7000 rows...\nIngested 8000 rows...\nIngested 9000 rows...\nIngested 10000 rows...\nIngested 20000 rows...\nIngested 30000 rows...\nIngested 40000 rows...\nIngested 50000 rows...\nIngested 60000 rows...\nIngested 70000 rows...\nIngested 80000 rows...\nIngested 90000 rows...\nIngested 100000 rows...\nIngested 200000 rows...\nIngested 300000 rows...\nIngested 400000 rows...\nIngested 500000 rows...\nIngested 600000 rows...\nIngested 700000 rows...\nIngested 800000 rows...\nIngested 900000 rows...\nIngested 1000000 rows...\nIngested 2000000 rows...\nIngested 3000000 rows...\nIngested 4000000 rows...\nIngested 5000000 rows...\nIngested 6000000 rows...\nIngested 7000000 rows...\nIngested 8000000 rows...\nIngested 9000000 rows...\nIngested 10000000 rows...\nIngested 11000000 rows...\nIngested 12000000 rows...\nIngested 12312312 rows in total\n\nCo-authored-by: Tash Boyse <57753415+nboyse@users.noreply.github.com>\nCo-authored-by: Michal Charemza \nCo-authored-by: Mohizur Khan \nCo-authored-by: Josh Wong <166488409+joshwong-cs@users.noreply.github.com>","shortMessageHtmlLink":"feat: logging during ingest"}},{"before":"2608380236b80abd8fe5b532e3c74d18e34121e5","after":"92e9ab7b347becd9c664c03a43301c3457f4315d","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-30T08:05:30.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"feat: logging during ingest\n\nIn some cases people thing pg-bulk-ingest is either doing nothing during\ningest, or loading all the data into memory. Adding in a bit more logging to\nmake it clearer data is going into the database, including interleaved with\nfetching data from source (due to the magic of iterables/generators).\n\nThe complex logic on when to log is an attempt to make the logging suitable for\nboth small and large numbers of rows without per ingest configuration, and\nkeeping the noise down in the logs. It logs every 100 rows initially, going up\nto logging every million rows later in the ingest. For example it could log:\n\nIngesting from source into the database...\nIngested 100 rows...\nIngested 200 rows...\nIngested 300 rows...\nIngested 400 rows...\nIngested 500 rows...\nIngested 600 rows...\nIngested 700 rows...\nIngested 800 rows...\nIngested 900 rows...\nIngested 1000 rows...\nIngested 2000 rows...\nIngested 3000 rows...\nIngested 4000 rows...\nIngested 5000 rows...\nIngested 6000 rows...\nIngested 7000 rows...\nIngested 8000 rows...\nIngested 9000 rows...\nIngested 10000 rows...\nIngested 20000 rows...\nIngested 30000 rows...\nIngested 40000 rows...\nIngested 50000 rows...\nIngested 60000 rows...\nIngested 70000 rows...\nIngested 80000 rows...\nIngested 90000 rows...\nIngested 100000 rows...\nIngested 200000 rows...\nIngested 300000 rows...\nIngested 400000 rows...\nIngested 500000 rows...\nIngested 600000 rows...\nIngested 700000 rows...\nIngested 800000 rows...\nIngested 900000 rows...\nIngested 1000000 rows...\nIngested 2000000 rows...\nIngested 3000000 rows...\nIngested 4000000 rows...\nIngested 5000000 rows...\nIngested 6000000 rows...\nIngested 7000000 rows...\nIngested 8000000 rows...\nIngested 9000000 rows...\nIngested 10000000 rows...\nIngested 11000000 rows...\nIngested 12000000 rows...\nIngested 12312312 rows in total\n\nCo-authored-by: Tash Boyse <57753415+nboyse@users.noreply.github.com>\nCo-authored-by: Michal Charemza \nCo-authored-by: Mohizur Khan \nCo-authored-by: Josh Wong <166488409+joshwong-cs@users.noreply.github.com>","shortMessageHtmlLink":"feat: logging during ingest"}},{"before":"b51e6c9d80779244f45ed0ef210bf395fa426e58","after":"2608380236b80abd8fe5b532e3c74d18e34121e5","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-30T08:03:34.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"feat: logging during ingest\n\nIn some cases people thing pg-bulk-ingest is either doing nothing during\ningest, or loading all the data into memory. Adding in a bit more logging to\nmake it clearer data is going into the database, including interleaved with\nfetching data from source (due to the magic of iterables/generators).\n\nThe complex logic on when to log is an attempt to make the logging suitable for\nboth small and large numbers of rows without per ingest configuration, and\nkeeping the noise down in the logs. It logs every 100 rows initially, going up\nto logging every million rows later in the ingest. For example it could log:\n\nIngesting from source into the database...\nIngested 100 rows...\nIngested 200 rows...\nIngested 300 rows...\nIngested 400 rows...\nIngested 500 rows...\nIngested 600 rows...\nIngested 700 rows...\nIngested 800 rows...\nIngested 900 rows...\nIngested 1000 rows...\nIngested 2000 rows...\nIngested 3000 rows...\nIngested 4000 rows...\nIngested 5000 rows...\nIngested 6000 rows...\nIngested 7000 rows...\nIngested 8000 rows...\nIngested 9000 rows...\nIngested 10000 rows...\nIngested 20000 rows...\nIngested 30000 rows...\nIngested 40000 rows...\nIngested 50000 rows...\nIngested 60000 rows...\nIngested 70000 rows...\nIngested 80000 rows...\nIngested 90000 rows...\nIngested 100000 rows...\nIngested 200000 rows...\nIngested 300000 rows...\nIngested 400000 rows...\nIngested 500000 rows...\nIngested 600000 rows...\nIngested 700000 rows...\nIngested 800000 rows...\nIngested 900000 rows...\nIngested 1000000 rows...\nIngested 2000000 rows...\nIngested 3000000 rows...\nIngested 4000000 rows...\nIngested 5000000 rows...\nIngested 6000000 rows...\nIngested 7000000 rows...\nIngested 8000000 rows...\nIngested 9000000 rows...\nIngested 10000000 rows...\nIngested 11000000 rows...\nIngested 12000000 rows...\nIngested 12312312 rows in total\n\nCo-authored-by: Tash Boyse <57753415+nboyse@users.noreply.github.com>\nCo-authored-by: Michal Charemza \nCo-authored-by: Mohizur Khan \nCo-authored-by: Josh Wong <166488409+joshwong-cs@users.noreply.github.com>","shortMessageHtmlLink":"feat: logging during ingest"}},{"before":"7637b64e630062acd9f6a3dfb7bc81f4e2a68513","after":"b51e6c9d80779244f45ed0ef210bf395fa426e58","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-29T16:12:28.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"Mohizurkhan","name":"Mohizur Khan","path":"/Mohizurkhan","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/147046629?s=80&v=4"},"commit":{"message":"TEMP","shortMessageHtmlLink":"TEMP"}},{"before":"c3a27b2d237e0ef47b3734e26e3f5890bf244361","after":"7637b64e630062acd9f6a3dfb7bc81f4e2a68513","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-29T16:01:34.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"Mohizurkhan","name":"Mohizur Khan","path":"/Mohizurkhan","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/147046629?s=80&v=4"},"commit":{"message":"TEMP","shortMessageHtmlLink":"TEMP"}},{"before":"f0c9775f662c6b5281fceb2fd248b092b8d47792","after":"c3a27b2d237e0ef47b3734e26e3f5890bf244361","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-29T15:46:14.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"Mohizurkhan","name":"Mohizur Khan","path":"/Mohizurkhan","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/147046629?s=80&v=4"},"commit":{"message":"TEMP","shortMessageHtmlLink":"TEMP"}},{"before":null,"after":"f0c9775f662c6b5281fceb2fd248b092b8d47792","ref":"refs/heads/feat/adding-login-during-ingest","pushedAt":"2024-05-29T15:42:00.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"Mohizurkhan","name":"Mohizur Khan","path":"/Mohizurkhan","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/147046629?s=80&v=4"},"commit":{"message":"feat: Adding in login during ingest","shortMessageHtmlLink":"feat: Adding in login during ingest"}},{"before":"8dc50efa4d62ed19389cec2f65b53b624dbb4bb6","after":null,"ref":"refs/heads/dependabot/npm_and_yarn/pug-3.0.3","pushedAt":"2024-05-29T09:21:24.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"nboyse","name":"Tash Boyse","path":"/nboyse","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57753415?s=80&v=4"}},{"before":"37c13ee64c7e003711184fcbb370ee7f97faf593","after":"fd74b2ab7cb45865c7154d99cb01488e14ad381e","ref":"refs/heads/main","pushedAt":"2024-05-29T09:21:24.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"nboyse","name":"Tash Boyse","path":"/nboyse","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/57753415?s=80&v=4"},"commit":{"message":"Merge pull request #218 from uktrade/dependabot/npm_and_yarn/pug-3.0.3\n\nbuild(deps): bump pug from 3.0.2 to 3.0.3","shortMessageHtmlLink":"Merge pull request #218 from uktrade/dependabot/npm_and_yarn/pug-3.0.3"}},{"before":null,"after":"8dc50efa4d62ed19389cec2f65b53b624dbb4bb6","ref":"refs/heads/dependabot/npm_and_yarn/pug-3.0.3","pushedAt":"2024-05-29T09:18:40.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"dependabot[bot]","name":null,"path":"/apps/dependabot","primaryAvatarUrl":"https://avatars.githubusercontent.com/in/29110?s=80&v=4"},"commit":{"message":"build(deps): bump pug from 3.0.2 to 3.0.3\n\nBumps [pug](https://github.com/pugjs/pug) from 3.0.2 to 3.0.3.\n- [Release notes](https://github.com/pugjs/pug/releases)\n- [Commits](https://github.com/pugjs/pug/compare/pug@3.0.2...pug@3.0.3)\n\n---\nupdated-dependencies:\n- dependency-name: pug\n dependency-type: indirect\n...\n\nSigned-off-by: dependabot[bot] ","shortMessageHtmlLink":"build(deps): bump pug from 3.0.2 to 3.0.3"}},{"before":null,"after":"33c3ed37e648ca3e45665501ec0abbea39366c00","ref":"refs/heads/fix/float-field-no-migration","pushedAt":"2024-05-28T14:39:24.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"fix: float field does not cause a migration","shortMessageHtmlLink":"fix: float field does not cause a migration"}},{"before":"34920fdd511c440cea4cefa90b936d1149b0524f","after":"37c13ee64c7e003711184fcbb370ee7f97faf593","ref":"refs/heads/main","pushedAt":"2024-05-23T14:08:40.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"joshwong-cs","name":"Josh Wong","path":"/joshwong-cs","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/166488409?s=80&v=4"},"commit":{"message":"Add failsafe against pushing example db connection into prod (#216)\n\n* soft copy pasta failsafe","shortMessageHtmlLink":"Add failsafe against pushing example db connection into prod (#216)"}},{"before":"05d9f482e15918912e745df6550bc363054f75d6","after":null,"ref":"refs/heads/feat/example-conn","pushedAt":"2024-05-23T14:08:40.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"joshwong-cs","name":"Josh Wong","path":"/joshwong-cs","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/166488409?s=80&v=4"}},{"before":"95e4dd3dcb444d6eedefc67d0e281c9675937a00","after":"05d9f482e15918912e745df6550bc363054f75d6","ref":"refs/heads/feat/example-conn","pushedAt":"2024-05-23T14:03:51.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"joshwong-cs","name":"Josh Wong","path":"/joshwong-cs","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/166488409?s=80&v=4"},"commit":{"message":"soft failsafe","shortMessageHtmlLink":"soft failsafe"}},{"before":null,"after":"95e4dd3dcb444d6eedefc67d0e281c9675937a00","ref":"refs/heads/feat/example-conn","pushedAt":"2024-05-23T13:35:37.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"joshwong-cs","name":"Josh Wong","path":"/joshwong-cs","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/166488409?s=80&v=4"},"commit":{"message":"copy pasta failsafe","shortMessageHtmlLink":"copy pasta failsafe"}},{"before":"e297988602221c7b54dc4bc2ee9184dfc9dcf6a9","after":null,"ref":"refs/heads/build/bump-docs-packages","pushedAt":"2024-05-08T05:12:20.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"}},{"before":"e91434f961a118bb534e69febd2528491cf6a900","after":"34920fdd511c440cea4cefa90b936d1149b0524f","ref":"refs/heads/main","pushedAt":"2024-05-08T05:12:19.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"Merge pull request #215 from uktrade/build/bump-docs-packages\n\nbuild(deps): simplify package.json","shortMessageHtmlLink":"Merge pull request #215 from uktrade/build/bump-docs-packages"}},{"before":null,"after":"e297988602221c7b54dc4bc2ee9184dfc9dcf6a9","ref":"refs/heads/build/bump-docs-packages","pushedAt":"2024-05-08T05:11:44.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"build(deps): simplify package.json\n\nAnd update packages. Triggered by other docs sites taking many hours to do `npm\nci`. It didn't on this one for some reason, but doing this for consistency.","shortMessageHtmlLink":"build(deps): simplify package.json"}},{"before":"ac85faa6ba580ab0b25cc7fae28cfa679030eb51","after":null,"ref":"refs/heads/dependabot/npm_and_yarn/ejs-3.1.10","pushedAt":"2024-05-07T16:04:07.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"}},{"before":"3d9a74ca0eb9f5f03a4a20f06f2662236ebeec98","after":"e91434f961a118bb534e69febd2528491cf6a900","ref":"refs/heads/main","pushedAt":"2024-05-07T16:04:06.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"Merge pull request #214 from uktrade/dependabot/npm_and_yarn/ejs-3.1.10\n\nbuild(deps): bump ejs from 3.1.9 to 3.1.10","shortMessageHtmlLink":"Merge pull request #214 from uktrade/dependabot/npm_and_yarn/ejs-3.1.10"}},{"before":null,"after":"ac85faa6ba580ab0b25cc7fae28cfa679030eb51","ref":"refs/heads/dependabot/npm_and_yarn/ejs-3.1.10","pushedAt":"2024-05-02T13:47:14.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"dependabot[bot]","name":null,"path":"/apps/dependabot","primaryAvatarUrl":"https://avatars.githubusercontent.com/in/29110?s=80&v=4"},"commit":{"message":"build(deps): bump ejs from 3.1.9 to 3.1.10\n\nBumps [ejs](https://github.com/mde/ejs) from 3.1.9 to 3.1.10.\n- [Release notes](https://github.com/mde/ejs/releases)\n- [Commits](https://github.com/mde/ejs/compare/v3.1.9...v3.1.10)\n\n---\nupdated-dependencies:\n- dependency-name: ejs\n dependency-type: indirect\n...\n\nSigned-off-by: dependabot[bot] ","shortMessageHtmlLink":"build(deps): bump ejs from 3.1.9 to 3.1.10"}},{"before":"4d3a2f16c158fd49609a0ad02dc9fac864199e9a","after":null,"ref":"refs/heads/perf/combine-grants","pushedAt":"2024-03-19T18:24:57.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"}},{"before":"26d5ba9aadf830b61c11c486b218afdf683f23b1","after":"3d9a74ca0eb9f5f03a4a20f06f2662236ebeec98","ref":"refs/heads/main","pushedAt":"2024-03-19T18:24:56.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"Merge pull request #213 from uktrade/perf/combine-grants\n\nperf: GRANT all SELECT privileges in a single query","shortMessageHtmlLink":"Merge pull request #213 from uktrade/perf/combine-grants"}},{"before":"8dbf1a05c6b694df26ea7a9fc0223b833441caea","after":"4d3a2f16c158fd49609a0ad02dc9fac864199e9a","ref":"refs/heads/perf/combine-grants","pushedAt":"2024-03-19T18:21:20.000Z","pushType":"force_push","commitsCount":0,"pusher":{"login":"michalc","name":"Michal Charemza","path":"/michalc","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/13877?s=80&v=4"},"commit":{"message":"perf: GRANT all SELECT privileges in a single query\n\nThis reduces the number of queries run when copying existing SELECT privileges,\nwhich happens at the end of the first batch in the cases when the ingest is not\ndirectly into the live table.\n\nThe existing test that privileges are preserved is extended to make sure the\n`','.join(...` behaviour is correct","shortMessageHtmlLink":"perf: GRANT all SELECT privileges in a single query"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEWRTn-gA","startCursor":null,"endCursor":null}},"title":"Activity ยท uktrade/pg-bulk-ingest"}