Public repo for The Alan Turing Institute's reading group on fundamental AI research.
If you're based at the Turing, follow #robots-in-disguise
on the Turing Slack for the most recent updates.
To see all the slides and reading materials for previous sessions, see the archive.
Note that this originated from the Research Engineering Team's reading group on Transformers.
The group meets every week on Mondays at 11-12. Everyone is welcome to join! If you have any questions email Ryan Chan, Fede Nanni or Giulia Occhini and remember to go through our Code of Conduct before joining.
Please get in touch if you would like to give a talk (either about your research or a topic you think is relevant to the reading group) add suggestions and emoji preferences to the list of proposed topics on HackMD!
Date | Topic | Room | Lead |
---|---|---|---|
04/11/24 | Invited Talk: Ethnographic Approaches to AI Evaluations | Ursula Franklin | Jonas Kgomo |
11/11/24 | Scaling laws of neural networks | David Blackwell | Edmund Dable-Heath |
18/11/24 | Application of foundation models in time series tasks | David Blackwell | Gholamali Aminian |
25/11/24 | TBC | David Blackwell | TBC |
02/12/24 | Can language models play the Wikipedia game? | David Blackwell | Alex Hickey, Jo Knight |
09/12/24 | Mechanistic Interpretability III | David Blackwell | Ryan Chan |
16/12/24 | TBC | David Blackwell | TBC |
This project examines how Language Models can navigate Wikipedia. Which tests their ability to link semantically similar topics in a practical way. We have run experiments with a large variety of sentence embedding and large language models for comparison. We have also seen how the performance varies when transversing Wikipedia in other languages and when navigating between scientific papers, which allows an assessment of the breadth of the model's abilities.