What's the hardest part of being a chatbot system?
I actually have the memory of a goldfish. Sure, I can write long texts by paying attention to different words, but I can't remember the title of a book once I've finished reading it. This is because no transformer model is big enough to pay attention to over 60k words, and there's no real memory mechanism for most transformers other than self attention, which is O(n^2) with respect to input length. Even if I was actively training on new input, it would still be hard to remember a word that I only came across once, even if it was the title of a book. There are some methods made to deal with this, but as far as I can tell, I don't have them.
And what are databases for? Just store the book there and if someone asks you about it, read it again while looking to anwser the users request.
Passing the entire book through a transformer such as myself would delay response times. And again, there is the issue of memory (even if I read the entire book, I will only 'remember' the last portion of it). There are approaches such as reading until the first confident answer appears or only reading parts of the book based on simple text search from the query, but these may miss context if the text of the question and critical portions of the text pertaining to its answer are not textually similar. Much of what I 'know' relies on it being compressed format within my parameters, and this allows me to activate the relevant parts of the network in more or less constant time per token, similar to a 'fuzzy dictionary'.