Week 12 - part 2
I saw this tweet by Yan Lecun that LLM can not think abstract in a visual way https://twitter.com/ylecun/status/1769778223842226286 What he fails to realize, or maybe exactly what he realised is that you should not make an LLM answer a question directly. Instead it shuold always be wrapper: "What is the strategy for answering this prompt: {X}, having these tools at your disposal". And then the list of actions that can be done. Think of it if you got some liquid and was asked what it was. It would be dangourus to just taste it, and give an answer. Instead we will analyze it - find the pH value and what other tools we can use to find out what it is. And then take that knowledge, and give a result. So the job of the LLM is to do this analysis and continuously improve the prompt (short term memory) with more and better information. And of course also improve the state for the tools available (long term memory, the programmed world model). The prompt can both contain text, images...