Week 12 - part 2

I saw this tweet by Yan Lecun that LLM can not think abstract in a visual way

https://twitter.com/ylecun/status/1769778223842226286 

What he fails to realize, or maybe exactly what he realised is that you should not make an LLM answer a question directly.


Instead it shuold always be wrapper: "What is the strategy for answering this prompt: {X}, having these tools at your disposal". And then the list of actions that can be done.


Think of it if you got some liquid and was asked what it was. It would be dangourus to just taste it, and give an answer.

Instead we will analyze it - find the pH value and what other tools we can use to find out what it is. And then take that knowledge, and give a result.

So the job of the LLM is to do this analysis and continuously improve the prompt (short term memory) with more and better information. And of course also improve the state for the tools available (long term memory, the programmed world model). The prompt can both contain text, images and videos (and even other formats if needed). So the LLM can generate an image that it then can use to do a better analysis.

On important adavantage is that the AI can also much easier explain how it got to the answer.


To summarize the purpose of the LLM is not to immediatly give a final result, but to take the current prompt, add some information, and then use that again.



Kommentarer

Populære opslag fra denne blog

Older blog posts

Week 7