Discussion about this post

User's avatar
Kevin's avatar

I agree, it is becoming more clear how the LLMs function as “fuzzy lookup systems”. Hallucination is still a problem, too - it does not seem able to admit that it can’t figure out a problem, when it’s a little bit beyond its ability.

Expand full comment
sprachzarathustra's avatar

How does this relate to the thesis of the (I know, unpublished) Diminishing Returns in Machine Learning installment on algorithmic improvements? On one hand o1‘s accuracy scales logarithmically with compute, on the other test time compute is a new type of scaling district from training compute. Will the next installment of Diminishing Returns still be published? I enjoyed the first one.

Expand full comment
5 more comments...

No posts