Skip to content

Blog

Large Reasoning Models and Intelligence

Toward the end of 2024, a new type of language model was released: the Large Reasoning Model, or LRM. Examples include OpenAI's o1, Qwen's QwQ and Deepseek's R1. Unlike traditional Large Language Models (LLMs), these models improve their accuracy by performing test-time compute, generating long reasoning chains before outputting their answer.

In the prior discussion on the measures of intelligence, the intelligence of a system is defined as its skill-acquisition efficiency, when given a set of priors and experience. A more intelligent system would end up with greater skill after undergoing a similar amount of experience as a less intelligent system. In essence, this is the measure of the generalization ability of the system in a particular domain.

This article is an exploration into whether LRMs and the method of generating reasoning chains represent a path toward higher intelligence as defined above.

Book Summary: The Goal

This is a book primarily about plant management/supply chain management, but the lessons and concepts translate well to other fields e.g. software development.

Book Summary: The Courage to be Disliked

I listened to the audiobook version of this several times and here is my summary of it.

My biggest takeaways are:

  • We lack the courage to be disliked, to be normal, to try, and so on. Instead, we manufacture 'life lies' (if only I could stop blushing, if only I did not have a traumatic childhood, etc.)
  • Being disliked by others is necessary to achieve freedom
  • Separation of tasks: do not interfere, or care about, other people's tasks
  • Happiness is the feeling of contribution to the community