Recent Results Towards Reasoning
In this presentation, I will go over a few recent topics towards understanding the limits of reasoning capabilities of large language models. I will start with an analysis of the hardness of problems like syllogisms when tackling them with the Transformer architecture; I will propose a metric to measure that hardness, and an approach to make it easier for LLMs to reduce such hardness. I will show that these limitations are not specific to the textual domain and also exist for other domains like visual tasks. Finally, I will discuss recent results in better measuring the capabilities of LLMs for math problems.

