Are LLMs capable of non-verbal reasoning?

Date:

Share:



Large language models have found great success so far by using their transformer architecture to effectively predict the next words (i.e., language tokens) needed to respond to queries. When it comes to complex reasoning tasks that require abstract logic, though, some researchers have found that interpreting everything through this kind of “language space” can start to cause some problems, even for modern “reasoning” models.

Now, researchers are trying to work around these problems by crafting models that can work out potential logical solutions completely in “latent space”—the hidden computational layer just before the transformer generates language. While this approach doesn’t cause a sea change in an LLM’s reasoning capabilities, it does show distinct improvements in accuracy for certain types of logical problems and shows some interesting directions for new research.

Wait, what space?

Modern reasoning models like ChatGPT’s o1 tend to work by generating a “chain of thought.” Each step of the logical process in these models is expressed as a sequence of natural language word tokens that are fed back through the model.

Read full article

Comments



Source link

━ more like this

Xbox’s VR headset with Meta could release sooner than we thought

Xbox has come a long way since its humble beginnings as a chunky console. It's recently taken on the form of an Asus...

Chinese company Netease is making an AAA action-adventure game called ‘Blood Message’

NetEase, the Chinese video game company that published Marvel Rivals and Bungie's Destiny: Rising, has announced its first single-player AAA game. It's a...

Tesla inaugural Robotaxi rides will have a human ‘safety monitor’ on board

A select few will soon get to experience Tesla's robotaxi service for the first time, but they won't be alone in the car....
spot_img