12.1 C
London
Thursday, 11 December, 2025

An Interview with Prof. Matthew Dixon: A Quant and Rock Star

Dr Paul Bilokon (PB) has interviewed Prof....

Not Yet Broken: RSA in the NISQ Era

An excerpt from “Chapter 1: Why Quantum...

The Origins of High-Frequency Trading

The Long Pursuit of Speed in Finance How...

Reward is enough

AIReward is enough

In a recent paper by David Silver, Satinder Baveja, Doina Precup, and Richard Sutton, the authors hypothesize that the objective of maximizing reward is enough to drive behaviour that exhibits most if not all attributes of intelligence that are studied in natural and artificial intelligence, including knowledge, learning, perception, social intelligence, language, and generalization. This is in contrast to the view that specialized problem formulations are needed for each attribute of intelligence, based on other signals or objectives. The reward-is-enough hypothesis suggests that agents with powerful reinforcement learning algorithms when placed in rich environments with simple rewards could develop the kind of broad, multi-attribute intelligence that constitutes an artificial general intelligence.

Check out our other content

Check out other tags:

Most Popular Articles