[Submitted on 16 Feb 2024 (v1), last revised 19 Jul 2024 (this version, v7)]
Abstract:It has long been hypothesised that causal reasoning plays a fundamental role in robust and general intelligence. However, it is not known if agents must learn causal models in order to generalise to new domains, or if other inductive biases are sufficient. We answer this question, showing that any agent capable of satisfying a regret bound under a large set of distributional shifts must have learned an approximate causal model of the data generating process, which converges to the true causal model for optimal agents. We discuss the implications of this result for several research areas including transfer learning and causal inference.Submission history
 From: Jonathan Richens [view email]      
            [v1]
        Fri, 16 Feb 2024 18:29:19 UTC (2,663 KB)
            [v2]
        Fri, 23 Feb 2024 10:50:13 UTC (2,637 KB)
            [v3]
        Mon, 26 Feb 2024 17:05:33 UTC (2,637 KB)
            [v4]
        Thu, 14 Mar 2024 21:30:42 UTC (2,636 KB)
            [v5]
        Tue, 9 Apr 2024 12:33:53 UTC (2,637 KB)
            [v6]
        Mon, 15 Apr 2024 11:34:52 UTC (2,637 KB)
    [v7]
        Fri, 19 Jul 2024 11:12:08 UTC (2,872 KB)
.png)
  

