Breaking the spurious link: How causal models fix offline reinforcement learning's generalization problem
phys.org
Researchers from Nanjing University and Carnegie Mellon University have introduced an AI approach that improves how machines learn from past data—a process known as offline reinforcement learning. This type of machine learning is essential for allowing systems to make decisions using only historical information without needing real-time interaction with the world.
By focusing on the authentic cause-and-effect relationships within the data, the new method enables autonomous systems—like driverless cars and medical decision-support systems—to make safer and more reliable choices. The work is published in the journal Frontiers of Computer Science.
From misleading signals to true causality: A new learning paradigm
Traditionally, offline reinforcement learning has struggled because it sometimes picks up misleading patterns from biased historical data. To illustrate, imagine learning how to ...
Copyright of this story solely belongs to phys.org . To see the full text click HERE