Difference between revisions of "AlfWorld"

From
Jump to: navigation, search
m
m
Line 14: Line 14:
 
* [https://arxiv.org/abs/2010.03768 AlfWorld: Aligning Text and Embodied Environments for Interactive Learning | M. Shridhar, X. Yuan, M. Côté, Y. Bisk, A. Trischler, M. Hausknecht - arXiv]
 
* [https://arxiv.org/abs/2010.03768 AlfWorld: Aligning Text and Embodied Environments for Interactive Learning | M. Shridhar, X. Yuan, M. Côté, Y. Bisk, A. Trischler, M. Hausknecht - arXiv]
 
* [https://github.com/alfworld/alfworld AlfWorld | GitHub]
 
* [https://github.com/alfworld/alfworld AlfWorld | GitHub]
* [[Embodied AI]]
+
 
 +
* [[ALFRED]]
 
* [[Assistants]]  ... [[Agents]]  ... [[Negotiation]] ... [[LangChain]]
 
* [[Assistants]]  ... [[Agents]]  ... [[Negotiation]] ... [[LangChain]]
 
* [[Singularity]] ... [[Artificial Consciousness / Sentience|Sentience]] ... [[Artificial General Intelligence (AGI)| AGI]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
 
* [[Singularity]] ... [[Artificial Consciousness / Sentience|Sentience]] ... [[Artificial General Intelligence (AGI)| AGI]] ... [[Inside Out - Curious Optimistic Reasoning| Curious Reasoning]] ... [[Emergence]] ... [[Moonshots]] ... [[Explainable / Interpretable AI|Explainable AI]] ...  [[Algorithm Administration#Automated Learning|Automated Learning]]
 +
* [[Embodied AI]]
  
<b>AlfWorld</b>: Aligning Text and Embodied Environments for Interactive Learning. a simulator that enables agents to learn abstract, text based policies in TextWorld (Côté et al., 2018) and then execute goals from the ALFRED benchmark (Shridhar et al., 2020) in a rich visual environment. ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions. In turn, as we demonstrate empirically, this fosters better agent generalization than training only in the visually grounded environment. BUTLER's simple, modular design factors the problem to allow researchers to focus on models for improving every piece of the pipeline (language understanding, planning, navigation, and visual scene understanding). The aligned environments allow agents to reason and learn high-level policies in an abstract space before solving embodied tasks through low-level actuation.
+
<b>AlfWorld</b>: Aligning Text and [[Embodied AI]] Environments for Interactive Learning. a simulator that enables agents to learn abstract, text based policies in TextWorld (Côté et al., 2018) and then execute goals from the [[ALFRED]] benchmark (Shridhar et al., 2020) in a rich visual environment. ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions. In turn, as we demonstrate empirically, this fosters better agent generalization than training only in the visually grounded environment. BUTLER's simple, modular design factors the problem to allow researchers to focus on models for improving every piece of the pipeline (language understanding, planning, navigation, and visual scene understanding). The aligned environments allow agents to reason and learn high-level policies in an abstract space before solving embodied tasks through low-level actuation.
  
  
Line 25: Line 27:
  
 
<youtube>Iu2S1wWJ2T4</youtube>
 
<youtube>Iu2S1wWJ2T4</youtube>
 
= [[ALFRED]] =
 
ALFRED (Action Learning From Realistic Environments and Directives) is a benchmark for learning a mapping from natural language instructions and egocentric vision to sequences of actions for household tasks. It includes long, compositional tasks with non-reversible state changes to shrink the gap between research benchmarks and real-world applications
 
 
<youtube>-YmHT2fSQDo</youtube>
 
<youtube>1XoRLNmXffo</youtube>
 

Revision as of 08:45, 20 May 2023

Youtube ... Quora ...Google search ...Google News ...Bing News

AlfWorld: Aligning Text and Embodied AI Environments for Interactive Learning. a simulator that enables agents to learn abstract, text based policies in TextWorld (Côté et al., 2018) and then execute goals from the ALFRED benchmark (Shridhar et al., 2020) in a rich visual environment. ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions. In turn, as we demonstrate empirically, this fosters better agent generalization than training only in the visually grounded environment. BUTLER's simple, modular design factors the problem to allow researchers to focus on models for improving every piece of the pipeline (language understanding, planning, navigation, and visual scene understanding). The aligned environments allow agents to reason and learn high-level policies in an abstract space before solving embodied tasks through low-level actuation.