Loading
Warm Up Round: Completed Round 1: 76 days left #neurips #reinforcement_learning #nlp
6869
176
18
92

Problem Statements

NeurIPS 2022 IGLU Challenge - RL Task

Use an RL agent to build a structure with natural language inputs

1473
60
NeurIPS 2022 IGLU Challenge - NLP Task

Language assisted Human - AI Collaboration

1212
32

Important Links

πŸš€ RL Task Starter Kit

πŸš€ NLP Task Starter Kit

πŸš€ NLP Task Official Baseline

πŸ‘₯ Looking for teammates or advice?

πŸ•΅οΈ Introduction

The general goal of the IGLU challenge is to facilitate research in the area of Human-AI collaboration through the natural language. The aim of this edition is to build interactive agents that learn to solve a task while provided with grounded natural language instructions in a collaborative environment. By interactive agent, we mean agents that can follow instructions in natural language and ask for clarification when needed. Ultimately, the agent should be able to quickly adapt newly acquired skills, just like humans do in collaborative interaction with each other. Despite all the recent progress in interactive problem solving, the task of interactive learning is far from solved. To facilitate research in this direction, we present IGLU – a voxel-based collaborative environment and set of tasks to study interactive grounded language understanding and learning.

In the IGLU setup, human and embodied AI agents have to exchange information using language to accomplish a common goal. Specifically, the human – the Architect – gets to see a 3D structure made of colored cubes and has to provide language instructions to the other agent – the Builder – who can place blocks and interact within the environment. The Builder can also ask clarifying questions to the Architect whenever the provided instructions are ambiguous. IGLU is naturally related, but not limited, to two main areas of AI research: Natural Language Understanding and Generation (NLU/G) and Reinforcement Learning (RL). With this challenge, we hope to bring the RL and NLU communities together and work towards a common goal: building language-grounded interactive agents (as demonstrated in the example below).

Top: architect's instruction was clear, builder proceeds with placing block. Bottom: builder asks a clarifying question, then proceeds.

πŸ“‘ Tasks

Understanding the complexity of the challenge, we offer the participants two tracks they can tackle separately.

πŸ‘· RL Task: Building Structures

This task is about following natural language instructions to build a target structure without seeing what it should look like at the end. The RL agent observes the environment from a first-person point-of-view and is able to move around and place different colored blocks within a predefined building zone. Its task is provided as a dialog between an Architect and a Builder. Specifically, the dialog is split into two parts: the context utterances defining blocks placed previously, and target utterances defining the rest of the blocks to be placed. At the end of an episode, the RL agent receives a score reflecting how complete is the built structure compared to the ground truth target structure.

Head over the RL Task - Building Structures challenge for more details and get started!

The best performing solutions for the Building Structures task will be further evaluated with human-in-the-loop, where the developed agents have a chance to interact with actual human users. The ultimate goal is to see how the proposed offline evaluation correlates or does not correlate with human perspectives of the task. The winners will be nominated according to the offline evaluation used on the leaderboard.

πŸ™‹ NLP Task: Asking Clarifying Questions

This task is about determining when and what clarifying questions to ask. Given the instruction from the Architect (e.g., β€œHelp me build a house.”), the Builder needs to decide whether it has sufficient information to carry out that described task or if further clarification is needed. For instance, the Builder might ask β€œWhat material should I use to build the house?” or β€œWhere do you want it?”. The NLP task is formulated independently from learning to interact with the 3D environment. The original instruction and clarification can be used as input for the Builder to guide its progress.

The NLP Task will be released soon! Come back later for updates on this.

πŸ“… Timeline

  • July: Releasing materials: IGLU framework and baselines code.
  • 25th July: The warm-up phase of the competition begins! Participants are invited to start submitting their solutions.
  • 13th August: End of warming up phase! The official competition begins.
  • October: Submission deadline. Submissions are closed and organizers begin the evaluation process.
  • November: Winners are announced and are invited to contribute to the competition writeup.
  • 2nd-3rd of December: Presentation at NeurIPS 2022 (online/virtual).

πŸ† Prizes

The challenge features a Total Cash Prize Pool of $15,000 USD.

This prize pool is divided as follows:

  1. NLP Task
    • 1st place: $4,000 USD
    • 2nd place: $1,500 USD
    • 3st place: $500 USD
  2. RL Task
    • 1st place: $4,000 USD
    • 2nd place: $1,500 USD
    • 3st place: $500 USD
  3. Research prizes: $3,000 USD

Task Winners. For each task, we will evaluate submissions as described in the Evaluation section. The three teams that score highest on this evaluation will receive prizes of $4,000, $1,500, and $500.

Research prizes. We have reserved $3,000 of the prize pool to be given out at the organizers’ discretion to submissions that we think made a particularly interesting or valuable research contribution. If you wish to be considered for a research prize, please include some details on interesting research-relevant results in the README for your submission. We expect to award around 2-5 research prizes in total.

Authorship. In addition to the cash prizes, we will invite the top three teams from both the RL and NLP tasks for authorship summary manuscript at the end of the competition. At our discretion, we may also include honourable mentions for academically interesting approaches. Honourable mentions will be invited to contribute a shorter section to the paper and have their names included inline.

πŸ‘₯ Team

The organizing team:

  • Julia Kiseleva (Microsoft Research)
  • Alexey Skrynnik (MIPT)
  • Artem Zholus (MIPT)
  • Shrestha Mohanty (Microsoft Research)
  • Negar Arabzadeh (University of Waterloo)
  • Marc-Alexandre CΓ΄tΓ© (Microsoft Research)
  • Mohammad Aliannejadi (University of Amsterdam)
  • Milagro Teruel (Microsoft Research)
  • Ziming Li (Amazon Alexa)
  • Mikhail Burtsev (DeepPavlov)
  • Maartje ter Hoeve (University of Amsterdam)
  • Zoya Volovikova (MIPT)
  • Aleksandr Panov (MIPT)
  • Yuxuan Sun (Meta AI)
  • Kavya Srinet (Meta AI)
  • Arthur Szlam (Meta AI)
  • Ahmed Awadallah (Microsoft Research)

The advisory board:

  • Tim RocktΓ€schel (UCL & DeepMind)
  • Julia Hockenmaier (University of Illinois at Urbana-Champaign)
  • Katja Hofmann (Microsoft Research)
  • Bill Dolan (Microsoft Research)
  • Ryen W. White (Microsoft Research)
  • Maarten de Rijke (University of Amsterdam)
  • Oleg Rokhlenko (Amazon Alexa Shopping)
  • Sharada Mohanty (AICrowd)

πŸ‘‰ Similar challenges

If you are interested in embodied agents interacting with Minecraft-like environments, you will be interested in the ongoing MineRL Basalt competition. They offer cutting edge pretrained agents ready to be finetuned!

🀝 Sponsors

Special thanks to our sponsors for their contributions.

πŸ“± Contact

Twitter URL

We encourage the participants to join our Slack workspace for discussions and asking questions.

You can also reach us at info@iglu-contest.net or via the AICrowd discussion forum.

Participants

Getting Started

Leaderboard

01 dipam 10.000
02
12.000
02
12.000
03 negar_arabzadeh 13.000
03 AndreyKrotkikh 13.000