
Location
Badges
Activity
Challenge Categories
Challenges Entered
The first, open autonomous racing challenge.
Latest submissions
| Participant | Rating |
|---|
| Participant | Rating |
|---|
Learn-to-Race: Autonomous Racing Virtual Challenge
[Call for Papers โ Extended] 2nd Workshop on AI for Autonomous Driving at IJCAI 2022
Over 3 years agoHello all,
We have extended the paper submission deadline for the 2nd Workshop on Artificial Intelligence for Autonomous Driving (AI4AD), co-located with IJCAI-ECAI 2022, to May 20, 2022.
We extend special invitation to all challenge participants!
Workshop website: https://learn-to-race.org/workshop-ai4ad-ijcai2022/
All papers related to autonomous driving are welcome (4-page extended abstracts or 8-page full papers; page count does not include references or appendices), especially those academic manuscripts that describe your research, development, and experiments on the L2R Autonomous Racing Virtual Challenge (L2R-ARVC). We provide a summary of the Challenge, for everyoneโs convenience: https://arxiv.org/pdf/2205.02953.pdf
As the goal is to aggregate all efforts in relevant areas, dual submission is allowed: feel free to submit work-in-progress, work under review, or work already accepted/published elsewhere.
Start a submission: https://cmt3.research.microsoft.com/AI4AD2022
Important dates (all deadlines are in Central European Time (CET), UTC +1, Paris, Brussels, Vienna):
- Paper submissions due:
13 May 202220 May 2022 - Author notification: 3 June 2022
- Workshop: 23 July 2022
Everyone is welcome to attend, in-person and/or online. If you are interested, you can subscribe to our mailing list for updates, here: https://lnkd.in/eBHUfFn
Best Regards,
Organizers
- Jonathan Francis; CMU + Bosch Research
- Xinshuo Weng; CMU + NVIDIA Research
- Hitesh Arora; Amazon
- Siddha Ganju; NVIDIA
- Bingqing Chen; CMU
- Daniel Omeiza; Oxford
- Jean Oh; CMU
- Eric Nyberg; CMU
- Sylvia L. Herbert; UCSD
Call for Papers: Workshop on AI for Autonomous Driving, at the International Joint Conference on Artificial Intelligence (IJCAI 2022)
Almost 4 years agoHello all,
We are happy to announce the 2nd Workshop on Artificial Intelligence for Autonomous Driving (AI4AD), co-located with the International Joint Conference on Artificial Intelligence (IJCAI 2022), to be held in Vienna and online.
Workshop website: https://learn-to-race.org/workshop-ai4ad-ijcai2022/
All papers related to autonomous driving are welcome (4-page extended abstracts or 8-page full papers; page count does not include references or appendices), especially those academic manuscripts that describe your research, development, and experiments on the L2R Autonomous Racing Virtual Challenge.
As the goal is to aggregate all efforts in relevant areas, dual submission is allowed: feel free to submit work-in-progress, work under review, or work already accepted/published elsewhere.
Start a paper submission: https://cmt3.research.microsoft.com/AI4AD2022
Important dates (all deadlines are in Central European Time (CET), UTC +1, Paris, Brussels, Vienna):
- Paper submissions due: 13 May 2022
- Author notification: 3 June 2022
- Workshop: 23 July 2022
Everyone is welcome to attend, in-person and/or online. If you are interested, you can subscribe to our mailing list for updates, here: https://lnkd.in/eBHUfFn
Organizers:
- Jonathan Francis; CMU + Bosch Research
- Xinshuo Weng; CMU + NVIDIA Research
- Hitesh Arora; Amazon
- Siddha Ganju; NVIDIA
- Bingqing Chen; CMU
- Daniel Omeiza; Oxford
- Jean Oh; CMU
- Eric Nyberg; CMU
- Sylvia L. Herbert; UCSD
Updates to timelines
Almost 4 years agoWe will send out a separate notice, before the launch of Stage 2 (see here for more info: [Round 2] Launch - Expected Date - #3 by jon_francis).
Not to worry about the number of submissions left. Once Stage 2 opens, participants will be able to make submissions at a much higher frequency (compared to Stage 1) and we will resume the standard request protocol (๐๏ธ Claim Your Training Credits) for AWS credits.
Clarification on input sensors during evaluation
Almost 4 years agoBottom line: yes, we will allow access to the semseg cameras, during the 1-hour practice period in Stage 2
Regarding Stage 2 evaluation
Almost 4 years agoThanks for the note. Yes, you will be able to upload pre-trained models in Stage 2. Additionally, those models can perform further optimizer updates, through the practice period. More info here: [Round 2] Launch - Expected Date - #3 by jon_francis
[Round 2] Launch - Expected Date
Almost 4 years agoThanks for your continued patience on the launch of Stage 2!
We are working to incorporate participant feedback/concerns, before the Stage 2 launch, to ensure that Stage 2 remains both fair and accommodating. We are again expecting a launch this week, as we are in the final stages of testing various model types and configurations.
There will be no restriction on the submission frequency, subject to the serverโs ability to perform the โpracticeโ phases and main evaluations, in a timely manner. We will adopt the same protocol as before, for allowing teams to request AWS credits: ๐๏ธ Claim Your Training Credits
A brief word about Stage 2: this phase of the competition really tests agentsโ abilities to safely generalise to unseen environments. We encourage participants to optimise their approaches, specifically for this safe generalisation capability, by experimenting offline with the Anglesey track as a target environment. Transfer learning techniques such as domain adaptation, *-shot learning, knowledge distillation, or self-supervision/self-training โ e.g., making use the sensory information that is available during the โpractice phaseโ but will not be available during the main evaluation โ may prove useful; leveraging domain knowledge about the road features may be crucial. Indeed, participants will submit their models for a 1-hour practice period, wherein agents will be free to perform optimizer updates. Afterwards, the resultant checkpoint from the practice phase will be tested on the simulated North Road track at Las Vegas Motor Speedway.
Very much looking forward to it!
Collecting Data through Arrival Simulator
Almost 4 years agoThere are a couple options here:
- Save an agentโs transitions, automatically: agents/sac_agent.py ยท master ยท Learn to Race / l2r-starter-kit ยท GitLab
- Save queue: L82-93
-
record_experienceconfiguration flag: L328-342, L357-359, L479-494
- Record transitions from the simulator, manually; see the following thread for some hints:
Get input actions given directly to simulator for creating Imitation Learning data
Almost 4 years agoYes, the reason why recording transitions from keyboard-based actions is not supported is, indeed, because it bypasses the L2R framework: the python code would then be unable to intercept the commands.
However, we have contacted the developers of the simulator and will continue to look into it!
Recent changes to the StarterKit and Code Documentation
Almost 4 years agoRecent changes to the StarterKit and Code Documentation
The post concerns recent changes and patches made to the starter kit. These patches deal with recent issues that contestants were facing, regarding: stability, metrics calculations, and agent initialisation. Additionally, camera configuration interfaces were optimised for simplicity, and codebase documentation was updated and extended. Some changes included in this patch necessitate re-evaluation of previous submissions, which may affect leaderboard ranking. See below.
Changelog:
- Simplified the camera interface code, for environment/simulator interaction.
- Added additional camera configurations for other sensors that are permitted for use during training.
- Resolved agent initialisation issues, related to yaw ambiguity; this corrects situations where agents re-spawned with incorrect orientation after failing track segments. Previously, this produced spurious results, where agents were assigned incorrect segment-completion metrics, during evaluation. This fix may affect leaderboard results.
- Provided additional agent tracking information, displayed on the console during training and evaluation.
- Revised code documentation, to incorporate recent inquiries:
- Environment description: https://learn-to-race.readthedocs.io/en/latest/env_overview.html
- Sensor configuration: https://learn-to-race.readthedocs.io/en/latest/sensors.html#creating-custom-sensor-configurations
- Getting started: https://learn-to-race.readthedocs.io/en/latest/getting_started.html
- [Edit, 17 Jan 2022 16:26 ET]: Migrated Anglesey track map json file from the official L2R repo to the StarterKit
We hope participants find these changes helpful.
Participants are strongly encouraged to incorporate these changes, as soon as possible. In order to do this, please initiate a merge request, from the upstream repository to your respective forked repositories. https://gitlab.aicrowd.com/learn-to-race/l2r-starter-kit
Need your Inputs for improving competition
Almost 4 years agoThanks โ added to the documentation, along with some other suggestions: https://learn-to-race.readthedocs.io/en/latest/getting_started.html
Increasing the Flow of Time in the Simulator
Almost 4 years agoThanks, weโve reached out to the developers of the simulator.
That said, we do remain excited about seeing approaches that have taken model inference and optimisation latencies into consideration, in the control policy design. These approaches would be particularly viable for simulation-to-real transfer.
Information Available on the Hidden Stage 2 Track
Almost 4 years agoYou are correct that using Las Vegas track information is not permitted.
While the map information has already been omitted from the simulator, itself, we have also removed the json layout, to avoid confusion.
Thanks!
Announcing L2R Challenge white paper
Over 3 years agoWe have posted a new white paper, which summarises the 2022 Learn-to-Race (L2R) Autonomous Racing Virtual Challenge and formally introduces the new L2R Task 2.0 evaluation benchmark that we used in this competition.
https://learn-to-race.org/2022/05/04/whitepaper_l2rarvc/