Loading
Feedback

junjie_li 264

Name

Junjie Li

Badges

0
0
1

Activity

Aug
Sep
Oct
Nov
Dec
Jan
Feb
Mar
Apr
May
Jun
Jul
Aug
Mon
Wed
Fri

Ratings Progression

Loading...

Challenge Categories

Loading...

Challenges Entered

Multi Agent Reinforcement Learning on Trains

Latest submissions

See All
submitted 76695
submitted 76694
graded 76592

Robots that learn to interact with the environment autonomously

Latest submissions

No submissions made in this challenge.
Gold 0
Silver 0
Bronze 1
Trustable
May 16, 2020

Badges


  • May 16, 2020

  • May 16, 2020

  • May 16, 2020
  • Has filled their profile page
    May 16, 2020

  • May 16, 2020
Participant Rating
vrv
MasterScrat
student 249
Participant Rating

NeurIPS 2020: Flatland Challenge

🚑 Addressing Round 1 pain points

7 days ago

(post withdrawn by author, will be automatically deleted in 24 hours unless flagged)

Submit both RL and OR method

7 days ago

What happen if have a OR submission then a RL submission?

Which result will be showed on leaderboard? or both?

🚑 Addressing Round 1 pain points

7 days ago

When is the deadline of Round 1? Within 1 day? @MasterScrat

Team merging deadline

9 days ago

Hi, @MasterScrat, it seems that we cannot invite member to our team now?

🚑 Addressing Round 1 pain points

15 days ago

Hi, @MasterScrat, thanks for the kind reply and explaination.

As there is no other teams(using RL) sharing simialr concerns with me, please moving forward.

🚑 Addressing Round 1 pain points

15 days ago

I may be wrong, but below is my feedback about adding many more evaluation episodes:

  • Currently RL’s complete rate is row even given current env settings. It may narrow the application of RL in order to compete with OR method.

  • It may ask us to focus more on OR method.

As I commented before, I think larger env is good, but it’s better to have much less test cases.

🧞 Pain points in Round 1 and wishes for Round 2?

21 days ago

Thanks for the thread for dicussion.

As a participant who really interested in usng RL to solve this problem, my concerns are:

  • Timing. When we use RL, likely we need to use GPU for inference. Unfortuntately, our GPU utilization should be low as it only serve one or a few states per batch. So I may expect that for larger grid size, RL with GPU is likely to be less efficient than OR method.
  • Diversity of env. When we have 14 different size of grid, it makes our RL training harder. If we further consider different speeds, it may require more effort for deadlock free planning.

My wishes for Round 2 are:

  • Use only a few large test cases(for example, # of test cases <= 10), while keep same overall running time. It may be even better to test with same grid size.
  • Use same speed for different agents. I personally prefer to focus more on RL related things, instead of dealing with dead-lock from different speeds.

I think one of OR’s shortage is that it’s not straightforward to optimize for global reward.
My understanding: RL’s advantage is finding a better solution(combining with OR), but not acting in a shorter time.
If we want to see RL performan better than OR, we should give RL enough time for planning/inference on large grid env. (both 5 min and 5s may not be enough for RL to do planning and inference. )

Number of test cases and video of each submission

About 1 month ago

How many test maps to generate the submission result?

After each submission, there will be a video for this submission. Is the video including all test cases?

Config of simulation environment during training and evaluation

About 2 months ago

As you mentioned, small map size may be better with operations search.

I am not sure if there will be test cases with small map size?

If yes, then we may need to implement an operations search algorithm, along with RL algorithm.

My question is: will you limit the minimal map size? For example, larger than K x K, ensuring that most operations search algorithm can not solve the problem in time limit. So that we can focus more on real large map size.

Config of simulation environment during training and evaluation

About 2 months ago

Thanks @MasterScrat for the quick reply.

I feel much clear with your reply.

Config of simulation environment during training and evaluation

About 2 months ago

Thanks @MasterScrat for the kind reply.

May I know how much difference it may be between round 1 and round 2?


Consider the example with two different settings:

  • when we just need our algorithm to work with map size 150 * 150
  • when we also need our algorithm to work with map size 1500 * 1500

It may be quite different to design a optimal state/algorithm when the problem settings are different.

Conda env creation errors...UPDATED: later EOF error when running evaluator

About 2 months ago

I am using WSL2 with Ubuntu(16.04) and docker.

It works well so far.

For the visualization, I have tried two ways, both work for me:

  1. Install GUI and XServer for WSL2.
    Some links I found helpful:
  1. After getting frames in png format, use the following function to generate a video:
    https://gitlab.aicrowd.com/flatland/flatland/blob/master/flatland/evaluators/aicrowd_helpers.py#L108

Overall, I feel the 2nd method is simpler and I am currently using it for visualization.

Config of simulation environment during training and evaluation

About 2 months ago

For RL to work well, it’s better to have similar configs between the simulation environment of training and evaluation.

To help properly setting up the training environment, can you provide some basic information in the evaluation environment?
For example, the range of the following settings:

  1. width and height of map
  2. num of trains
  3. num of cities
  4. type of city distribution
  5. speed ratio of trains
  6. max rails between cities
  7. max rails in cities
  8. type of schedule generator
  9. malfunction: rate, min/max duration.

Conda env creation errors...UPDATED: later EOF error when running evaluator

About 2 months ago

Same here.

In windows, I can only install the environment via: pip install flatland-rl

However, it failed to run the evaluator, error same as in MemoAI’s post. (… EOFError: Ran out of input)

RE: 🚑 Addressing Round 1 pain points

7 days ago

Hi @MasterScrat, when is the deadline of round 1? 1day or 7 days left?

junjie_li has not provided any information yet.