Ray: [rllib] PPO Pytorch support?

Created on 12 Dec 2019  路  8Comments  路  Source: ray-project/ray

What is your question?

Does RLLib support APPO with PyTorch models? The docs seem to hint that it does, but this Open Issue and the attached PR's left me confused:
https://github.com/ray-project/ray/issues/3365

Moreover do Impala and APE-X have Pytorch support? Can these algorithms be used with memory/LSTM models?

question rllib

Most helpful comment

We are currently working on making pytorch a first class citizen across all of RLlib.

  • Some work needs to be done on unifying interfaces/factories/etc.. (I'm currently working on it).
  • Then we'll go through all agents and remove duplicate/almost-duplicate code (tf vs pytorch versions) as much as possible to be left with a clean backend-independent Agent-design.

All 8 comments

There is an implementation of PG in Pytorch. I'm working on implementing PPO.

@toanngosy overall is PyTorch kind of a second class citizen here? Should I just rewrite my stuff in TF? I'd like to get access to the level of distributed training offered by RLlib without haven't to move to tensorflow

We are currently working on making pytorch a first class citizen across all of RLlib.

  • Some work needs to be done on unifying interfaces/factories/etc.. (I'm currently working on it).
  • Then we'll go through all agents and remove duplicate/almost-duplicate code (tf vs pytorch versions) as much as possible to be left with a clean backend-independent Agent-design.

@sven1977 What is the timeline looking like for that?

(simple) PPO is done and PR-submitted. APPO to follow. Not sure, may still take a few days or even weeks.

@sven1977 awesome! does it have LSTM support?

@sven1977 Following up here re: LSTM support.

I'm closing this issue.

Was this page helpful?
0 / 5 - 0 ratings