GithubHelp home page GithubHelp logo

cbfinn / maml_rl Goto Github PK

View Code? Open in Web Editor NEW
611.0 15.0 180.0 7.73 MB

Code for RL experiments in "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks"

License: Other

Python 96.23% Ruby 0.65% Mako 0.20% Shell 0.20% CSS 0.44% JavaScript 1.50% HTML 0.79%

maml_rl's Introduction

Model-Agnostic Meta-Learning

This repo contains code accompaning the paper, Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks (Finn et al., ICML 2017). It includes code for running the few-shot reinforcement learning experiments.

For the experiments in the supervised domain, see this codebase.

Dependencies

This code is based off of the rllab code repository and can be installed in the same way (see below). This codebase is not necessarily backwards compatible with rllab.

The MAML code uses the TensorFlow rllab version, so be sure to install TensorFlow v1.0+.

Usage

Scripts for running the experiments found in the paper are located in maml_examples/.

The pointmass environment is located in maml_examples/ whereas the MuJoCo environments are located in rllab/envs/mujoco/.

Speed of Code

One current limitation of the code is that it is particularly slow. We welcome contributions to speed it up. We expect the biggest speed improvements to come from better parallelization of sampling and meta-learning graph computation.

Contact

To ask questions or report issues, please open an issue on the issues tracker.

rllab

Docs Circle CI License Join the chat at https://gitter.im/rllab/rllab

rllab is a framework for developing and evaluating reinforcement learning algorithms. It includes a wide range of continuous control tasks plus implementations of the following algorithms:

rllab is fully compatible with OpenAI Gym. See here for instructions and examples.

rllab only officially supports Python 3.5+. For an older snapshot of rllab sitting on Python 2, please use the py2 branch.

rllab comes with support for running reinforcement learning experiments on an EC2 cluster, and tools for visualizing the results. See the documentation for details.

The main modules use Theano as the underlying framework, and we have support for TensorFlow under sandbox/rocky/tf.

Documentation

Documentation is available online: https://rllab.readthedocs.org/en/latest/.

Citing rllab

If you use rllab for academic research, you are highly encouraged to cite the following paper:

Credits

rllab was originally developed by Rocky Duan (UC Berkeley / OpenAI), Peter Chen (UC Berkeley), Rein Houthooft (UC Berkeley / OpenAI), John Schulman (UC Berkeley / OpenAI), and Pieter Abbeel (UC Berkeley / OpenAI). The library is continued to be jointly developed by people at OpenAI and UC Berkeley.

Slides

Slides presented at ICML 2016: https://www.dropbox.com/s/rqtpp1jv2jtzxeg/ICML2016_benchmarking_slides.pdf?dl=0

maml_rl's People

Contributors

alexbeloi avatar cbfinn avatar coorsbenjamin avatar dementrock avatar florensacc avatar lchenat avatar neocxi avatar openai-sys-okta-integration avatar paulhendricks avatar singulaire avatar tigerneil avatar zhongwen avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

maml_rl's Issues

file ../vendor/mujoco/libmujoco131 is missing

when I run test_maml_cheetah.py cause OSerror:
OSError: ../vendor/mujoco/libmujoco131.so: cannot open shared object file: No such file or directory
can you append the need part?

thank you!

Cannot set up anaconda environment

Hi,
I'm trying to set up the environment with Anaconda to run the code.
I'm running: conda env create -f environment.yml and I get error:

Warning: you have pip-installed dependencies in your environment file, but you do not list pip itself as one of your conda dependencies.  Conda may not use the correct pip to install your packages, and they may end up in the wrong place.  Please add an explicit pip dependency.  I'm adding one for you, but still nagging you.
Collecting package metadata (repodata.json): done
Solving environment: failed

ResolvePackageNotFound:
  - joblib==0.9.4
  - numpy==1.10.4
  - python==3.5.2

Do you know how to fix that?

ImportError: No module named 'sandbox.rocky.tf.algos.maml_trpo'

Here is my Error:

Traceback (most recent call last):
  File "maml_examples/cluster_maml_trpo_ant.py", line 1, in <module>
    from sandbox.rocky.tf.algos.maml_trpo import MAMLTRPO
ImportError: No module named 'sandbox.rocky.tf.algos.maml_trpo'

I have installed rllab3 , and set PYTHONPATH properly.
then source activate rllab3 , go to maml_rl directory and run python maml_examples/cluster_maml_trpo_ant.py , and this error occurs.
What's wrong? should I git clone maml_rl into some directory under rllab?

MAML for DDPG

Hi,
I'd like to ask few important questions,

  • is it possible to implement MAML with DDPG like other implemented MAML-RL algorithms?
  • if so, what are the right way to do that? and how much we expect the DDPG performance will improve after adding MAML?

regards,

CondaHTTPError: HTTP 000 CONNECTION FAILED for url <https://conda.binstar.org/tlatorre/noarch/repodata.json>

Hi, I run ./scripts/setup_linux.sh and get the following error:

Creating conda environment...
Using Anaconda API: https://api.anaconda.org
Solving environment: failed

CondaHTTPError: HTTP 000 CONNECTION FAILED for url <https://conda.binstar.org/tlatorre/noarch/repodata.json>
Elapsed: -

An HTTP error occurred when trying to retrieve this URL.
HTTP errors are often intermittent, and a simple retry will get you on your way.
SSLError(MaxRetryError('HTTPSConnectionPool(host=\'conda.binstar.org\', port=443): Max retries exceeded with url: /tlatorre/noarch/repodata.json (Caused by SSLError(SSLError("bad handshake: Error([(\'SSL routines\', \'ssl3_get_server_certificate\', \'certificate verify failed\')],)",),))',),)

I have tried many times and the error always occurs. I wonder whether the version you used can not work now ? or other problems.
Thanks.

Bug when init VariantGenerator

Hi, I try to run the code. Last day, I success. But today I meet a strange problem.
I don't modify any code, just use setup_mujoco.sh script from
to setup the mujoco env for running cluster_trpo_cheetah.py:
For bug information see:
tv
It seems that the VG class init with 4 arguments but default init is 2. But I check the class VariantGenerator, it's init with no arguments. So with this bug information I can't locate where the problem is. Have you ever meet this problem.

Installation/replicating point mass experiments

Hi, I've been trying to replicate some of your results and compare against Reptile, for the reinforcement learning case. Specifically, I'm trying to run the point mass experiments.

After installing everything and sourcing the conda env rllab3 and running python maml_examples/maml_trpo_point.py, it complains about Theano and Lasagne not being installed.

From the README, it seems like this was built on the Tensorflow version of rllab. Does it still need Theano or am I doing something wrong?

PS: Thanks for open-sourcing your work!

AttributeError: function 'mjv_makeObjects' not found

Hello,
I am trying to set up Mujoco for one of my projects. I installed all the dependencies and used the appropriate license file. When I am trying to run test_maml_ant.py, its giving me the following error

func = self._FuncPtr((name_or_ordinal, self))
AttributeError: function 'mjv_makeObjects' not found

in the file mjlib.py. Does anyone know if it is the installation issue or am I missing something? The examples provided by the mujoco_py repositories are working without any error though.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.