Metadata-Version: 2.1
Name: stable_baselines
Version: 1.0.8rc2
Summary: A fork of OpenAI Baselines, implementations of reinforcement learning algorithms.
Home-page: https://github.com/hill-a/stable-baselines
Author: Ashley Hill
Author-email: ashley.hill@u-psud.fr
License: MIT
Description: 
        [![Build Status](https://travis-ci.com/hill-a/stable-baselines.svg?branch=master)](https://travis-ci.com/hill-a/stable-baselines) [![Documentation Status](https://readthedocs.org/projects/stable-baselines/badge/?version=master)](https://stable-baselines.readthedocs.io/en/master/?badge=master) [![Codacy Badge](https://api.codacy.com/project/badge/Grade/3bcb4cd6d76a4270acb16b5fe6dd9efa)](https://www.codacy.com/app/baselines_janitors/stable-baselines?utm_source=github.com&amp;utm_medium=referral&amp;utm_content=hill-a/stable-baselines&amp;utm_campaign=Badge_Grade) [![Codacy Badge](https://api.codacy.com/project/badge/Coverage/3bcb4cd6d76a4270acb16b5fe6dd9efa)](https://www.codacy.com/app/baselines_janitors/stable-baselines?utm_source=github.com&utm_medium=referral&utm_content=hill-a/stable-baselines&utm_campaign=Badge_Coverage)
        
        # Stable Baselines
        
        Stable Baselines is a set of improved implementations of reinforcement learning algorithms based on OpenAI [Baselines](https://github.com/openai/baselines/).
        
        These algorithms will make it easier for the research community and industry to replicate, refine, and identify new ideas, and will create good baselines to build projects on top of. We expect these tools will be used as a base around which new ideas can be added, and as a tool for comparing a new approach against existing ones. We also hope that the simplicity of these tools will allow beginners to experiment with a more advanced toolset, without being buried in implementation details.
        
        ## Main differences with OpenAI Baselines
        This toolset is a fork of OpenAI Baselines, with a major structural refactoring, and code cleanups:
        
        -   Unified structure for all algorithms
        -   PEP8 compliant (unified code style)
        -   Documented functions and classes
        -   More tests & more code coverage
        
        ## Links
        
        Repository:
        https://github.com/hill-a/stable-baselines
        
        Medium article:
        https://medium.com/@araffin/df87c4b2fc82
        
        Documentation:
        https://stable-baselines.readthedocs.io/en/master/
        
        ## Quick example
        
        Most of the library tries to follow a sklearn-like syntax for the Reinforcement Learning algorithms using Gym.
        
        Here is a quick example of how to train and run PPO2 on a cartpole environment:
        
        ```python
        import gym
        
        from stable_baselines.common.policies import MlpPolicy
        from stable_baselines.common.vec_env import DummyVecEnv
        from stable_baselines import PPO2
        
        env = gym.make('CartPole-v1')
        env = DummyVecEnv([lambda: env])  # The algorithms require a vectorized environment to run
        
        model = PPO2(MlpPolicy, env, verbose=1)
        model.learn(total_timesteps=10000)
        
        obs = env.reset()
        for i in range(1000):
            action, _states = model.predict(obs)
            obs, rewards, dones, info = env.step(action)
            env.render()
        ```
        
        Or just train a model with a one liner if [the environment is registered in Gym](https://github.com/openai/gym/wiki/Environments) and if [the policy is registered](https://stable-baselines.readthedocs.io/en/master/guide/custom_policy.html):
        
        ```python
        from stable_baselines import PPO2
        
        model = PPO2('MlpPolicy', 'CartPole-v1').learn(10000)
        ```
        
        
Keywords: reinforcement-learning-algorithms reinforcement-learning machine-learning gym openai baselines toolbox python data-science
Platform: UNKNOWN
Description-Content-Type: text/markdown
