A fork of OpenAI Baselines, implementations of reinforcement learning algorithms
-
Updated
May 28, 2020 - Python
A fork of OpenAI Baselines, implementations of reinforcement learning algorithms
It would be great to have instructions on how to train a language model from scratch - not just loading the paper's model.
Evolutionary Algorithm using Python
Hi,
When we try to tokenize the following sentence:
If we use spacy
a = spacy.load('en_core_web_lg')
doc = a("I like the link http://www.idph.iowa.gov/ohds/oral-health-center/coordinator")
list(doc)
We got
[I, like, the, link, http://www.idph.iowa.gov, /, ohds, /, oral, -, health, -, center, /, coordinator]
But if we use the Spacy transformer tokenizer:
A collection of 100+ pre-trained RL agents using Stable Baselines, training and hyperparameter optimization included.
This repository contains model-free deep reinforcement learning algorithms implemented in Pytorch
PyTorch version of Stable Baselines, improved implementations of reinforcement learning algorithms.
An experimentation framework for Reinforcement Learning using OpenAI Gym, Tensorflow, and Keras.
✍🏻gpt2-client: Easy-to-use TensorFlow Wrapper for GPT-2 117M, 345M, 774M, and 1.5B Transformer Models
Implementation of reinforcement learning approach to make a car learn to drive smoothly in minutes
At the moment if the digital twin model does not have the plugins built, we can not communicate them them thus we get the following error,
Timeout communicating with flight control plugin.
It would be helpful if the error message included a message reminding the user to check to make sure the plugin is built, or better yet do the check automatically.
A PyTorch implementation of OpenAI's REPTILE algorithm
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
OpenAI's cartpole env solver.
Mirror of Stable-Baselines: a fork of OpenAI Baselines, implementations of reinforcement learning algorithms
Deep reinforcement learning using an asynchronous advantage actor-critic (A3C) model.
Similarly as we do for other projects (example) we should add a CONTRIBUTING file. It would be very helpful for external contributors to understand how to propose new changes.
AI research environment for the game of Snake
Reproducing MuJoCo benchmarks in a modern, commercial game /physics engine (Unity + PhysX).
openai/gym's popular toolkit for developing and comparing reinforcement learning algorithms port to C#.
OpenAI Gym wrapper for the DeepMind Control Suite
Reinforcement learning in Julia. Solving OpenAI gym.
An open AI environment using selenium + docker for training automated web agents
A collection of pre-trained RL agents using Stable Baselines3, training and hyperparameter optimization included.
A Repository with C++ implementations of Reinforcement Learning Algorithms (Pytorch)
Add a description, image, and links to the openai topic page so that developers can more easily learn about it.
To associate your repository with the openai topic, visit your repo's landing page and select "manage topics."
https://github.com/minimaxir/gpt-2-simple/blob/ca6bc61d958fd4c474af9a412ace27279b88dd90/gpt_2_simple/src/encoder.py#L8
According to the docs the lru_cache decorator uses memozation to speed up a function call when it's called with the same arguments.
But this function doesn't need any arguments...