Skip to content

AssistiveRoboticsUNH/bc_tutorial

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Imitation Learning Hello World (dev in progress)

Imitation learning is supervised learning where data comes as expert demonstration. The expert can be a human or any other agent. Input data is referred to as "state" and output data as "action." In discrete action spaces, it resembles classification; in continuous action spaces, it is regression.

Policy $\pi: S \rightarrow A$ is the function/model that takes a state as input and outputs an action. The goal of imitation learning is to learn a policy that mimics the expert's behavior.

Behavioral Cloning (BC) is offline imitation learning that use only the collected demonstrations and doesn't use simulator during learning.

  • This tutorial is educational purpose, so code isn't optimized for production but easy to understand.
  • Each policy training is done in a single jupyter notebook.
  • Each directory contain a readme file.

Installation

    pip install gym==0.26.2
    pip install readchar
    pip install imageio
    pip install -U scikit-learn

Demos

Videos Task State Space Action Space Expert Colab
MountainCar-v0 Continuous(2) Discrete(3) Human toadd
Pendulum-v1 Continuous(3) Continuous(1) RL toadd
CarRacing-v2 Image(96x96x3) Continuous(3) Human toadd
Ant-v3 Continuous(111) Continuous(8) RL toadd
Lift Low-dim(19) Continuous(7) Human toadd

Data format

  • We will use hdf5 file for robomimic (see the 'readme.md' in robomimic directory to understand the data format) and real robot.
  • For rest of the environment we will store as *.pkl file with the following structure.

*.pkl structure we are going to use.

About

Getting Started in Imitation Learning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published