Skip to content

gtbook/robotics

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Introduction to Robotics and Perception Draft

Copyright (c) 2022-2024 Frank Dellaert and Seth Hutchinson, All rights Reserved

The code in the notebooks and python .py files is covered by the BSD license, like GTSAM; see the LICENSE file for details.

The remainder (including all markdown cells in the notebooks and other prose) is not licensed for any redistribution or change of format or medium, other than making copies of the notebooks or forking this repo for your own private use. No commercial or broadcast use is allowed. We are making these materials freely available to help you learn robotics, so please respect our copyright and these restrictions.

This repository contains one python notebook per section.

deploy-book

Binder

Current Draft TOC (chapters will be viewable as we finish drafts):

Chapter 1: General Introduction

Chapter 2: Discrete: Trash Sorter

  • 2.1 State: plastic|paper|cardboard|metal|glass|unknown
  • 2.2 Actuators: gripper (4 actions or no-op)
  • 2.3 Sensors: Attributes -> colors (5), size (SML), weights (LH)
  • 2.4 Perception: Estimation probabilities of plastic|paper|cardboard|metal|unknown
  • 2.5 Planning: Decision Theory -> sort into three bins (plastics|mixed paper|metals|glass) or do nothing
  • 2.6 Learning: Learning sensor models (discrete)

Chapter 3: Discrete: Vacuum Cleaner

  • 3.1 State: house
  • 3.2 Actuators: noisy graph
  • 3.3 Sensors: noisy light
  • 3.4 Perception: HMM
  • 3.5 Planning: MDP
  • 3.6 Learning: Model-based RL and Policy Iteration

Chapter 4: 2D: Logistics Platform

  • 4.1 State: x, y
  • 4.2 Actuators: Swedish wheels, round robot
  • 4.3 Sensors: 2D LIDAR
  • 4.4 Perception: Localization
  • 4.5 Planning: Graph-based stuff, config space
  • 4.6 Learning: Learning motion models (continuous)

Chapter 5: x, y, theta: Duckiebot

  • 5.1 State: x, y, theta (not fully SE(2) and 3x3 matrices yet)
  • 5.2 Actuators: differential drive
  • 5.3 Sensors: Camera, image formation
  • 5.4 Perception: CNN Place Recognition
  • 5.5 Planning: Dubbins etc, Sampling-based planning
  • 5.6 Deep Learning: Stochastic Gradient descent for CNN

Chapter 6: SE(2): Autonomous Vehicle

  • 6.1 State: SE(2): SO(2) and 3x3 homogeneous
  • 6.2 Actuators: Ackerman steering
  • 6.3 Sensors: 3D LIDAR
  • 6.4 Perception: ICP, Pose-SLAM
  • 6.5 Planning: Motion Primitives
  • 6.6 Learning: Deep Reinforcement Learning (Policy Gradient)

Chapter 7: SE(3): Drone

  • 7.1 State: SE(3): SO(3) and 4x4 homogeneous
  • 7.2 Actuators: Quad-rotor
  • 7.3 Sensors: Camera in 3D, geometry, stereo
  • 7.4 Perception: Visual SLAM
  • 7.5 Planning: Trajectory Optimization
  • 7.6 Learning: Tentative: Safe Flying

About

Notebook-based book "Introduction to Robotics and Perception" by Frank Dellaert and Seth Hutchinson

Topics

Resources

License

Stars

Watchers

Forks

Languages