THE UNIVERSITY of EDINBURGH

DEGREE REGULATIONS & PROGRAMMES OF STUDY 2025/2026

Timetable information in the Course Catalogue may be subject to change.

University Homepage
DRPS Homepage
DRPS Search
DRPS Contact
DRPS : Course Catalogue : School of Informatics : Informatics

Undergraduate Course: Robot and Reinforcement Learning (INFR11285)

Course Outline
SchoolSchool of Informatics CollegeCollege of Science and Engineering
Credit level (Normal year taken)SCQF Level 11 (Year 4 Undergraduate) AvailabilityAvailable to all students
SCQF Credits20 ECTS Credits10
SummaryThis course is about enabling agents (e.g., robots, software systems) to plan and act in complex domains. Such agents are increasingly being deployed in applications such as healthcare, self-driving cars, and warehouse automation, but they still need considerable prior knowledge and/or human supervision. This course will explore the challenges and opportunities that arise when modern machine learning methods are leveraged to create adaptive and continually improving agents.

In particular, we will explore reinforcement learning methods that solve sequential decision making problems and are being used in many application domains. We will focus on the development and use of such learning algorithms in the challenging context of problems in robotics and cyber-physical systems that are characterised by continuous state and action spaces, under-actuation, partial observability, data sparsity, and the need for safe behaviour. No prior knowledge of robotics is assumed but students will be expected to engage with the robotics problems used to motivate the design of learning algorithms.
Course description The course will begin with an introduction to sequential decision making and optimal control as the foundation to planning and acting without learning. We will discuss the fundamentals of reinforcement learning, focusing on dynamic programming and Monte Carlo methods for sequential action selection. We will introduce the development of a model and its importance to (robot) planning and control algorithms. We will cover methods for adaptive control and learning a dynamics model from data, exploring how these models enable model-based reinforcement learning. We will also discuss model-free reinforcement learning methods (e.g., temporal difference methods) and their use when we are unable to model aspects of the environment. In addition, we will discuss how to bootstrap learning via demonstration and how physical simulation may aid exploration.

The field of robotics adapts quickly to the latest developments in AI. By grounding the learning methods in the context of robot decision-making and control, students will be exposed to the latest research, and they will critically analyse, implement, and evaluate modern learning algorithms in a physically-realistic simulation environment.

List of topics may include:
1. Introducing to sequential decision making
2. Fundamentals of reinforcement learning
3. Sequential action selection: Dynamic Programming, Monte Carlo
4. Adaptive control and dynamics model learning
5. Reinforcement learning applied to robotics: model-based, model-free
6. Imitation learning and function approximation
7. Actor-critic and gradient-based optimisation
8. Offline robot learning and the role of simulation
9. Robot foundation models
10. Training and evaluating performance
Entry Requirements (not applicable to Visiting Students)
Pre-requisites Co-requisites
Prohibited Combinations Students MUST NOT also be taking Robot and Reinforcement Learning (UG) (INFR11290)
Other requirements MSc students must register for this course, while Undergraduate students must register for INFR11290 instead.

Students are expected to be proficient in mathematical concepts, particularly:
Probability theory and statistics.
Linear algebra and geometry (e.g., vectors, matrices, rotations, trigonometry).
Calculus (e.g., integration, differentiation)

Students are also expected to be proficient in a programming language, ideally python, and familiar with working in Linux OS systems.
Information for Visiting Students
Pre-requisitesMSc students must register for this course, while Undergraduate students must register for INFR11290 instead.

Students are expected to be proficient in mathematical concepts, particularly:
Probability theory and statistics.
Linear algebra and geometry (e.g., vectors, matrices, rotations, trigonometry).
Calculus (e.g., integration, differentiation)

Students are also expected to be proficient in a programming language, ideally python, and familiar with working in Linux OS systems.
Course Delivery Information
Academic year 2025/26, Available to all students (SV1) Quota:  None
Course Start Semester 2
Timetable Timetable
Learning and Teaching activities (Further Info) Total Hours: 200 ( Lecture Hours 26, Seminar/Tutorial Hours 2, Summative Assessment Hours 2, Revision Session Hours 2, Programme Level Learning and Teaching Hours 4, Directed Learning and Independent Learning Hours 164 )
Assessment (Further Info) Written Exam 70 %, Coursework 30 %, Practical Exam 0 %
Additional Information (Assessment) Written Exam __70%
Coursework __30%
Feedback Example problem sets and revision lectures will be provided to help with exam preparation.
Exam Information
Exam Diet Paper Name Minutes
Main Exam Diet S2 (April/May)Robot and Reinforcement Learning (INFR11285)120
Learning Outcomes
On completion of this course, the student will be able to:
  1. explain key concepts of sequential decision making, reinforcement learning, and robot planning and control.
  2. analyze and evaluate decision-making problems in robot control, and identify those to which learning methods can be applied.
  3. implement and analyse a subset of established learning algorithms for dynamics learning and stochastic control.
  4. explain limitations of current learning and optimisation algorithms in the context of robot control problems.
Reading List
(i) Reinforcement Learning: An Introduction (second edition). R. Sutton and A. Barto. MIT Press, 2018.
(ii) Algorithms for Reinforcement Learning. C. Szepesvari. Morgan and Claypool Publishers, 2010.
Additional Information
Graduate Attributes and Skills Problem solving, critical/analytical thinking, handling ambiguity, knowledge integration, planning and organizing, independent learning, creativity, written communication.
KeywordsFFL
Contacts
Course organiserDr Michael Mistry
Tel: (0131 6)50 2937
Email: mmistry@exseed.ed.ac.uk
Course secretaryMiss Toni Noble
Tel: (0131 6)50 2692
Email: Toni.noble@ed.ac.uk
Navigation
Help & Information
Home
Introduction
Glossary
Search DPTs and Courses
Regulations
Regulations
Degree Programmes
Introduction
Browse DPTs
Courses
Introduction
Humanities and Social Science
Science and Engineering
Medicine and Veterinary Medicine
Other Information
Combined Course Timetable
Prospectuses
Important Information