Mandana Samiei

I am a PhD candidate in Computer Science at Mila - Quebec AI Institute and Reasoning and Learning Lab at McGill University in Montréal, where I am advised by Prof. Doina Precup and Prof. Blake Richards . I am interested in designing AI systems that are able to learn and adapt continually, instead of mastering specific tasks they are able to learn general skills/features which can be reused given a new task/dataset/environment. I would like to study whether certain optmizaiton strategies or sampling methods can be applied to achieve such generalizations.

Just as humans can continuously adapt, can AI agents achieve similar abilities?

My research centers on exploring how AI agents can efficiently learn abstractions of intricate sensory observations of the world, enabling them to reason, plan, and make decisions. Recently, I find myself more and more interested in understanding the role of structure in designing adaptable agents, which can recieve observations/images and output actions/policy or human readable texual content.

Email     CV     GitHub     Scholar     BlueSky     Twitter     LinkedIn

And, finally I would like to share a quote from Jean Piaget whose research has inspired me: "Scientific knowledge is in perpetual evolution; it finds itself changed from one day to the next".

profile photo

Highlights & News




Research

The Role of Schemas in Reinforcement Learning: Insights and Implications for Generalization

In cognitive psychology, schemas are considered to be “building blocks” of cognition, shaping how people view the world and interact with it. The goal of this paper is to propose a method for learning schemas in RL. We argue that, by representing tasks through schemas, agents can more effectively generalize from past experiences and adapt to new, unseen environments with minimal data.

with Prof.Doina Precup and Prof. Blake Richards

Beyond Multitask learning in Continual Learning

Continual Learning solutions often treat multitask learning as an upper-bound of what the learning process can achieve. This is a natural assumption, given that this objective directly addresses the catastrophic forgetting problem, which has been a central focus in early works. However, depending on the nature of the distributional shift in the data, the multi-task solution is not always optimal for the broader continual learning problem. We draw on principles from online learning to formalize the limitations of multitask objectives.

with Giulia Lanzillotta, Prof. Claire Vernade, and Razvan Pascanu

Testing Causal Hypotheses Through Hierarchical Reinforcement Learning

we propose hierarchical reinforcement learning (HRL) as a key ingredient to building agents that can systematically generate and test hypotheses that enables transferrable learning of the world, and discuss potential implementation strategies.

Presented at Intrinsically Motivated Open-ended Learning - IMOL Workshop at NeurIPS 2024 as a poster.

with Dongyan Lin and Anthony Chen

Towards Efficient Generalization in Continual RL using Episodic Memory

As an awardee of Microsoft Research Research grant, with Prof. Blake Richards, Prof. Guillaume Lajoie, Mehdi Fatemi, Ida Momennejad, Prof. Geoffrey J. Gordon, and John Langford. Gave an invited talk at Microsoft Research Summit 2021. Slides.

Possible Principles for Aligned Structure Learning Agents

This paper offers a roadmap for the development of scalable aligned artificial intelligence (AI) from first principle descriptions of natural intelligence. Read More.

with Lancelot Da Costa and Cristian Dragos-Manta

The Cancer Genome Atlas Program (TCGA) Clinical Benchmark

This paper provide a benchmark to study cancer and gene expression relations. Tasks are a combination of clinical features and cancer study. An example of a task would be predicting gender, age, alcohol document history, family history of stomach cancer, the code of the disease and other clinical attributes for different types of patients based on their gene expressions. Read More. Github.

with Joseph Paul Cohen and Prof. Thomas Fevens

Torchmeta: A meta-learning library for pytorch

We introduce Torchmeta, a library built on top of PyTorch that enables seamless and consistent evaluation of meta-learning algorithms on multiple datasets, by providing data-loaders for most of the standard benchmarks in few-shot classification and regression, with a new meta-dataset abstraction. Read More. Github.

with Tristan Deleu and Prof. Yoshua Bengio

Organizing Committee

collas

D&I chair at Fourth Conference on Lifelong Learning Agents - CoLLAs 2025

Local chair at Second Conference on Lifelong Learning Agents - CoLLAs 2023

Organizer at First Conference on Lifelong Learning Agents - CoLLAs 2022

wiml

Senior organizer with Eda Okur - 19th Women in Machine Learning Workshop NeurIPS 2024

Senior organizer with Caroline Weis - Women in Machine Learning Symposium ICML 2024

Senior organizer - Women in Machine Learning Un-workshop ICML 2023

Organizer - Women in Machine Learning Un-workshop ICML 2020

wiml Organizer at Machine Learning Reproducibility Challenge - MLRC 2023

Teaching

eeml EEML PyTorch and Colab intro, Summer 2024

Taught by Nemanja Rakićević and Matko Bošnjak

mcgill Teaching Assistant: COMP 767 Reinforcement Learning, Winter 2021

Taught by Prof. Doina Precup

Teaching Assistant: COMP 417 Intro to Robotics & Intelligent Systems, Fall 2020

Taught by Prof. Dave Meger

udem Teaching Assistant: IFT 6390 Fundamentals of Machine Learning , Fall 2021

Taught by Prof. Ioannis Mitliagkas and Prof. Guillaume Rabusseau

Reviewing

  • Reinforcement Learning Conference - RLC 2024
  • Conference on Lifelong Learning Agents - CoLLAs 2022 & 2024
  • The International Conference on Learning Representations - ICLR 2023 & 2024
  • Transactions on Machine Learning Research - TMLR 2023
  • Conference on Neural Information Processing Systems - NeurIPS 2020 & 2023
  • Generative Models for Decision Making workshop at ICLR 2024
  • Decision Awareness in Reinforcement Learning (DARL) at ICML 2022



Credit to Jon Barron for the template.
Last Update on Jan 2025.