Seohong Park

[Pronunciation: "suh-hong" ("Seo" as in "Seoul")]

Hey! I'm a Ph.D. student at UC Berkeley advised by Sergey Levine. I'm interested in developing scalable reinforcement learning (RL) algorithms and understanding their mathematical properties. The question that drives me these days is: Is RL really scalable? More specifically, my goal is to make off-policy Q-learning scalable so that we can leverage previously collected data. To this end, I work on data-driven RL, unsupervised RL, and representation learning.

CV | Email | Twitter | Google Scholar | Github


profile photo
Publications
(*: equal contribution)
Flow Q-Learning
OGBench: Benchmarking Offline Goal-Conditioned RL
Unsupervised-to-Online Reinforcement Learning
Is Value Learning Really the Main Bottleneck in Offline RL?
Seohong Park, Kevin Frans, Sergey Levine, Aviral Kumar
NeurIPS 2024
ICML 2024 ARLET Workshop (Oral)
paper | thread | talk (10 min) | blog post
Foundation Policies with Hilbert Representations
Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings
METRA: Scalable Unsupervised RL with Metric-Aware Abstraction
HIQL: Offline Goal-Conditioned RL with Latent States as Actions
Controllability-Aware Unsupervised Skill Discovery
Predictable MDP Abstraction for Unsupervised Model-Based RL
Constrained GPI for Zero-Shot Transfer in Reinforcement Learning
Lipschitz-constrained Unsupervised Skill Discovery
Time Discretization-Invariant Safe Action Repetition for Policy Gradient Methods
Unsupervised Skill Discovery with Bottleneck Option Learning
Talks

Invited Talks

Scalable Data-Driven RL @ Reinforcement Learning Lab, Princeton University (Mar 2025)

Scalable Data-Driven RL @ Robotics and Embodied Artificial Intelligence Lab, Stanford University (Mar 2025)

Toward Scalable Unsupervised RL @ Intel AI Lab (Feb 2024)

Unsupervised Reinforcement Learning @ Reading group on Unsupervised RL, MILA (Apr 2023)

Education

University of California, Berkeley (Aug 2022 - Present)

Ph.D. student in Computer Science

Seoul National University (Mar 2014 - Aug 2022)

B.S. in Computer Science and Engineering
Leave of absence for military service: Sep 2017 - Sep 2020 (3 years)

The University of Tokyo (Sep 2016 - Feb 2017)

Exchange student

Work Experience

Devsisters (Sep 2018 - Sep 2020)

Machine Learning Engineer
Worked as part of the mandatory military service in the Republic of Korea

Ace Project (Sep 2017 - Aug 2018)

Software Engineer
Worked as part of the mandatory military service in the Republic of Korea

Honors and Awards

Scholarships

KFAS Overseas PhD Scholarship (Aug 2022 - Present)

Korea Foundation for Advanced Studies (KFAS)
Full tuition, insurance, and living expenses support for graduate studies

Berkeley Fellowship (Aug 2022 - Aug 2023)

Presidential Science Scholarship (Mar 2014 - Aug 2022)

Korea Student Aid Foundation (KOSAF)
Full tuition and living expenses support for undergraduate studies

Awards

Gold Prize (1st Place in Signal Processing), Samsung Humantech Paper Award (Jan 2022)

Programming Contests (Selected)

2nd Place, ACM-ICPC Asia Daejeon Regional Contest (Nov 2016)

1st Place, Google Code Jam Round 1C (May 2016)

3rd Place, ACM-ICPC Asia Daejeon Regional Contest (Nov 2015)

1st Place, ACM-ICPC Asia Daejeon Regional Preliminary Contest (Oct 2015)

1st Place, Korea Olympiad in Informatics (KOI) (Jul 2012)

Codeforces: polequoll

Services

Reviews

Conferences: ICML (2023, 2024), NeurIPS (2023, 2024), ICLR (2023, 2024, 2025), IROS (2024)

Workshops: ICML Frontiers4LCD (2023), NeurIPS FMDM (2023), ICML ARLET (2024)

Coursework

Graduate Courses, UC Berkeley (Selected)

MATH 202A Introduction to Topology and Analysis

MATH 202B Introduction to Topology and Analysis

MATH 214 Differential Topology

MATH 240 Riemannian Geometry

MATH 250A Groups, Rings, and Fields

MATH 250B Commutative Algebra

MATH C218A/STAT C205A Probability Theory

STAT 210B Theoretical Statistics

CS 229A Information Theory and Coding

EE 227BT Convex Optimization


Template