Seohong Park
[Pronunciation: "suh-hong" ("Seo" as in "Seoul")]
Hey! I'm a Ph.D. student at UC Berkeley advised by Sergey Levine.
I'm interested in developing scalable reinforcement learning (RL) algorithms and understanding their mathematical properties.
The question that drives me these days is: Is RL really scalable?
More specifically, my goal is to make off-policy Q-learning scalable so that we can leverage previously collected data.
To this end, I work on data-driven RL, unsupervised RL, and representation learning.
CV |
Email |
Twitter |
Google Scholar |
Github
|
|
Publications (*: equal contribution)
|
Flow Q-Learning
|
OGBench: Benchmarking Offline Goal-Conditioned RL
|
Unsupervised-to-Online Reinforcement Learning
|
Is Value Learning Really the Main Bottleneck in Offline RL?
|
Foundation Policies with Hilbert Representations
|
Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings
|
METRA: Scalable Unsupervised RL with Metric-Aware Abstraction
|
HIQL: Offline Goal-Conditioned RL with Latent States as Actions
|
Controllability-Aware Unsupervised Skill Discovery
|
Predictable MDP Abstraction for Unsupervised Model-Based RL
|
Constrained GPI for Zero-Shot Transfer in Reinforcement Learning
|
Lipschitz-constrained Unsupervised Skill Discovery
|
Time Discretization-Invariant Safe Action Repetition for Policy Gradient Methods
|
Unsupervised Skill Discovery with Bottleneck Option Learning
|
Invited Talks
|
Scalable Data-Driven RL @ Reinforcement Learning Lab, Princeton University (Mar 2025)
Scalable Data-Driven RL @ Robotics and Embodied Artificial Intelligence Lab, Stanford University (Mar 2025)
Toward Scalable Unsupervised RL @ Intel AI Lab (Feb 2024)
Unsupervised Reinforcement Learning @ Reading group on Unsupervised RL, MILA (Apr 2023)
|
University of California, Berkeley
(Aug 2022 - Present)
Ph.D. student in Computer Science
Seoul National University
(Mar 2014 - Aug 2022)
B.S. in Computer Science and Engineering
Leave of absence for military service: Sep 2017 - Sep 2020 (3 years)
The University of Tokyo
(Sep 2016 - Feb 2017)
Exchange student
|
Devsisters
(Sep 2018 - Sep 2020)
Machine Learning Engineer Worked as part of the mandatory military service in the
Republic of Korea
Ace Project
(Sep 2017 - Aug 2018)
Software Engineer Worked as part of the mandatory military service in the Republic of
Korea
|
Scholarships
|
KFAS Overseas PhD Scholarship (Aug 2022 - Present)
Korea Foundation for Advanced Studies (KFAS)
Full tuition, insurance, and living expenses support for graduate studies
|
Berkeley Fellowship (Aug 2022 - Aug 2023)
|
Presidential Science Scholarship (Mar 2014 - Aug 2022)
Korea Student Aid Foundation (KOSAF)
Full tuition and living expenses support for undergraduate studies
|
Awards
|
Gold Prize (1st Place in Signal Processing), Samsung Humantech Paper Award (Jan 2022)
|
Programming Contests (Selected)
|
2nd Place, ACM-ICPC Asia Daejeon Regional Contest (Nov 2016)
1st Place, Google Code Jam Round 1C (May 2016)
3rd Place, ACM-ICPC Asia Daejeon Regional Contest (Nov 2015)
1st Place, ACM-ICPC Asia Daejeon Regional Preliminary Contest (Oct 2015)
1st Place, Korea Olympiad in Informatics (KOI) (Jul 2012)
Codeforces: polequoll
|
Reviews
|
Conferences:
ICML (2023, 2024),
NeurIPS (2023, 2024),
ICLR (2023, 2024, 2025),
IROS (2024)
Workshops:
ICML Frontiers4LCD (2023),
NeurIPS FMDM (2023),
ICML ARLET (2024)
|
Graduate Courses, UC Berkeley (Selected)
|
MATH 202A Introduction to Topology and Analysis
MATH 202B Introduction to Topology and Analysis
MATH 214 Differential Topology
MATH 240 Riemannian Geometry
MATH 250A Groups, Rings, and Fields
MATH 250B Commutative Algebra
MATH C218A/STAT C205A Probability Theory
STAT 210B Theoretical Statistics
CS 229A Information Theory and Coding
EE 227BT Convex Optimization
|
|