Skip to content
View SoanKim's full-sized avatar
๐Ÿ€
lucky
๐Ÿ€
lucky

Block or report SoanKim

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please donโ€™t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this userโ€™s behavior. Learn more about reporting abuse.

Report abuse
SoanKim/README.md

Hi there, I'm Soan Kim. ๐Ÿ‘‹

AI Researcher & Systems Builder | Ph.D. Candidate in Reinforcement Learning & Neuroscience

I sit at the intersection of Theoretical Reinforcement Learning and Closed-Loop Systems Engineering. My work focuses on building resource-rational agents that can reason, reflect, and adapt in sparse-reward environments.


๐Ÿ”ญ Current Focus: The AI Scientist

I am building metacognitive architecturesโ€”AI systems that don't just solve tasks but monitor their own uncertainty to allocate compute efficiently.

  • Dual-Process RL: Modeling the arbitration between fast heuristics (System 1) and costly planning (System 2).
  • Closed-Loop Pipelines: Architecting real-time fMRI neurofeedback systems that couple biological substrates with autonomous agents.
  • Sparse Reward Learning: Developing "Insight" mechanisms that allow agents to discover structural shortcuts when external feedback is removed.

๐Ÿ› ๏ธ Featured Projects

A modular RL framework for modeling metacognitive control.

  • The Build: A custom Monte Carlo Tree Search (MCTS) engine integrated with a Model-Free heuristic agent.
  • The Tech: Implemented a "Singleton Latch" mechanism for one-shot schema induction and a parallelized Bayesian optimization pipeline (scikit-optimize) for parameter fitting.
  • The Result: Validated against human behavioral data ($N=49$), proving that "Insight" acts as a structural gatekeeper for resource-rational control.

End-to-end Deep RL agent for abstract reasoning.

  • The Build: Architected a custom OpenAI Gym-compatible environment for the card game SET.
  • The Tech: Trained a Deep Q-Network (DQN) to navigate high-dimensional state spaces, optimizing for optimal solution paths.

๐Ÿ”„ Real-Time Decoded Neurofeedback (Closed Source/Lab)

Real-time signal processing pipeline for causal neural modulation.

  • The Build: Implemented a closed-loop system that decodes fMRI patterns in real-time (MVPA) to induce reward prediction errors in human subjects.
  • The Impact: Demonstrated a causal link between induced neural states and improved problem-solving performance.

๐Ÿ’ป Tech Stack

  • Languages: Python Bash MATLAB
  • ML & RL: PyTorch Scikit-learn OpenAI Gym
  • Tools: Git Linux

Pinned Loading

  1. MetaReasoningHumanExp MetaReasoningHumanExp Public

    โš ๏ธ DEPRECATED: This repository is the legacy research code. For the active, modular framework used in current papers, please see Sparse-Reward-RL-Metacontroller.

    Python 1