Search Results for author: Yeoneung Kim

Found 5 papers, 0 papers with code

Approximate Thompson Sampling for Learning Linear Quadratic Regulators with $O(\sqrt{T})$ Regret

no code implementations29 May 2024 Yeoneung Kim, Gihun Kim, Insoon Yang

We propose an approximate Thompson sampling algorithm that learns linear quadratic regulators (LQR) with an improved Bayesian regret bound of $O(\sqrt{T})$.

Thompson Sampling

Acceleration of Grokking in Learning Arithmetic Operations via Kolmogorov-Arnold Representation

no code implementations26 May 2024 Yeachan Park, Minseok Kim, Yeoneung Kim

Focusing on the grokking phenomenon that arises in learning arithmetic binary operations via the transformer model, we begin with a discussion on data augmentation in the case of commutative binary operations.

Data Augmentation Decoder +1

On the stability of Lipschitz continuous control problems and its application to reinforcement learning

no code implementations20 Apr 2024 Namkyeong Cho, Yeoneung Kim

We address the crucial yet underexplored stability properties of the Hamilton--Jacobi--Bellman (HJB) equation in model-free reinforcement learning contexts, specifically for Lipschitz continuous optimal control problems.

Continuous Control reinforcement-learning

Improved Regret Analysis for Variance-Adaptive Linear Bandits and Horizon-Free Linear Mixture MDPs

no code implementations5 Nov 2021 Yeoneung Kim, Insoon Yang, Kwang-Sung Jun

For linear bandits, we achieve $\tilde O(\min\{d\sqrt{K}, d^{1. 5}\sqrt{\sum_{k=1}^K \sigma_k^2}\} + d^2)$ where $d$ is the dimension of the features, $K$ is the time horizon, and $\sigma_k^2$ is the noise variance at time step $k$, and $\tilde O$ ignores polylogarithmic dependence, which is a factor of $d^3$ improvement.

LEMMA

Training Wasserstein GANs without gradient penalties

no code implementations27 Oct 2021 Dohyun Kwon, Yeoneung Kim, Guido Montúfar, Insoon Yang

We propose a stable method to train Wasserstein generative adversarial networks.

Cannot find the paper you are looking for? You can Submit a new open access paper.