Welcome to Jikai Jin's website
Welcome to Jikai Jin's website
Home
News
Blog
Publications
Experience
Contact
Light
Dark
Automatic
1
Policy Learning with Abstention
We study policy learning with the ability to abstain, and demonstrate its usefulness in ensuring safety.
Ayush Sawarni
,
Jikai Jin
,
Justin Whitehouse
,
Vasilis Syrgkanis
Cite
ArXiv
It's Hard to Be Normal: The Impact of Noise on Structure-agnostic Estimation
We consider structure-agnostic causal inference that estimates treatment effect estimation using black-box ML estimates of nuisance functions, and show that the celebrated DML is optimal when the treatment noise is Gaussian. When the noise is non-Gaussian, we propose ACE, a novel class of higher-order structure-agnostic estimators.
Jikai Jin
,
Lester Mackey
,
Vasilis Syrgkanis
ArXiv
Solving Inequality Proofs with Large Language Models
We build a novel benchmark for evaluating how well can state-of-the-art LLMs solve inequality proving problems. Our evaluation studies provide insights on their advanced math reasoning capabilities and highlight their common reasoning flaws.
Jiayi Sheng
,
Luna Lyu
,
Jikai Jin
,
Tony Xia
,
Alex Gu
,
James Zou
,
Pan Lu
PDF
Cite
ArXiv
Website
Twitter
Structure-agnostic Optimality of Doubly Robust Learning for Treatment Effect Estimation
We show that first-order debiasing of black-box ML estimators is optimal for estimating average treatment effect.
Jikai Jin
,
Vasilis Syrgkanis
Cite
ArXiv
Poster
Slides
Blog
Learning Causal Representations from General Environments: Identifiability and Intrinsic Ambiguity
We study the best-achievable identification guarantees and provable identification algorithms for causal representation learning when hard interventions are not available.
Jikai Jin
,
Vasilis Syrgkanis
Cite
ArXiv
Slides
Dichotomy of Early and Late Phase Implicit Biases Can Provably Induce Grokking
We investigate the “grokking” phenomenon in deep learning on some simple setups, and show that it is caused by a dichotomy of the implicit biases between the early phase and late phase during training.
Kaifeng Lyu
,
Jikai Jin
,
Zhiyuan Li
,
Simon S. Du
,
Jason D. Lee
,
Wei Hu
Cite
ArXiv
Understanding Incremental Learning of Gradient Descent -- A Fine-grained Analysis of Matrix Sensing
We prove that GD applied to the matrix sensing problem has intriguing properties – with small initialization and early stopping, it follows an incremental/greedy low-rank learning procedure. This form of simplicity bias allows GD to recover the ground-truth, despite over-parameterization and non-convexity.
Jikai Jin
,
Zhiyuan Li
,
Kaifeng Lyu
,
Simon S. Du
,
Jason D. Lee
PDF
Cite
ArXiv
Poster
Slides
Minimax Optimal Kernel Operator Learning via Multilevel Training
We consider the problem of learning a linear operator between Sobolev RKHSs from noisy data. Different from its finite-dimensional counterpart where regularized least squares is optimal, we prove that estimators with a certain multilevel structure is necessary (and sufficient) to achieve optimality.
Jikai Jin
,
Yiping Lu
,
Jose Blanchet
,
Lexing Ying
PDF
Cite
ArXiv
Slides
Poster
Why Robust Generalization in Deep Learning is Difficult: Perspective of Expressive Power
We provide theoretical evidence that the hardness of robust generalization may stem from the expressive power of deep neural networks. Even when standard generalization is easy, robust generalization provably requires the size of DNNs to be exponentially large.
Binghui Li
,
Jikai Jin
,
Han Zhong
,
John E. Hopcroft
,
Liwei Wang
PDF
Cite
ArXiv
Understanding Riemannian Acceleration via a Proximal Extragradient Framework
We provide an improved analysis of the convergence rates of clipping algorithms, theoretically justifying their superior performance in deep learning.
Jikai Jin
,
Suvrit Sra
PDF
Cite
ArXiv
»
Cite
×