팔로우
Nan Jiang
Nan Jiang
Assistant Professor of Computer Science, UIUC
illinois.edu의 이메일 확인됨 - 홈페이지
제목
인용
인용
연도
Doubly Robust Off-policy Value Evaluation for Reinforcement Learning
N Jiang, L Li
Proceedings of the 33rd International Conference on Machine Learning (ICML-16), 2015
4242015
Contextual Decision Processes with Low Bellman Rank are PAC-Learnable
N Jiang, A Krishnamurthy, A Agarwal, J Langford, RE Schapire
Proceedings of the 34th International Conference on Machine Learning (ICML-17), 2016
2402016
Information-Theoretic Considerations in Batch Reinforcement Learning
J Chen, N Jiang
Proceedings of the 36th International Conference on Machine Learning (ICML …, 2019
1512019
Hierarchical Imitation and Reinforcement Learning
HM Le, N Jiang, A Agarwal, M Dudík, Y Yue, H Daumé III
Proceedings of the 35th International Conference on Machine Learning (ICML-18), 2018
1332018
Provably efficient RL with Rich Observations via Latent State Decoding
SS Du, A Krishnamurthy, N Jiang, A Agarwal, M Dudík, J Langford
Proceedings of the 36th International Conference on Machine Learning (ICML …, 2019
1192019
Model-based RL in Contextual Decision Processes: PAC bounds and Exponential Improvements over Model-free Approaches
W Sun, N Jiang, A Krishnamurthy, A Agarwal, J Langford
Conference on Learning Theory, 2019
118*2019
The Dependence of Effective Planning Horizon on Model Accuracy
N Jiang, A Kulesza, S Singh, R Lewis
Proceedings of the 2015 International Conference on Autonomous Agents and …, 2015
1082015
Minimax Weight and Q-Function Learning for Off-Policy Evaluation
M Uehara, J Huang, N Jiang
arXiv preprint arXiv:1910.12809, 2019
942019
On Oracle-Efficient PAC Reinforcement Learning with Rich Observations
C Dann, N Jiang, A Krishnamurthy, A Agarwal, J Langford, RE Schapire
Advances in Neural Information Processing Systems, 2018, 2018
772018
Sample complexity of reinforcement learning using linearly combined model ensembles
A Modi, N Jiang, A Tewari, S Singh
International Conference on Artificial Intelligence and Statistics, 2010-2020, 2020
702020
Empirical study of off-policy policy evaluation for reinforcement learning
C Voloshin, HM Le, N Jiang, Y Yue
arXiv preprint arXiv:1911.06854, 2019
672019
Abstraction Selection in Model-based Reinforcement Learning
N Jiang, A Kulesza, S Singh
Proceedings of the 32nd International Conference on Machine Learning (ICML …, 2015
612015
Repeated Inverse Reinforcement Learning
K Amin, N Jiang, S Singh
Advances in Neural Information Processing Systems, 2017, 2017
542017
Open Problem: The Dependence of Sample Complexity Lower Bounds on Planning Horizon
N Jiang, A Agarwal
Conference On Learning Theory, 3395-3398, 2018
512018
Reinforcement Learning: Theory and Algorithms
A Agarwal, N Jiang, SM Kakade
502019
Provably efficient q-learning with low switching cost
Y Bai, T Xie, N Jiang, YX Wang
Advances in Neural Information Processing Systems, 8004-8013, 2019
482019
Batch value-function approximation with only realizability
T Xie, N Jiang
International Conference on Machine Learning, 11404-11413, 2021
392021
Bellman-consistent pessimism for offline reinforcement learning
T Xie, CA Cheng, N Jiang, P Mineiro, A Agarwal
Advances in neural information processing systems 34, 6683-6694, 2021
372021
Improving UCT planning via approximate homomorphisms
N Jiang, S Singh, R Lewis
Proceedings of the 2014 international conference on Autonomous agents and …, 2014
362014
Minimax Value Interval for Off-Policy Evaluation and Policy Optimization
N Jiang, J Huang
arXiv preprint arXiv:2002.02081, 2020
342020
현재 시스템이 작동되지 않습니다. 나중에 다시 시도해 주세요.
학술자료 1–20