Martha White
Martha White
ualberta.ca의 이메일 확인됨 - 홈페이지
제목
인용
인용
연도
Off-Policy Actor-Critic
T Degris, M White, RS Sutton
Twenty-Ninth International Conference on Machine Learning, 2012
3542012
An emphatic approach to the problem of off-policy temporal-difference learning
RS Sutton, AR Mahmood, M White
The Journal of Machine Learning Research 17 (1), 2603-2631, 2016
1702016
Convex Multi-view Subspace Learning.
M White, Y Yu, X Zhang, D Schuurmans
Nips, 1682-1690, 2012
1512012
Meta-learning representations for continual learning
K Javed, M White
arXiv preprint arXiv:1905.12588, 2019
1302019
Supervised autoencoders: Improving generalization performance with unsupervised regularizers
L Le, A Patterson, M White
Advances in neural information processing systems 31, 107-117, 2018
1062018
Estimating the class prior and posterior from noisy positives and unlabeled data
S Jain, M White, P Radivojac
Advances in neural information processing systems 29, 2693-2701, 2016
872016
Unifying task specification in reinforcement learning
M White
International Conference on Machine Learning, 2016
592016
Relaxed clipping: A global training method for robust regression and classification
Y Yu, M Yang, L Xu, M White, D Schuurmans
Advances in Neural Information Processing Systems 23, 2011
452011
Nonparametric semi-supervised learning of class proportions
S Jain, M White, MW Trosset, P Radivojac
arXiv preprint arXiv:1601.01944, 2016
442016
An off-policy policy gradient theorem using emphatic weightings
E Imani, E Graves, M White
arXiv preprint arXiv:1811.09013, 2018
412018
Recovering true classifier performance in positive-unlabeled learning
S Jain, M White, P Radivojac
Proceedings of the AAAI Conference on Artificial Intelligence 31 (1), 2017
412017
Convex Sparse Coding, Subspace Learning, and Semi-Supervised Extensions.
X Zhang, Y Yu, M White, R Huang, D Schuurmans
Proceedings of the AAAI Conference on Artificial Intelligence, 2011
392011
Maxmin Q-learning: Controlling the Estimation Bias of Q-learning
Q Lan, Y Pan, A Fyshe, M White
International Conference on Learning Representations, 2020
372020
Organizing experience: a deeper look at replay mechanisms for sample-based planning in continuous state domains
Y Pan, M Zaheer, A White, A Patterson, M White
International Joint Conference on Artificial Intelligence, 2018
352018
Optimal reverse prediction: a unified perspective on supervised, unsupervised and semi-supervised learning
L Xu, M White, D Schuurmans
Proceedings of the 26th International Conference on Machine Learning, 1137-1144, 2009
312009
A greedy approach to adapting the trace parameter for temporal difference learning
M White, A White
International Conference on Autonomous Agents & Multiagent Systems, 557-565, 2016
302016
Investigating practical, linear temporal difference learning
A White, M White
Autonomous Agents and Multiagent Sytems, 2016
292016
Interval Estimation for Reinforcement-Learning Algorithms in Continuous-State Domains
M White, A White
Advances in Neural Information Processing Systems, 2433–2441, 2010
292010
The utility of sparse representations for control in reinforcement learning
V Liu, R Kumaraswamy, L Le, M White
AAAI Conference on Artificial Intelligence, 2019
272019
Emphatic temporal-difference learning
AR Mahmood, H Yu, M White, RS Sutton
European Workshop on Reinforcement Learning, 2015
262015
현재 시스템이 작동되지 않습니다. 나중에 다시 시도해 주세요.
학술자료 1–20