Æȷοì
Naomi Saphra
Naomi Saphra
Kempner Institute at Harvard University
fas.harvard.eduÀÇ À̸ÞÀÏ È®ÀÎµÊ - ȨÆäÀÌÁö
Á¦¸ñ
Àοë
Àοë
¿¬µµ
Dynet: The dynamic neural network toolkit
G Neubig, C Dyer, Y Goldberg, A Matthews, W Ammar, A Anastasopoulos, ...
arXiv preprint arXiv:1701.03980, 2017
2782017
Understanding objects in detail with fine-grained attributes
A Vedaldi, S Mahendran, S Tsogkas, S Maji, R Girshick, J Kannala, ...
Proceedings of the IEEE conference on computer vision and pattern ¡¦, 2014
1332014
A taxonomy and review of generalization research in NLP
D Hupkes, M Giulianelli, V Dankers, M Artetxe, Y Elazar, T Pimentel, ...
Nature Machine Intelligence 5 (10), 1161-1174, 2023
121*2023
Understanding learning dynamics of language models with SVCCA
N Saphra, A Lopez
arXiv preprint arXiv:1811.00225, 2018
116*2018
Understanding privacy-related questions on stack overflow
M Tahaei, K Vaniea, N Saphra
Proceedings of the 2020 CHI conference on human factors in computing systems ¡¦, 2020
1012020
The multiberts: Bert reproductions for robustness analysis
T Sellam, S Yadlowsky, J Wei, N Saphra, A D'Amour, T Linzen, J Bastings, ...
arXiv preprint arXiv:2106.16163, 2021
932021
An algerian arabic-french code-switched corpus
R Cotterell, A Renduchintala, N Saphra, C Callison-Burch
Workshop on free/open-source arabic corpora and corpora processing tools ¡¦, 2014
762014
Pareto probing: Trading off accuracy for complexity
T Pimentel, N Saphra, A Williams, R Cotterell
arXiv preprint arXiv:2010.02180, 2020
632020
Linear connectivity reveals generalization strategies
J Juneja, R Bansal, K Cho, J Sedoc, N Saphra
arXiv preprint arXiv:2205.12411, 2022
502022
Sudden drops in the loss: Syntax acquisition, phase transitions, and simplicity bias in MLMs
A Chen, R Shwartz-Ziv, K Cho, ML Leavitt, N Saphra
arXiv preprint arXiv:2309.07311, 2023
382023
A non-linear structural probe
JC White, T Pimentel, N Saphra, R Cotterell
arXiv preprint arXiv:2105.10185, 2021
312021
A framework for (under) specifying dependency syntax without overloading annotators
N Schneider, B O'Connor, N Saphra, D Bamman, M Faruqui, NA Smith, ...
arXiv preprint arXiv:1306.2091, 2013
312013
LSTMs compose (and learn) bottom-up
N Saphra, A Lopez
arXiv preprint arXiv:2010.04650, 2020
19*2020
Benchmarking compositionality with formal languages
J Valvoda, N Saphra, J Rawski, A Williams, R Cotterell
arXiv preprint arXiv:2208.08195, 2022
162022
First tragedy, then parse: History repeats itself in the new era of large language models
N Saphra, E Fleisig, K Cho, A Lopez
arXiv preprint arXiv:2311.05020, 2023
152023
Amrica: an amr inspector for cross-language alignments
N Saphra, A Lopez
Proceedings of the 2015 conference of the north american chapter of the ¡¦, 2015
132015
Transcendence: Generative Models Can Outperform The Experts That Train Them
E Zhang, V Zhu, N Saphra, A Kleiman, BL Edelman, M Tambe, ...
arXiv preprint arXiv:2406.11741, 2024
82024
Latent state models of training dynamics
MY Hu, A Chen, N Saphra, K Cho
arXiv preprint arXiv:2308.09543, 2023
52023
Benchmarks as microscopes: A call for model metrology
M Saxon, A Holtzman, P West, WY Wang, N Saphra
arXiv preprint arXiv:2407.16711, 2024
42024
TRAM: Bridging Trust Regions and Sharpness Aware Minimization
T Sherborne, N Saphra, P Dasigi, H Peng
arXiv preprint arXiv:2310.03646, 2023
42023
ÇöÀç ½Ã½ºÅÛÀÌ ÀÛµ¿µÇÁö ¾Ê½À´Ï´Ù. ³ªÁß¿¡ ´Ù½Ã ½ÃµµÇØ ÁÖ¼¼¿ä.
ÇмúÀÚ·á 1–20