팔로우
Fedor Moiseev
Fedor Moiseev
google.com의 이메일 확인됨
제목
인용
인용
연도
Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned
E Voita, D Talbot, F Moiseev, R Sennrich, I Titov
arXiv preprint arXiv:1905.09418, 2019
10402019
Gemini: a family of highly capable multimodal models
G Team, R Anil, S Borgeaud, Y Wu, JB Alayrac, J Yu, R Soricut, ...
arXiv preprint arXiv:2312.11805, 2023
5202023
SKILL: Structured knowledge infusion for large language models
F Moiseev, Z Dong, E Alfonseca, M Jaggi
arXiv preprint arXiv:2205.08184, 2022
612022
Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned. arXiv 2019
E Voita, D Talbot, F Moiseev, R Sennrich, I Titov
arXiv preprint arXiv:1905.09418, 0
22
Analyzing multi-head self-attention: Specialized heads do the heavy lifting, the rest can be pruned. arXiv
E Voita, D Talbot, F Moiseev, R Sennrich, I Titov
arXiv preprint arXiv:1905.09418, 2019
132019
SamToNe: Improving Contrastive Loss for Dual Encoder Retrieval Models with Same Tower Negatives
F Moiseev, GH Abrego, P Dornbach, I Zitouni, E Alfonseca, Z Dong
arXiv preprint arXiv:2306.02516, 2023
22023
현재 시스템이 작동되지 않습니다. 나중에 다시 시도해 주세요.
학술자료 1–6