עקוב אחר
Aidan Gomez
Aidan Gomez
Cohere
כתובת אימייל מאומתת בדומיין cohere.ai - דף הבית
כותרת
צוטט על ידי
צוטט על ידי
שנה
Attention is all you need (arXiv: 1706.03762). arXiv
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ...
162035*2017
Tensor2tensor for neural machine translation
A Vaswani, S Bengio, E Brevdo, F Chollet, AN Gomez, S Gouws, L Jones, ...
arXiv preprint arXiv:1803.07416, 2018
6622018
The reversible residual network: Backpropagation without storing activations
AN Gomez, M Ren, R Urtasun, RB Grosse
Advances in neural information processing systems 30, 2017
6372017
Disease variant prediction with deep generative models of evolutionary data
J Frazer, P Notin, M Dias, A Gomez, JK Min, K Brock, Y Gal, DS Marks
Nature 599 (7883), 91-95, 2021
5862021
One model to learn them all
L Kaiser, AN Gomez, N Shazeer, A Vaswani, N Parmar, L Jones, ...
arXiv preprint arXiv:1706.05137, 2017
4042017
Depthwise Separable Convolutions for Neural Machine Translation
L Kaiser, AN Gomez, F Chollet
International Conference on Learning Representations, 2018
3952018
Attention is all you need. arXiv. org
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ...
arXiv preprint arXiv:1706.03762, 2017
238*2017
Ł. Kaiser, Polosukhin I (2017) Attention is all you need
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez
Advances in neural information processing systems, 5998-6008, 0
232
Tranception: protein fitness prediction with autoregressive transformers and inference-time retrieval
P Notin, M Dias, J Frazer, J Marchena-Hurtado, AN Gomez, D Marks, ...
International Conference on Machine Learning, 16990-17017, 2022
1932022
A systematic comparison of bayesian deep learning robustness in diabetic retinopathy tasks
A Filos, S Farquhar, AN Gomez, TGJ Rudner, Z Kenton, L Smith, ...
arXiv preprint arXiv:1912.10481, 2019
151*2019
Prioritized training on points that are learnable, worth learning, and not yet learnt
S Mindermann, JM Brauner, MT Razzak, M Sharma, A Kirsch, W Xu, ...
International Conference on Machine Learning, 15630-15649, 2022
1422022
Self-attention between datapoints: Going beyond individual input-output pairs in deep learning
J Kossen, N Band, C Lyle, AN Gomez, T Rainforth, Y Gal
Advances in Neural Information Processing Systems 34, 28742-28756, 2021
1392021
Learning Sparse Networks Using Targeted Dropout
AN Gomez, I Zhang, S Rao Kamalakara, D Madaan, K Swersky, Y Gal, ...
arXiv preprint arXiv:1905.13678, 2019
1292019
Attention is all you need, 2023
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ...
arXiv preprint arXiv:1706.03762, 2023
1142023
The difficulty of training sparse neural networks
U Evci, F Pedregosa, A Gomez, E Elsen
arXiv preprint arXiv:1906.10732, 2019
1082019
Unsupervised cipher cracking using discrete GANs
AN Gomez, S Huang, I Zhang, BM Li, M Osama, L Kaiser
arXiv preprint arXiv:1801.04883, 2018
852018
Attention is all you need, December 2017
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ...
URL http://arxiv. org/abs/1706.03762 66, 0
64
Aya 23: Open weight releases to further multilingual progress
V Aryabumi, J Dang, D Talupuru, S Dash, D Cairuz, H Lin, B Venkitesh, ...
arXiv preprint arXiv:2405.15032, 2024
592024
Attention is all you need. DOI: 10.48550
A Vaswani, N Shazeer, N Parmar, J Uszkoreit, L Jones, AN Gomez, ...
arXiv preprint ARXIV.1706.03762, 2017
552017
Attention-based sequence transduction neural networks
NM Shazeer, AN Gomez, LM Kaiser, JD Uszkoreit, LO Jones, NJ Parmar, ...
US Patent 10,452,978, 2019
402019
המערכת אינה יכולה לבצע את הפעולה כעת. נסה שוב מאוחר יותר.
מאמרים 1–20