追蹤
Vinh Q. Tran
Vinh Q. Tran
Research Scientist, Google DeepMind
在 google.com 的電子郵件地址已通過驗證 - 首頁
標題
引用次數
引用次數
年份
UL2: Unifying Language Learning Paradigms
Y Tay, M Dehghani, VQ Tran, X Garcia, J Wei, X Wang, HW Chung, ...
ICLR 2023, 2022
368*2022
Transformer memory as a differentiable search index
Y Tay, VQ Tran, M Dehghani, J Ni, D Bahri, H Mehta, Z Qin, K Hui, Z Zhao, ...
NeurIPS 2022, 2022
2032022
ExT5: Towards Extreme Multi-Task Scaling for Transfer Learning
V Aribandi, Y Tay, T Schuster, J Rao, HS Zheng, SV Mehta, H Zhuang, ...
ICLR 2022, 2021
1952021
A new generation of perspective api: Efficient multilingual character-level transformers
A Lees, VQ Tran, Y Tay, J Sorensen, J Gupta, D Metzler, L Vasserman
KDD'22 ADS, 2022
1492022
Confident adaptive language modeling
T Schuster, A Fisch, J Gupta, M Dehghani, D Bahri, VQ Tran, Y Tay, ...
NeurIPS 2022, 2022
1342022
Charformer: Fast character transformers via gradient-based subword tokenization
Y Tay, VQ Tran, S Ruder, J Gupta, HW Chung, D Bahri, Z Qin, ...
ICLR 2022, 2021
1302021
Attributed question answering: Evaluation and modeling for attributed large language models
B Bohnet, VQ Tran, P Verga, R Aharoni, D Andor, LB Soares, M Ciaramita, ...
arXiv preprint arXiv:2212.08037, 2022
892022
Recommender Systems with Generative Retrieval
S Rajput, N Mehta, A Singh, RH Keshavan, T Vu, L Heldt, L Hong, Y Tay, ...
NeurIPS 2023, 2023
832023
Scaling laws vs model architectures: How does inductive bias influence scaling?
Y Tay, M Dehghani, S Abnar, HW Chung, W Fedus, J Rao, S Narang, ...
EMNLP 2023 Findings, 2022
722022
Transcending scaling laws with 0.1% extra compute
Y Tay, J Wei, HW Chung, VQ Tran, DR So, S Shakeri, X Garcia, HS Zheng, ...
EMNLP 2023, 2022
602022
Making the case for Query-by-Voice with EchoQuery
G Lyons, V Tran, C Binnig, U Cetintemel, T Kraska
SIGMOD 2016, 2129-2132, 2016
572016
How Does Generative Retrieval Scale to Millions of Passages?
R Pradeep, K Hui, J Gupta, AD Lelkes, H Zhuang, J Lin, D Metzler, ...
EMNLP 2023, 2023
492023
Quiz-Style Question Generation for News Stories
AD Lelkes, VQ Tran, C Yu
WWW '21: Proceedings of the Web Conference 2021, Pages 2501–2511, 2021
472021
DSI++: Updating Transformer Memory with New Documents
SV Mehta, J Gupta, Y Tay, M Dehghani, VQ Tran, J Rao, M Najork, ...
EMNLP 2023, 2022
402022
AgreeSum: Agreement-Oriented Multi-Document Summarization
RY Pang, AD Lelkes, VQ Tran, C Yu
ACL-IJCNLP 2021 Findings, 3377–3391, 2021
182021
Smaller, Weaker, Yet Better: Training LLM Reasoners via Compute-Optimal Sampling
H Bansal, A Hosseini, R Agarwal, VQ Tran, M Kazemi
arXiv preprint arXiv:2408.16737, 2024
12024
Fractal Patterns May Illuminate the Success of Next-Token Prediction
I Alabdulmohsin, VQ Tran, M Dehghani
NeurIPS 2024, 2024
1*2024
Crossword puzzle generator
A Lelkes, C Keogh, RMH Gaughan III, K Tempero, C Yu, VQ Tran, ...
US Patent 10,967,248, 2021
12021
Character-level attention neural networks
Y Tay, D Bahri, DA Metzler, HW Chung, JP Gupta, SN Ruder, ...
US Patent App. 18/564,859, 2024
2024
Efficient Decoding of Output Sequences Using Adaptive Early Exiting
T Schuster, AJ Fisch, JP Gupta, M Dehghani, D Bahri, VQ Tran, Y Tay, ...
US Patent App. 18/222,395, 2024
2024
系統目前無法執行作業,請稍後再試。
文章 1–20