Folgen
Katherine Lee
Katherine Lee
Researcher, Google DeepMind
Bestätigte E-Mail-Adresse bei google.com - Startseite
Titel
Zitiert von
Zitiert von
Jahr
Exploring the limits of transfer learning with a unified text-to-text transformer
C Raffel, N Shazeer, A Roberts, K Lee, S Narang, M Matena, Y Zhou, W Li, ...
The Journal of Machine Learning Research 21 (1), 5485-5551, 2020
197842020
Palm: Scaling language modeling with pathways
A Chowdhery, S Narang, J Devlin, M Bosma, G Mishra, A Roberts, ...
Journal of Machine Learning Research 24 (240), 1-113, 2023
51462023
Gemini: a family of highly capable multimodal models
G Team, R Anil, S Borgeaud, Y Wu, JB Alayrac, J Yu, R Soricut, ...
arXiv preprint arXiv:2312.11805, 2023
21462023
Extracting Training Data from Large Language Models.
N Carlini, F Tramer, E Wallace, M Jagielski, A Herbert-Voss, K Lee, ...
USENIX Security Symposium 6, 2021
17802021
PaLM 2 Technical Report
R Anil, AM Dai, O Firat, M Johnson, D Lepikhin, A Passos, S Shakeri, ...
arXiv preprint arXiv:2305.10403, 2023
14112023
Gemma: Open Models Based on Gemini Research and Technology
G Team, T Mesnard, C Hardin, R Dadashi, S Bhupatiraju, S Pathak, ...
arXiv preprint arXiv:2403.08295, 2024
7112024
Quantifying Memorization Across Neural Language Models
N Carlini, D Ippolito, M Jagielski, K Lee, F Tramer, C Zhang
arXiv preprint arXiv:2202.07646, 2022
6192022
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
M Reid, N Savinov, D Teplyashin, D Lepikhin, T Lillicrap, J Alayrac, ...
arXiv preprint arXiv:2403.05530, 2024
5852024
Deduplicating training data makes language models better
K Lee, D Ippolito, A Nystrom, C Zhang, D Eck, C Callison-Burch, N Carlini
arXiv preprint arXiv:2107.06499, 2021
5382021
Are aligned neural networks adversarially aligned?
N Carlini, M Nasr, CA Choquette-Choo, M Jagielski, I Gao, PWW Koh, ...
Advances in Neural Information Processing Systems 36, 2024
2362024
Scalable Extraction of Training Data from (Production) Language Models
M Nasr, N Carlini, J Hayase, M Jagielski, AF Cooper, D Ippolito, ...
arXiv preprint arXiv:2311.17035, 2023
2282023
What Does it Mean for a Language Model to Preserve Privacy?
H Brown, K Lee, F Mireshghallah, R Shokri, F Tramèr
2022 ACM Conference on Fairness, Accountability, and Transparency, 2280-2292, 2022
2192022
WT5?! Training Text-to-Text Models to Explain their Predictions
S Narang, C Raffel, K Lee, A Roberts, N Fiedel, K Malkan
arXiv preprint arXiv:2004.14546, 2020
2052020
Counterfactual memorization in neural language models
C Zhang, D Ippolito, K Lee, M Jagielski, F Tramèr, N Carlini
Advances in Neural Information Processing Systems 36, 39321-39362, 2023
1352023
Hallucinations in neural machine translation
K Lee, O Firat, A Agarwal, C Fannjiang, D Sussillo
1332018
Preventing Verbatim Memorization in Language Models Gives a False Sense of Privacy
D Ippolito, F Tramèr, M Nasr, C Zhang, M Jagielski, K Lee, ...
arXiv preprint arXiv:2210.17546, 2022
1312022
A Pretrainer's Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity
S Longpre, G Yauney, E Reif, K Lee, A Roberts, B Zoph, D Zhou, J Wei, ...
arXiv preprint arXiv:2305.13169, 2023
1082023
Measuring Forgetting of Memorized Training Examples
M Jagielski, O Thakkar, F Tramèr, D Ippolito, K Lee, N Carlini, E Wallace, ...
arXiv preprint arXiv:2207.00099, 2022
842022
Madlad-400: A multilingual and document-level large audited dataset
S Kudugunta, I Caswell, B Zhang, X Garcia, D Xin, A Kusupati, R Stella, ...
Advances in Neural Information Processing Systems 36, 2024
772024
Propagation of information along the cortical hierarchy as a function of attention while reading and listening to stories
M Regev, E Simony, K Lee, KM Tan, J Chen, U Hasson
Cerebral Cortex 29 (10), 4017-4034, 2019
762019
Das System kann den Vorgang jetzt nicht ausführen. Versuchen Sie es später erneut.
Artikel 1–20