FOCUS: Effective Embedding Initialization for Monolingual Specialization of Multilingual Models K Dobler, G De Melo EMNLP 2023, 2023 | 30 | 2023 |
Art Creation with Multi-Conditional StyleGANs K Dobler, F Hübscher, J Westphal, A Sierra-Múnera, G de Melo, R Krestel International Joint Conference on Artificial Intelligence (IJCAI 2022), 2022 | 12 | 2022 |
Efficient Parallelization Layouts for Large-Scale Distributed Model Training J Hagemann, S Weinbach, K Dobler, M Schall, G de Melo COLM 2024; WANT@NeurIPS 2023 (Best Workshop Paper), 2023 | 3 | 2023 |
Focus: Effective embedding initialization for specializing pretrained multilingual models on a single language K Dobler, G de Melo arXiv preprint arXiv:2305.14481, 2023 | 3 | 2023 |
Knowledge Acquisition through Continued Pretraining is Difficult: A Case Study on r/AskHistorians J Hoffbauer, S Sawicki, M Ulrich, T Buz, K Dobler, M Schneider, ... Proceedings of the 1st Workshop on Towards Knowledgeable Language Models …, 2024 | | 2024 |
Language Adaptation on a Tight Academic Compute Budget: Tokenizer Swapping Works and Pure bfloat16 Is Enough K Dobler, G de Melo WANT@ICML 2024, 2024 | | 2024 |
I Don't Know: Explicit Modeling of Uncertainty with an [IDK] Token R Cohen, K Dobler, E Biran, G de Melo NeurIPS 2024, 2024 | | 2024 |
Image Classification in the Art Domain JM Burmeister, K Dobler, N Müller Deep learning for computer vision in the art domain: proceedings of the …, 2021 | | 2021 |