Follow
Konstantin Dobler
Title
Cited by
Cited by
Year
FOCUS: Effective Embedding Initialization for Monolingual Specialization of Multilingual Models
K Dobler, G De Melo
EMNLP 2023, 2023
302023
Art Creation with Multi-Conditional StyleGANs
K Dobler, F Hübscher, J Westphal, A Sierra-Múnera, G de Melo, R Krestel
International Joint Conference on Artificial Intelligence (IJCAI 2022), 2022
122022
Efficient Parallelization Layouts for Large-Scale Distributed Model Training
J Hagemann, S Weinbach, K Dobler, M Schall, G de Melo
COLM 2024; WANT@NeurIPS 2023 (Best Workshop Paper), 2023
32023
Focus: Effective embedding initialization for specializing pretrained multilingual models on a single language
K Dobler, G de Melo
arXiv preprint arXiv:2305.14481, 2023
32023
Knowledge Acquisition through Continued Pretraining is Difficult: A Case Study on r/AskHistorians
J Hoffbauer, S Sawicki, M Ulrich, T Buz, K Dobler, M Schneider, ...
Proceedings of the 1st Workshop on Towards Knowledgeable Language Models …, 2024
2024
Language Adaptation on a Tight Academic Compute Budget: Tokenizer Swapping Works and Pure bfloat16 Is Enough
K Dobler, G de Melo
WANT@ICML 2024, 2024
2024
I Don't Know: Explicit Modeling of Uncertainty with an [IDK] Token
R Cohen, K Dobler, E Biran, G de Melo
NeurIPS 2024, 2024
2024
Image Classification in the Art Domain
JM Burmeister, K Dobler, N Müller
Deep learning for computer vision in the art domain: proceedings of the …, 2021
2021
The system can't perform the operation now. Try again later.
Articles 1–8