Antal van den Bosch
In my research I develop machine learning and language technology. Most of my work involves the intersection of the two fields: computers that learn to understand and generate natural language, nowadays known as Generative AI and Large Language Models. The computational models that this work produces have all kinds of applications in other areas of scholarly research as well as in society and industry. They also link in interesting ways to theories and developments in linguistics, psycholinguistics, neurolinguistics, and sociolinguistics. I love multidisciplinary collaborations to make advances in all these areas.
Session
Transformer-based LLMs, at scale, are prohibitively expensive to train, requiring massive GPU capacity. Alternative technologies do exist, producing functionally equivalent LLMs at a fraction of the training costs, using high-memory CPU nodes. I will illustrate this with a memory-based LLM trained on Snellius' hi-mem nodes.