Home NewsPublicationsTeachingSoftwareAwardsOther

Lars Kotthoff

Lars Kotthoff

Templeton Associate Professor

Derecho Professor

Presidential Faculty Fellow

larsko@uwyo.edu

EERB 422b
Department of Electrical Engineering and Computer Science
School of Computing
University of Wyoming
1000 E University Ave
Laramie, WY 82071-2000

My research combines artificial intelligence and machine learning to build robust systems with state-of-the-art performance. I develop techniques to induce models of how algorithms for solving computationally difficult problems behave in practice. Such models allow to select the best algorithm and choose the best parameter configuration for solving a given problem. I lead the Meta-Algorithmics, Learning and Large-scale Empirical Testing (MALLET) lab at the University of Wyoming.

More broadly, I am interested in innovative ways of modelling and solving challenging problems and applying such approaches to the real world. Part of this is making cutting edge research available to and usable by non-experts. Machine learning often plays a crucial role in this, and I am also working on making machine learning more accessible and easier to use.

Interested in coming to beautiful Wyoming and joining MALLET? Please drop me an email or, if you are already here, come by my office. I also have Master's projects and projects for undergraduates seeking research experience available.

News

Publications

For citation numbers, please see my Google Scholar page.

2025

  • Raponi, Elena, Lars Kotthoff, Hyunsun Alicia Kim, and Marius Lindauer. “Automated Machine Learning For Computational Mechanics (Dagstuhl Seminar 24282).” Edited by Elena Raponi, Lars Kotthoff, Hyunsun Alicia Kim, and Marius Lindauer. Dagstuhl Reports 14, no. 7 (2025): 17–34. https://doi.org/10.4230/DagRep.14.7.17. preprint PDF bibTeX abstract

    Machine learning (ML) has achieved undeniable success in computational mechanics, an ever-growing discipline that impacts all areas of engineering, from structural and fluid dynamics to solid mechanics and vehicle simulation. Computational mechanics uses numerical models and time- and resource-consuming simulations to reproduce physical phenomena, usually with the goal of optimizing the parameter configuration of the model with respect to the desired properties of the system. ML algorithms enable the construction of surrogate models that approximate the outcome of the simulations, allowing faster identification of well-performing configurations. However, determining the best ML approach for a given task is not straightforward and depends on human experts. Automated machine learning (AutoML) aims to reduce the need for experts to obtain effective ML pipelines. It provides off-the-shelf solutions that can be used without prior knowledge of ML, allowing engineers to spend more time on domain-specific tasks. AutoML is underutilized in computational mechanics; there is almost no communication between the two communities, and engineers spend unnecessary effort selecting and configuring ML algorithms. Our Dagstuhl Seminar aimed to (i) raise awareness of AutoML in the computational mechanics community, (ii) discover strengths and challenges for applying AutoML in practice, and (iii) create a bilateral exchange so that researchers can mutually benefit from their complementary goals and needs.
  • Stein, Niki van, Anna V. Kononova, Lars Kotthoff, and Thomas Bäck. “Code Evolution Graphs: Understanding Large Language Model Driven Design of Algorithms.” In Genetic and Evolutionary Computation Conference (GECCO), 2025. preprint PDF bibTeX abstract

    Large Language Models (LLMs) have demonstrated great promise in generating code, especially when used inside an evolutionary computation framework to iteratively optimize the generated algorithms. However, in some cases they fail to generate competitive algorithms or the code optimization stalls, and we are left with no recourse because of a lack of understanding of the generation process and generated codes. We present a novel approach to mitigate this problem by enabling users to analyze the generated codes inside the evolutionary process and how they evolve over repeated prompting of the LLM. We show results for three benchmark problem classes and demonstrate novel insights. In particular, LLMs tend to generate more complex code with repeated prompting, but additional complexity can hurt algorithmic performance in some cases. Different LLMs have different coding ``styles'' and generated code tends to be dissimilar to other LLMs. These two findings suggest that using different LLMs inside the code evolution frameworks might produce higher performing code than using only one LLM.

2024

  • Bischl, Bernd, Raphael Sonabend, Lars Kotthoff, and Michel Lang, eds. Applied Machine Learning Using mlr3 in R. 1st ed. CRC Press, 2024. preprint PDF bibTeX

  • Rogers, Jen, Marie Anastacio, Jürgen Bernard, Mehdi Chakhchoukh, Rebecca Faust, Andreas Kerren, Steffen Koch, Lars Kotthoff, Cagatay Turkay, and Emily Wall. “Visualization and Automation in Data Science: Exploring the Paradox of Humans-in-the-Loop.” In IEEE Visualization in Data Science (VDS), 1–5, 2024. https://doi.org/10.1109/VDS63897.2024.00005. preprint PDF bibTeX abstract

    We explore the interplay between automation and human involvement in data science. Emerging from in-depth discussions at a Dagstuhl seminar, we synthesize perspectives from Automated Data Science (AutoDS) and Interactive Data Visualization (VIS) – two fields that traditionally represent opposing ends of the human-machine spectrum. While AutoDS seeks to enhance efficiency through increasing automation, VIS underscores the critical value of human involvement in providing nuanced understanding, creativity, innovation, and contextual relevance. We explore these dichotomies through an online survey and advocate for a balanced approach that harmonizes the speed and consistency of effective automation with the indispensable insights of human expertise and thought. Ultimately, we confront the essential question: what aspects of data science should we automate?
  • Crisan, Anamaria, Lars Kotthoff, Marc Streit, and Kai Xu. “Human-Centered Approaches for Provenance in Automated Data Science (Dagstuhl Seminar 23372).” Edited by Anamaria Crisan, Lars Kotthoff, Marc Streit, and Kai Xu. Dagstuhl Reports 13, no. 9 (2024): 116–36. https://doi.org/10.4230/DagRep.13.9.116. preprint PDF bibTeX abstract

    The scope of automated machine learning (AutoML) technology has extended beyond its initial boundaries of model selection and hyperparameter tuning and towards end-to-end development and refinement of data science pipelines. These advances, both theoretical and realized, make the tools of data science more readily available to domain experts that rely on low- or no-code tooling options to analyze and make sense of their data. To ensure that automated data science technologies are applied both effectively and responsibly, it becomes increasingly urgent to carefully audit the decisions made both automatically and with guidance from humans. This Dagstuhl Seminar examines human-centered approaches for provenance in automated data science. While prior research concerning provenance and machine learning exists, it does not address the expanded scope of automated approaches and the consequences of applying such techniques at scale to the population of domain experts. In addition, most of the previous works focus on the automated part of this process, leaving a gap on the support for the sensemaking tasks users need to perform, such as selecting the datasets and candidate models and identifying potential causes for poor performance. The seminar brought together experts from across provenance, information visualization, visual analytics, machine learning, and human-computer interaction to articulate the user challenges posed by AutoML and automated data science, discuss the current state of the art, and propose directions for new research. More specifically, this seminar: - articulates the state of the art in AutoML and automated data science for supporting the provenance of decision making, - describes the challenges that data scientists and domain experts face when interfacing with automated approaches to make sense of an automated decision, - examines the interface between data-centric, model-centric, and user-centric models of provenance and how they interact with automated techniques, and - encourages exploration of human-centered approaches; for example leveraging visualization.
See all

Software

  • Core contributor to the mlr3 R ecosystem (book) for all things machine learning in R.

  • Maintainer of the FSelector R package.

  • Author and maintainer of LLAMA, an R package to simplify common algorithm selection tasks such as training a classifier as portfolio selector.

  • Leading the Auto-WEKA project, which brings automated machine learning to WEKA.

Teaching

  • I am teaching COSC 3020 (Algorithm Design and Analysis) and COSC 4552/5552 (Advanced Topics in AI). Lecture materials, assignments, announcements, etc. are available on WyoCourses.

  • I am teaching and developing a course on Practical Machine Learning (materials offline because of a major revision, sorry).

Awards

Other

Apart from my main affiliation, I am a research associate with the Maya Research Program. If I'm not in the office, it's possible that you can find me in the jungle of Belize excavating and/or mapping Maya ruins. Check out the interactive map.

I am also involved with the OpenML project project and a core contributor to ASlib, the benchmark library for algorithm selection.

While you're here, have a look at my overview of the Algorithm Selection literature. For something more visual, have a look at my pictures on Flickr.