Invited speakers

Abhik Roychoudhury

Abhik Roychoudhury is Provost’s Chair Professor of Computer Science at the National University of Singapore (NUS), where he leads a research team on Trustworthy and Secure Software (TSS). He is also a Senior Advisor at SonarSource, subsequent to the acquisition of his spinoff AutoCodeRover on AI-based coding, by Sonar. Abhik received his PhD in Computer Science from the Stony Brook University in 2000, and has been a faculty member at NUS School of Computing since 2001. Abhik’s research group at NUS is known for foundational contributions to software testing and analysis. Specifically the team has made contributions to automatic programming and automated program repair, as well as to fuzz testing for finding security vulnerabilities in software systems. These works have been honored with various awards including an International Conference on Software Engineering (ICSE) Most Influential Paper Award (Test-of-time award) for program repair, IEEE New Directions Award 2022 (jointly with Cristian Cadar) for contributions to symbolic execution.

Abhik was the inaugural recipient of the NUS Outstanding Graduate Mentor Award 2024. Doctoral students graduated from his research team have taken up faculty positions in many academic institutions. He has served the software engineering research community in various capacities including as chair of the major conferences of the field, ICSE and FSE. Currently, he serves as chair of the FSE steering committee. He is the current Editor-in-Chief of the ACM Transactions on Software Engineering and Methodology (TOSEM), and is a member of the editorial board of Communications of the ACM. Abhik is a Fellow of the ACM.

AutoCodeRover:  from research on automatic programming to spinoff acquisition

Large Language Models (LLMs) have shown surprising proficiency in generating code snippets, promising to automate large parts of software engineering via artificial intelligence (AI). We argue that successfully deploying AI software engineers requires a level of trust equal to or even greater than the trust established by human-driven software engineering practices. The recent trend toward LLM agents offers a path toward integrating the power of LLMs to create new code with the power of program analysis tools to increase trust in the code.
In this talk, we will share our experience with students in coming up with the design of AutoCodeRover, an early approach towards agentic Artificial Intelligence (AI) in coding, which is seeing increased attention in 2025. The main differentiation of AutoCodeRover with other proposals was its focus on using program analysis tools autonomously. This allows AutoCodeRover to infer developer intent or specification inference, thereby successfully conducting program improvement such as program repair or feature addition. AutoCodeRover was a spinoff from NUS which has been acquired by SonarSource in February 2025.

We will conclude the talk with a discussion on how agentic AI may be shifting the balance in programming — programming with trust becoming more important than programming at scale.

Fabian Wenz

Fabian Wenz is a researcher at the Computer Science and Artificial Intelligence Laboratory (CSAIL) at the Massachusetts Institute of Technology (MIT), where he works with Cagatay Demiralp and Michael Stonebraker. He holds bachelor’s degrees in Mathematics and Computer Science and a master’s degree in Mathematics and Data Science from the Technical University of Munich (TUM).
His research lies at the intersection of large language models (LLMs), data management, and machine learning systems. He focuses on improving LLM performance for complex reasoning over enterprise databases and structured data. He currently leads the development of BenchPress, a suite of tools and benchmarks for measuring the accuracy of LLM-generated SQL on production database logs.

This popup may have any size and colors

Etiam euismod lobortis iaculis. Fusce in nisi in enim facilisis pellentesque. Maecenas pulvinar lacus sed dui scelerisque pharetra. Praesent interdum velit pretium rutrum volutpat. Proin porttitor neque sapien.

Donec et urna in metus eleifend condimentum vel et lacus. Cras ac aliquam nulla. In luctus pretium ligula, et mollis velit laoreet rhoncus. Morbi at nibh ac ligula porta convallis. Lorem ipsum dolor sit amet, consectetur adipiscing elit.

 

Till Döhmen

Till Döhmen leads the AI team at MotherDuck in Amsterdam. He is also pursuing an external part-time PhD at BIFOLD, TU Berlin, focusing on the intersection of Data Management and ML. His background is in software engineering, with significant experience in data science and ML engineering. Over the past years, he has maintained a presence in both research and industry, embracing the combination of real-world applications and scientific insights. His journey with DuckDB began, with writing the first implementation of DuckDB’s CSV sniffer, and he subsequently worked on a series of DuckDB-related projects (data quality evaluation for ML pipelines; building a DuckDB-based lakehouse query engine for the Hopsworks ML Feature Store). At MotherDuck, he is now working on harnessing the power of LLMs for database users, from Text-to-SQL to leveraging LLMs for unstructured data analysis right within the database.

This content will appear inside a popup…

Effy Xue Li

Effy Xue Li is currently a postdoc at CWI (Centrum Wiskunde & Informatica). Her research focuses on Knowledge Graph Construction from Conversational Data, with a particular interest in leveraging large language models (LLMs) to extract structured information such as entities and relations. She also explores how to make LLMs more data-efficient, robust, and adaptable, especially in the context of data management. She recently completed an internship at MotherDuck, working on the application of LLMs for data management tasks. Prior to that, she was an AI Resident at Microsoft Research Cambridge in the UK. She holds a PhD from the University of Amsterdam in the INDE Lab and a Master’s degree from the University of Edinburgh.

This content will appear inside a popup…

Royal Sequeira

Royal is a Machine Learning Engineer at Georgian’s AI Lab, where he helps portfolio companies build AI-driven product features and accelerate their go-to-market (GTM) strategies. He also supports sourcing and diligence for new investments. Previously, he worked at Ada Support as one of their first ML hires, LG Toronto AI Lab, and Microsoft Research India. In 2018, he founded Sushiksha, a mentorship organization that has mentored hundreds of medical and engineering students across rural India with both technical and soft skills.

This content will appear inside a popup…

Jenn McArthur

Dr. Jenn McArthur is an Associate Professor in the Department of Architectural Science at Toronto Metropolitan University. She holds a PhD in Management (University of Edinburgh) and MASc and BASc in Mechanical Engineering (both University of Waterloo). Jenn also serves as  Associate Chair for the Graduate Program in Project Management in the Built Environment. Jenn’ research program centers on the development of Smart & Sustainable Building Solutions including Smart and Continuous Commissioning applications, Cognitive Digital Twins for Smart Buildings, and supporting industry adoption of Smart technologies. Her other research focuses on building performance improvement through Smart and Ongoing Commissioning (SOCx), Smart Campus Integration, FM-enabled BIM, and  workplace design to improve productivity and health. Currently, Jenn is working with Infrastructure Ontario to develop a provincial Digital Twins strategy. Prior to her academic career, Jenn worked for over a decade as a mechanical engineer in building design, rural development, and disaster relief on three continents.

Cognitive Digital Twins for Decarbonization – Integrating AI with Building Technologies

Cognitive Digital Twins (CDTs) are well-established in the manufacturing sector but have had very little implementation in the buildings sector, which has a low degree of digitization. This presentation highlights the key steps needed to create the first known proof-of-concept CDT for a real-world building, the Daphne Cockwell Complex at TMU and how it is being used to support decarbonization efforts at TMU. The CDT development to date will be presented, describing the approaches and algorithms used, results achieved, and lessons learned. Topics covered will include the integrated data model and supporting ontology; building automation system data acquisition and streaming; data lake; event detection algorithms; integration, and visualization. Insights regarding approach selection, implementation considerations, limitations, and alternatives are presented for each to guide the remaining steps (learning to support true cognition) in the CDT development. Other decarbonization CDT projects in progress will also be presented, including those at the urban scale and at the equipment scale.

Istvan David

Dr. Istvan David is an Assistant Professor of Computer Science and Software Engineering in the Faculty of Engineering at McMaster University. His research is situated under the broader systems engineering umbrella, with a particular interest in the engineering of complex and sustainable systems by digital accelerators, such as digital twins and artificial intelligence. The mission of Dr. David’s program is to promote sustainable practices in complex systems engineering, especially through modeling and simulation and model-driven engineering. He maintains additional lines of research in select topics in modeling and simulation, cyber-physical systems, automated software and systems construction, and collaborative modeling.

In this talk, Dr. David outlines select applications of digital twins for engineering complex and sustainable systems and will touch upon the topics of cyber-biophysical systems, AI simulation, and twin transition.

Ian Arawjo

Ian Arawjo is an Assistant Professor of Human-Computer Interaction at the University of Montréal, in the Department of Computer Science and Operations Research (DIRO), and is affiliated with Mila. He leads the Montréal HCI group. He was previously a Postdoctoral Fellow at Harvard University and holds a Ph.D. in Information Science from Cornell University. His research explores the social and cultural dimensions of programming, combining methods such as ethnographic fieldwork, archival research, system design, and usability studies. His work has received awards at top HCI conferences including CHI, CSCW, and UIST.

He is the creator and lead developer of ChainForge, the first open-source visual programming environment for prompt engineering, developed in collaboration with colleagues at Harvard CS. He also introduced notational programming, a paradigm that blends handwritten notation with traditional code.

This content will appear inside a popup…

Daniel Jaroslawicz

Daniel is a Research Engineer at Distyl AI, where he develops compound AI systems to automate complex workflows at Fortune 500 companies. His work focuses on transforming organizational knowledge into structured representations that power AI-driven workflows and enable continuous improvement through both automated and expert-driven feedback loops. Prior to Distyl, he was an AI Scientist at BenevolentAI where he worked on hypothesis generation with LLMs for novel drug target discovery. He holds a BA and MS in Computer Science from Columbia University.

This content will appear inside a popup…

Olivier Nguye

Olivier is an Applied Scientist at Twitch working on the Safety ML team building machine learning applications to detect and remove harmful content and behavior from the platform. Previously, he was an Applied Research Scientist at ServiceNow that joined through the acquisition of Element AI. His previous work focused on problems in NLP (question answering, semantic search, natural language understanding) and everything around it (research, engineering, scaling).

Using LLMs for content moderation at Twitch

Recent advances in Large Language Models (LLMs) have opened new possibilities for automating complex decision-making tasks that traditionally required extensive human expertise. This talk presents a case study on leveraging LLMs to enhance content moderation systems for large-scale trust and safety operations. We demonstrate how modern LLMs’ capabilities – including large context windows, multilingual understanding, and structured reasoning – can be applied to interpret nuanced platform policies and make consistent moderation decisions while providing detailed explanations. We discuss the technical architecture, prompt engineering approaches, and evaluation frameworks developed for this system. Our findings show promising results in handling complex moderation scenarios, while highlighting important considerations around scalability and edge cases. The presentation presents exploration of future directions, particularly the potential of composable AI agents to handle increasingly complex trust and safety workflows.

Damien Masson

Damien Masson is an assistant professor in human-computer interaction (HCI) at the Université de Montréal, where he co-leads the Montréal HCI group. He is also an associate academic member of Mila and IVADO professor. Previously, he was a postdoctoral researcher at the University of Toronto. He obtained his PhD in the HCI Lab at the University of Waterloo. His research focuses on building AI-infused systems that prioritize thoughtful interaction design to effectively solve tasks. Currently, his projects relate to designing intelligent system for assisting creative writers, visualizing complex information, and making digital documents easier to understand. His work is published in top-tier HCI venues such as CHI and UIST and received multiple honours, including: a best demo award (CHI 2021); a best paper award (CHI 2023); the AFIHM dissertation award (2023); the Bill Buxton Dissertation Award (2023); and best paper honourable mentions (CHI 2024, CHI 2025).

Using LLMs for content moderation at Twitch

Recent advances in Large Language Models (LLMs) have opened new possibilities for automating complex decision-making tasks that traditionally required extensive human expertise. This talk presents a case study on leveraging LLMs to enhance content moderation systems for large-scale trust and safety operations. We demonstrate how modern LLMs’ capabilities – including large context windows, multilingual understanding, and structured reasoning – can be applied to interpret nuanced platform policies and make consistent moderation decisions while providing detailed explanations. We discuss the technical architecture, prompt engineering approaches, and evaluation frameworks developed for this system. Our findings show promising results in handling complex moderation scenarios, while highlighting important considerations around scalability and edge cases. The presentation presents exploration of future directions, particularly the potential of composable AI agents to handle increasingly complex trust and safety workflows.

Alexandre Lacoste

Alexandre is a Staff Research Manager leading the UI-Assist team at ServiceNow. Since his PhD in theoretical machine learning, he has published various influential works around foundation models, sequential decision-making, causality, and various works in collaboration with Yoshua Bengio. Prior to ServiceNow, he was the first research scientist to join Element AI and he worked for 3 years at Google. He also is a founding member of ClimateChangeAI and an author of the seminal paper Tackling Climate Change with Machine Learning.

Autonomous UI Agents: Open-Source Tools, Best Practices, and Safety

This talk presents an overview of the current open-source ecosystem for building and evaluating autonomous web agents. We will cover key platforms including BrowserGym, AgentLab, WebArena, WorkArena, and DoomArena, highlighting their design choices, capabilities, and how they can be used together to support research on UI-based decision-making. The talk is aimed at researchers interested in language agents, reinforcement learning, and interactive environments. We will also introduce core concepts for developing agents, including DOM parsing, AXTree, action spaces, and evaluation protocols. A dedicated section will address emerging security concerns such as prompt injection and unsafe tool use. The goal is to provide a clear understanding of the available tools, common workflows, and practical considerations for safely developing robust web agents.

Orlando Marquez

Orlando is a ServiceNow Lead Applied Research Scientist with a strong background in software engineering. One of his passions is shipping state-of-the-art AI to end-users through rigorous and careful experimentation as well as sound engineering. He has worked on several NLP tasks such as question answering, natural language understanding and summarization. He most recently led the development of Flow Generation, a GenAI feature with hundreds of enterprise users that automatically generates low-code workflows from text. Nowadays, he is trying to figure out how to reliably ship web agents. He holds a Bachelor of Software Engineering from the University of Waterloo and a Masters of Computer Science from the Université de Montréal (MILA).

Autonomous UI Agents: Open-Source Tools, Best Practices, and Safety

Weiyi Shang

Weiyi Shang is an Associate Professor at the University of Waterloo. His research interests include AIOps, big data software engineering, software log analytics and software performance engineering. He serves as a Steering committee member of the SPEC Research Group. He is ranked top worldwide SE research stars in a recent bibliometrics assessment of software engineering scholars. He is a recipient of various premium awards, including the CSCAN-INFOCAN Outstanding Early Career Computer Science Researcher Prize in 2021, the SIGSOFT Distinguished paper award at ICSE 2013, ICSE 2020 and 2025, best paper award at WCRE 2011 and the Distinguished reviewer award for the Empirical Software Engineering journal. His research has been adopted by industrial collaborators (e.g., BlackBerry and Ericsson) to improve the quality and performance of their software systems that are used by millions of users worldwide. Contact him at wshang@uwaterloo.cauwaterloo.ca/electrical-computer-engineering/profile/wshang.

Evaluating the efficiency of LLM-generated code

The integration of Large Language Models (LLMs) into software development holds the promise of transforming code generation processes. While AIdriven code generation presents numerous advantages for software development, code generated by large language models may introduce sub-optional efficiency of the generated code. In this talk, I will share our recent study on the efficiency of LLM generated code and associated benchmarks. I will also share our initial try out of using prompt engineering as a potential strategy for optimizing efficiency of LLM-generated code.