Date: Thu 19 Sep 2024

Abstract

Most algorithms for optimising nonlinear functions rely on access to (possibly stochastic) derivative information. However, for problems including adversarial example generation for neural networks and fine-tuning large language models, good derivative information can be difficult to obtain and "derivative-free" optimisation (DFO) algorithms are beneficial. Although there are many approaches for DFO, they generally struggle to solve large-scale problems such as those arising in machine learning. In this talk, I will introduce new scalable DFO algorithms based on random subspaces and develop a novel average-case analysis of such algorithms.

Speaker

Lindon Roberts

Research Area

Applied Mathematics

Affiliation

University of Sydney

Date

Thu 19 Sep 2024

Venue

Anita B. Lawrence 4082 and online via Zoom (Link below; password: 454808)