Skip navigation
The Australian National University

Student research opportunities

Large Scale Optimization for Sparse Modeling

Project Code: CECS_977

This project is available at the following levels:
Honours, Masters, PhD

Keywords:

Optimization, sparse coding

Supervisor:

Dr Xinhua Zhang

Outline:

At the heart of many machine learning algorithms is the problem of convex optimization. It characterizes the desirable properties of a model by regularizers and loss functions, and the convexity allows modelling to be decoupled with optimization. In the era of big data, the computational cost of conventional optimization methods has become exorbitant, and therefore customized algorithms are required to restore efficiency. This requires a refined exploitation of the structures in the problem, such as decomposition of loss, sparsity or low-rankness of the optimal model, and other efficient approximations of convex sets or functions.

Goals of this project

The project aims to develop novel efficient optimization algorithms for advanced sparse models, which originate from convex relaxation of submodular functions. It will leverage techniques from discrete optimization, convex analysis, and statistics. The developed approach can be readily used in applications such as computer vision and language modelling.

Requirements/Prerequisites

Good knowledge of linear algebra.

Ability to program, and willingness to learn.

Background Literature

See my publication page.


Contact:



Updated:  26 September 2013 / Responsible Officer:  JavaScript must be enabled to display this email address. / Page Contact:  JavaScript must be enabled to display this email address.