@inproceedings{xiao2025verbalized,title={Verbalized Machine Learning: Revisiting Machine Learning with Language Models},author={Xiao, Tim Z. and Bamler, Robert and Schölkopf, Bernhard and Liu, Weiyang},booktitle={Transactions on Machine Learning Research (TMLR)},video={https://www.youtube.com/watch?v=LCl_np5oPWA},year={2025},}
@inproceedings{liu2025fast,title={Fast Diversity-Preserving Reward Finetuning of Diffusion Models via Nabla-GFlowNets},author={Liu, Zhen and Xiao, Tim Z. and Liu, Weiyang and Bengio, Yoshua and Zhang, Dinghuai},booktitle={International Conference on Learning Representations (ICLR)},year={2025},}
@inproceedings{qiu2025can,title={Can Large Language Models Understand Symbolic Graphics Programs?},author={Qiu, Zeju and Liu, Weiyang and Feng, Haiwen and Liu, Zhen and Xiao, Tim Z. and Collins, Katherine M. and Tenenbaum, Joshua B. and Weller, Adrian and Black, Michael J. and Schölkopf, Bernhard},booktitle={International Conference on Learning Representations (ICLR)},year={2025},}
@inproceedings{ou2025improving,title={Improving Probabilistic Diffusion Models With Optimal Covariance Matching},author={Ou, Zijing and Zhang, Mingtian and Zhang, Andi and Xiao, Tim Z. and Li, Yingzhen and Barber, David},booktitle={International Conference on Learning Representations (ICLR)},year={2025},}
@inproceedings{zhang2025your,title={Your Finetuned Large Language Model is Already a Powerful Out-of-distribution Detector},author={Zhang, Andi and Xiao, Tim Z. and Liu, Weiyang and Bamler, Robert and Wischik, Damon},booktitle={International Conference on Artificial Intelligence and Statistics (AISTATS)},year={2025},}
2024
A Note on Generalization in Variational Autoencoders: How Effective Is Synthetic Data and Overparameterization?
Tim Z. Xiao*, Johannes Zenn*, and Robert Bamler
In Transactions on Machine Learning Research (TMLR), 2024
@inproceedings{xiao2024a,title={A Note on Generalization in Variational Autoencoders: How Effective Is Synthetic Data and Overparameterization?},author={Xiao, Tim Z. and Zenn, Johannes and Bamler, Robert},booktitle={Transactions on Machine Learning Research (TMLR)},year={2024},}
2023
A Compact Representation for Bayesian Neural Networks By Removing Permutation Symmetry
Tim Z. Xiao, Weiyang Liu, and Robert Bamler
arXiv preprint arXiv:2401.00611, 2023
NeurIPS 2023 Workshop on Unifying Representations in Neural Models
@article{xiao2023compact,title={A Compact Representation for Bayesian Neural Networks By Removing Permutation Symmetry},author={Xiao, Tim Z. and Liu, Weiyang and Bamler, Robert},journal={arXiv preprint arXiv:2401.00611},year={2023},}
The SVHN Dataset Is Deceptive for Probabilistic Generative Models Due to a Distribution Mismatch
@article{xiao2023the,title={The SVHN Dataset Is Deceptive for Probabilistic Generative Models Due to a Distribution Mismatch},author={Xiao, Tim Z. and Zenn, Johannes and Bamler, Robert},journal={arXiv preprint arXiv:2312.02168},year={2023},}
Trading Information between Latents in Hierarchical Variational Autoencoders
Tim Z. Xiao, and Robert Bamler
In International Conference on Learning Representations (ICLR), 2023
@inproceedings{xiao2023trading,title={Trading Information between Latents in Hierarchical Variational Autoencoders},author={Xiao, Tim Z. and Bamler, Robert},booktitle={International Conference on Learning Representations (ICLR)},year={2023},}
@inproceedings{Qiu2023DHT,title={Iterative Teaching by Data Hallucination},author={Qiu, Zeju and Liu, Weiyang and Xiao, Tim Z. and Liu, Zhen and Bhatt, Umang and Luo, Yucen and Weller, Adrian and Schölkopf, Bernhard},booktitle={International Conference on Artificial Intelligence and Statistics (AISTATS)},year={2023},}
@article{zhang2022out,title={Out-of-Distribution Detection with Class Ratio Estimation},author={Zhang, Mingtian and Zhang, Andi and Xiao, Tim Z. and Sun, Yitong and McDonagh, Steven},journal={arXiv preprint arXiv:2206.03955},year={2022},}
@article{zhang2022improving,title={Improving VAE-Based Representation Learning},author={Zhang, Mingtian and Xiao, Tim Z. and Paige, Brooks and Barber, David},journal={arXiv preprint arXiv:2205.14539},year={2022},}
@article{shah2021locally,title={Locally-Contextual Nonlinear CRFs for Sequence Labeling},author={Shah, Harshil and Xiao, Tim Z. and Barber, David},journal={arXiv preprint arXiv:2103.16210},year={2021},}
Exploiting Semi-Supervised Generative Model in Active Learning
My master project at UCL. Active learning tries to solve a practical problem for machine learning, which is to create a good model with only limited labelling budget. In this project, we exploit useful properties from semi-supervised generative model and use them in active learning. Our experiments in the half-moon and MNIST dataset show that by using semi-supervised generative model with simple acquisition function such as predictive entropy, we are able to improve the performance of active learning. Further experiments on our proposed acquisition functions expose interesting challenges in using data density provided by the model, which can be a valuable pointer for future active learning research.
@mastersthesis{xiao2021exploiting,author={Xiao, Tim Z.},school={University College London},title={Exploiting Semi-Supervised Generative Model in Active Learning},year={2021},}
2020
You Need Only Uncertain Answers: Data Efficient Multilingual Question Answering
Zhihao Lyu, Danier Duolikun, Bowei Dai, Yuan Yao, Pasquale Minervini, Tim Z. Xiao, and Yarin Gal
ICML 2020 Workshop on Uncertainty and Robustness in Deep Learning, 2020
I am an advisor for this project. Data scarcity is a major barrier for multilingual question answering: current systems work well with languages such as English where data is affluent, but face challenges with small corpora. As data labelling is expensive, previous works have resorted to pre-tuning systems on larger multilingual corpora followed by fine-tuning on the smaller ones. Instead of curating and labelling large corpora, we demonstrate a data efficient multi-lingual question answering system which only selects uncertain questions for labelling, reducing labelling efforts and costs. To realise this Bayesian active learning framework, we develop methodology to quantify uncertainty in several state-of-art attention-based Transfer question answering models. We then propose an uncertainty measure based on the variance of BLEU scores, and computed via Monte Carlo Dropout, to detect out-of-distribution questions. We finish by showing the effectiveness of our uncertainty measures in various out-of-distribution question answering settings.
@article{lyu2020you,title={You Need Only Uncertain Answers: Data Efficient Multilingual Question Answering},author={Lyu, Zhihao and Duolikun, Danier and Dai, Bowei and Yao, Yuan and Minervini, Pasquale and Xiao, Tim Z. and Gal, Yarin},journal={ICML 2020 Workshop on Uncertainty and Robustness in Deep Learning},year={2020},}
Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
Tim Z. Xiao, Aidan N. Gomez, and Yarin Gal
arXiv preprint arXiv:2006.08344, 2020
Spotlight talk, NeurIPS 2019 Workshop on Bayesian Deep Learning
My master project at Oxford. In the project, we detect out-of-training-distribution sentences in Neural Machine Translation using the Bayesian Deep Learning equivalent of Transformer models. For this we develop a new measure of uncertainty designed specifically for long sequences of discrete random variables – i.e. words in the output sentence. Our new measure of uncertainty solves a major intractability in the naive application of existing approaches on long sentences. We use our new measure on a Transformer model trained with dropout approximate inference. On the task of German-English translation using WMT13 and Europarl, we show that with dropout uncertainty our measure is able to identify when Dutch source sentences, sentences which use the same word types as German, are given to the model instead of German.
@article{xiao2020wat,title={Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers},author={Xiao, Tim Z. and Gomez, Aidan N. and Gal, Yarin},journal={arXiv preprint arXiv:2006.08344},year={2020},}
2016
A partial reconfiguration controller for Altera Stratix V FPGAs
Zhenzhong Xiao, Dirk Koch, and Mikel Lujan
In International Conference on Field Programmable Logic and Applications (FPL), 2016
@inproceedings{xiao2016fpga,author={Xiao, Zhenzhong and Koch, Dirk and Lujan, Mikel},booktitle={International Conference on Field Programmable Logic and Applications (FPL)},title={A partial reconfiguration controller for Altera Stratix V FPGAs},year={2016},doi={10.1109/FPL.2016.7577349},}