@inproceedings{xiao2023trading,title={Trading Information between Latents in Hierarchical Variational Autoencoders},author={Xiao, Tim Z. and Bamler, Robert},booktitle={International Conference on Learning Representations (ICLR)},year={2023},}
Iterative Teaching by Data Hallucination
Zeju Qiu, Weiyang Liu, Tim Z. Xiao, Zhen Liu, Umang Bhatt, Yucen Luo, Adrian Weller, and 1 more author
In International Conference on Artificial Intelligence and Statistics (AISTATS), 2023
@inproceedings{Qiu2023DHT,title={Iterative Teaching by Data Hallucination},author={Qiu, Zeju and Liu, Weiyang and Xiao, Tim Z. and Liu, Zhen and Bhatt, Umang and Luo, Yucen and Weller, Adrian and Schölkopf, Bernhard},booktitle={International Conference on Artificial Intelligence and Statistics (AISTATS)},year={2023},}
@article{zhang2022out,title={Out-of-Distribution Detection with Class Ratio Estimation},author={Zhang, Mingtian and Zhang, Andi and Xiao, Tim Z. and Sun, Yitong and McDonagh, Steven},journal={arXiv preprint arXiv:2206.03955},year={2022},}
@article{zhang2022improving,title={Improving VAE-Based Representation Learning},author={Zhang, Mingtian and Xiao, Tim Z. and Paige, Brooks and Barber, David},journal={arXiv preprint arXiv:2205.14539},year={2022},}
@article{shah2021locally,title={Locally-Contextual Nonlinear CRFs for Sequence Labeling},author={Shah, Harshil and Xiao, Tim Z. and Barber, David},journal={arXiv preprint arXiv:2103.16210},year={2021},}
Exploiting Semi-Supervised Generative Model in Active Learning
My master project at UCL. Active learning tries to solve a practical problem for machine learning, which is to create a good model with only limited labelling budget. In this project, we exploit useful properties from semi-supervised generative model and use them in active learning. Our experiments in the half-moon and MNIST dataset show that by using semi-supervised generative model with simple acquisition function such as predictive entropy, we are able to improve the performance of active learning. Further experiments on our proposed acquisition functions expose interesting challenges in using data density provided by the model, which can be a valuable pointer for future active learning research.
@mastersthesis{xiao2021exploiting,author={Xiao, Tim Z.},school={University College London},title={Exploiting Semi-Supervised Generative Model in Active Learning},year={2021},}
2020
You Need Only Uncertain Answers: Data Efficient Multilingual Question Answering
Zhihao Lyu, Danier Duolikun, Bowei Dai, Yuan Yao, Pasquale Minervini, Tim Z. Xiao, and Yarin Gal
ICML 2020 Workshop on Uncertainty and Robustness in Deep Learning, 2020
I am an advisor for this project. Data scarcity is a major barrier for multilingual question answering: current systems work well with languages such as English where data is affluent, but face challenges with small corpora. As data labelling is expensive, previous works have resorted to pre-tuning systems on larger multilingual corpora followed by fine-tuning on the smaller ones. Instead of curating and labelling large corpora, we demonstrate a data efficient multi-lingual question answering system which only selects uncertain questions for labelling, reducing labelling efforts and costs. To realise this Bayesian active learning framework, we develop methodology to quantify uncertainty in several state-of-art attention-based Transfer question answering models. We then propose an uncertainty measure based on the variance of BLEU scores, and computed via Monte Carlo Dropout, to detect out-of-distribution questions. We finish by showing the effectiveness of our uncertainty measures in various out-of-distribution question answering settings.
@article{lyu2020you,title={You Need Only Uncertain Answers: Data Efficient Multilingual Question Answering},author={Lyu, Zhihao and Duolikun, Danier and Dai, Bowei and Yao, Yuan and Minervini, Pasquale and Xiao, Tim Z. and Gal, Yarin},journal={ICML 2020 Workshop on Uncertainty and Robustness in Deep Learning},year={2020},}
Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
My master project at Oxford. In the project, we detect out-of-training-distribution sentences in Neural Machine Translation using the Bayesian Deep Learning equivalent of Transformer models. For this we develop a new measure of uncertainty designed specifically for long sequences of discrete random variables – i.e. words in the output sentence. Our new measure of uncertainty solves a major intractability in the naive application of existing approaches on long sentences. We use our new measure on a Transformer model trained with dropout approximate inference. On the task of German-English translation using WMT13 and Europarl, we show that with dropout uncertainty our measure is able to identify when Dutch source sentences, sentences which use the same word types as German, are given to the model instead of German.
@article{xiao2020wat,title={Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers},author={Xiao, Tim Z. and Gomez, Aidan N. and Gal, Yarin},journal={arXiv preprint arXiv:2006.08344},year={2020},}
2016
A partial reconfiguration controller for Altera Stratix V FPGAs
Zhenzhong Xiao, Dirk Koch, and Mikel Lujan
In International Conference on Field Programmable Logic and Applications (FPL), 2016
@inproceedings{xiao2016fpga,author={Xiao, Zhenzhong and Koch, Dirk and Lujan, Mikel},booktitle={International Conference on Field Programmable Logic and Applications (FPL)},title={A partial reconfiguration controller for Altera Stratix V FPGAs},year={2016},doi={10.1109/FPL.2016.7577349},}