Aston Zhang
Applied Scientist
Autonomous user interface (UI) agents aim to facilitate task automation ...
A particularly successful class of approaches for few-shot learning comb...
Transformer models are foundational to natural language processing (NLP)...
Diffusion models that are based on iterative denoising have been recentl...
This work proposes POMP, a prompt pre-training method for vision-languag...
Spurred by advancements in scale, large language models (LLMs) have
demo...
Large language models (LLMs) have shown impressive performance on comple...
Parameter-efficient fine-tuning aims to achieve performance comparable t...
The ability to jointly learn from multiple modalities, such as text, aud...
Pre-trained large language models can efficiently interpolate human-writ...
The mixture of Expert (MoE) parallelism is a recent advancement that sca...
Deep neural networks (DNNs) are vulnerable to backdoor attacks. Previous...
Large language models (LLMs) can perform complex reasoning by generating...
Existing out-of-distribution (OOD) detection methods are typically
bench...
Adversarial training (AT) defends deep neural networks against adversari...
Data augmentation is a necessity to enhance data efficiency in deep lear...
Hypercomplex neural networks have proved to reduce the overall number of...
This open-source book represents our attempt to make deep learning
appro...
In E-commerce, a key challenge in text generation is to find a good trad...
Recent works have demonstrated reasonable success of representation lear...
With graphs rapidly growing in size and deeper graph neural networks (GN...
Modeling user interests is crucial in real-world recommender systems. In...
This paper reviews the novel concept of controllable variational autoenc...
This paper demonstrates a fatal vulnerability in natural language infere...
Pretrained Transformer-based language models (LMs) display remarkable na...
GitHub has become a popular social application platform, where a large n...
Variational Autoencoders (VAE) and their variants have been widely used ...
Transformer has been widely used thanks to its ability to capture sequen...
Traditionally, many text-mining tasks treat individual word-tokens as th...
We present GluonCV and GluonNLP, the deep learning toolkits for computer...
Many state-of-the-art neural models for NLP are heavily parameterized an...
This paper proposes Quaternion Collaborative Filtering (QCF), a novel
re...
This paper tackles the problem of reading comprehension over long narrat...
Expert finding is an important task in both industry and academia. It is...
Recent advances in deep learning motivate the use of deep neutral networ...
Mobile sensing applications usually require time-series inputs from sens...