Publications

You can also find my articles on my Google Scholar profile.

Preprints


Conference Papers


Generative Pretrained Structured Transformers: Unsupervised Syntactic Language Models at Scale

Published in ACL, 2024

We propose GPST, a syntactic language model which could be pre-trained on raw text efficiently without any human-annotated trees. When GPST and GPT-2 are both pre-trained on OpenWebText from scratch, GPST can outperform GPT-2 on various downstream tasks. Moreover, it significantly surpasses previous methods on generative grammar induction tasks, exhibiting a high degree of consistency with human syntax.

Download Paper

Augmenting Transformers with Recursively Composed Multi-Grained Representations

Published in ICLR, 2024

We reduce the space complexity of the deep inside-outside algorithm from cubic to linear and further reduce the parallel time complexity to approximately log N thanks to the new pruning algorithm proposed in this paper. Furthermore, we find that joint pre-training of Transformers and composition models can enhance a variety of NLP downstream tasks. We push unsupervised constituency parsing performance to 65% and demonstrate that our model could outperform vanillar Trasformers around 5% on span-level tasks.

Download Paper