My research interests are Natural Language Processing/Understanding and Information Retrieval. Recently I am
focusing on following topics:
Continue learning and evolving of Large Language Models: Large language models are powerful, but they
need to be updated. How to efficiently but effectively update them so that they can adapt to new task, new domains
and incorporate new knowledge ?
Language modeling with structured data: Large Language (Multimodal) Models succedded in modeling and
reasoning on seqential
data or
unstructured data, such as text, images, and videos. But how can we incooperate language models with structured
data, such as information network, knowledge graphs, taxonomy, event triplets, and tabular data ?
Representation learning, information retrieval and their applications: Good represetations matters, they
connect
different modalities and models. Training representations for various task is a long lasting problem.
News
[2024-05-01] New persoanl page initial setup
Publications
Full publications on Google
Scholar. ‡ indicates equal contribution.
Selected
Patton: Language Model Pretraining on Text-Rich Networks
Bowen Jin, Wentao Zhang, Yu Zhang, Yu Meng, Xinyang Zhang, Qi Zhu and Jiawei Han