Skip to main content

Jian Pei

Arthur S. Pearse Distinguished Professor of Computer Science
Computer Science
308 Research Drive, Durham, NC 27708

Overview


Data science, data mining, databases, information retrieval, computational statistics, applied machine learning and AI.

Current Appointments & Affiliations


Arthur S. Pearse Distinguished Professor of Computer Science · 2023 - Present Computer Science, Trinity College of Arts & Sciences
Professor of Computer Science · 2023 - Present Computer Science, Trinity College of Arts & Sciences
Chair of the Department of Computer Science · 2023 - Present Computer Science, Trinity College of Arts & Sciences
Professor of Biostatistics & Bioinformatics · 2023 - Present Biostatistics & Bioinformatics, Division of Biostatistics, Biostatistics & Bioinformatics
Professor of Electrical and Computer Engineering · 2023 - Present Electrical and Computer Engineering, Pratt School of Engineering

In the News


Published September 3, 2024
Six Duke Scholars Embark on Leadership Journey with Ivy+ Peers
Published November 12, 2023
Five Decades of Creating History and Pushing Boundaries at Duke Computer Science
Published July 25, 2023
Duke Effort Aims to End Bottlenecks Preventing Secondary Use of Big Data

View All News

Recent Publications


A Comprehensive Survey on Data Augmentation

Journal Article IEEE Transactions on Knowledge and Data Engineering · January 1, 2026 Data augmentation is a series of techniques that generate high-quality artificial data by manipulating existing data samples. By leveraging data augmentation techniques, AI models can achieve significantly improved applicability in tasks involving scarce o ... Full text Cite

A comprehensive survey on pretrained foundation models: a history from BERT to ChatGPT

Journal Article International Journal of Machine Learning and Cybernetics · December 1, 2025 Pretrained Foundation Models (PFMs) are regarded as the foundation for various downstream tasks across different data modalities. A PFM (e.g., BERT, ChatGPT, GPT-4) is trained on large-scale data, providing a solid parameter initialization for a wide range ... Full text Cite

Domain Specialization as the Key to Make Large Language Models Disruptive: A Comprehensive Survey

Journal Article ACM Computing Surveys · October 6, 2025 Large language models (LLMs) have significantly advanced the field of natural language processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of applications. However, directly applying LLMs to solve sophisticated problems i ... Full text Cite
View All Publications

Recent Grants


Collaborative Research: AI Tools to Knowledge Discovery and Rigorous Reasoning in Polyhedral Geometry

ResearchCo-Principal Investigator · Awarded by National Science Foundation · 2024 - 2027

EAGER: End-to-End Learning of Paradoxes and Interpretations for Data Storytelling

ResearchPrincipal Investigator · Awarded by National Science Foundation · 2023 - 2024

View All Grants

Education, Training & Certifications


Simon Fraser University · 2002 Ph.D.

External Links


Jian Pei Website