Pei (Patrick) Chen

Pei (Patrick) Chen

Ph.D. in Computer Science

Texas A&M University

Welcome!

I am a Ph.D. in Computer Science graduated from the TAMU NLP group, Department of Computer Science and Engineering at Texas A&M University. I enjoy doing research in the fields of Natural Language Processing, Deep Learning and applying deep learning models to practical NLP problems and applications. I joined Amazon.com as an Applied Scientist since Jan, 2024.

Email: chenpei.net@gmail.com
Links: GitHub LinkedIn Google Scholar

Interests
  • Large Language Models
  • Information Extraction
Education
  • Ph.D. in Computer Science, 2019 - 2024

    Texas A&M University

  • MS in Finance, 2016 - 2018

    Southwestern University of Finance and Economics

  • B.Eng. in Simulation Engineering, 2010 - 2014

    National University of Defense Technology

News

  • May/24 One paper accepted to NAACL-2024 and one accepted to ACL-2024. ✨
  • Sep/23 One paper accepted to NeurIPS-2023 as a spotlight. πŸ‘‹

Selected Publications

(2024). ItD: Large Language Models Can Teach Themselves Induction through Deduction.

Cite arXiv

(2024). Mosaic IT: Enhancing Instruction Tuning with Data Mosaics.

Cite arXiv

(2023). HYTREL: Hypergraph-enhanced Tabular Data Representation Learning. NeurIPS, 2023 (spotlight).

Cite arXiv code

(2023). ZeroKBC: A Comprehensive Benchmark for Zero-Shot Knowledge Base Completion. ICDMW 2022.

PDF Cite Dataset

(2022). Crossroads, Buildings and Neighborhoods: A Dataset for Fine-grained Location Recognition. NAACL 2022, long paper, acceptance rate: 21.96%.

PDF Cite Code Dataset

(2021). Explicitly Capturing Relations between Entity Mentions via Graph Neural Networks for Domain-specific Named Entity Recognition. ACL 2021, short paper, acceptance rate: 21.2%.

PDF Cite Code Dataset

(2021). Probing into the Root: A Dataset for Reason Extraction of Structural Events from Financial Documents. EACL 2021, short paper, acceptance rate: 24.7%.

PDF Cite Code Dataset

(2020). Reconstructing Event Regions for Event Extraction via Graph Attention Networks. AACL 2020, long paper, acceptance rate: 28.3%.

PDF Cite Dataset

Experience

 
 
 
 
 
Applied Scientist
Jan 2024 – Present Palo Alto, U.S.
Working on large language models and their applications.
 
 
 
 
 
Applied Scientist (Intern)
May 2023 – Aug 2023 Sata Clara, U.S.
Working on large language model prompting.
 
 
 
 
 
Applied Scientist (Intern)
Jun 2022 – Jan 2023 Sata Clara, U.S.
Working on language model pretraining.
 
 
 
 
 
NLP Researcher (Intern)
Jun 2021 – Aug 2021 Seattle, U.S.
Working on Knowledge Fusion and Representation.
 
 
 
 
 
Research Engineer
Jan 2018 – Aug 2019 Beijing, China
Working on event extraction and causality detection from financial domain texts.
 
 
 
 
 
Data Analyst (Intern)
Jul 2017 – Jan 2018 Hangzhou, China
Working on data tiding, analysis, and visualization for innovative financial applications.

Misc.

I love sports! I go to the recreation center every week for a diverse range of activities, including:

  • Swimming
  • Badminton
  • Boxing (1st DAN holder in ITF Taekwon-Do)