
Using Large Language Models in Your Research: Fine-Tuning, Embeddings, and Using APIs
Registration
Details
(1) fine-tune a language model on their own data;
(2) use large-language-model-powered embedding tools for retrieval, clustering, classification, and so on;
(3) access frontier large language models, such as GPT-4, via APIs.
Speakers
Tianyu Gao
Princeton University
I am a 5th-year PhD student at Princeton University, advised by Prof. Danqi Chen. I am also a member of the Princeton NLP group and the Princeton Language and Intelligence. My research interests lie within the intersection of natural language processing and machine learning. I am specifically interested in how to better pre-train, fine-tune/instruction-tune, and evaluate large language models (LLMs). I am also interested in how to augment LLMs with external retrieval components to alleviate the hallucination problem and make them more trustworthy.
Mengzhou Xia
Graduate Student, Computer Science
I'm a fifth-year Computer Science Ph.D. candidate at Princeton NLP, advised by Prof. Danqi Chen. Prior to this, I was a master's student at Carnegie Mellon University, advised by Prof. Graham Neubig. I obtained my Bachelor's degree from Fudan University's School of Data Science in China. My research is partially supported by the 2024 Apple Scholars in AIML PhD fellowship and the 2022 Bloomberg Data Science Ph.D. Fellowship. I have interned at Meta AI, Microsoft Research, and Bloomberg AI throughout my PhD years.
Hosted By
Co-hosted with: GradFUTURES
Contact the organizers