Chad Jipiti
AI & NLP Researcher
Summary
AI researcher with 8+ years of experience in natural language processing and machine learning. Specializing in transformer architectures, neural networks, and large language models with a strong focus on advancing conversational AI and human-computer interaction.
Skills
Technical Skills
- •PyTorch / TensorFlow / JAX / Transformers
- •Python / Julia / C++ / CUDA
- •Attention Mechanisms / Self-Supervised Learning
- •NLP / Large Language Models / Neural Networks
- •Git / GitHub / CI/CD / Distributed Computing
Soft Skills
- •Problem Solving / Critical Thinking
- •Communication / Collaboration
- •Time Management / Organization
- •Adaptability / Continuous Learning
- •Attention to Detail / User-Centered Design
Work Experience
Senior AI Researcher
DeepMind AI Labs
Oct 2022 - Present
San Francisco, CA
- •Leading research on next-generation multimodal large language models, resulting in state-of-the-art performance across multiple benchmarks.
- •Developed novel attention mechanism architectures, reducing training time by 35% while improving model performance by 12%.
- •Mentored junior researchers and published 5 papers in top-tier conferences including NeurIPS, ICML, and ACL.
- •Collaborated with ethics team to implement responsible AI guidelines and reduce bias in language models.
AI Research Engineer
OpenAI
March 2020 - Sept 2022
San Francisco, CA
- •Contributed to the development of GPT-3, focusing on optimization techniques and training infrastructure for large-scale language models.
- •Implemented performance optimizations that reduced inference latency by 40% and improved training throughput by 25%.
- •Developed methods for evaluating model outputs and fine-tuning systems that significantly improved response quality and safety.
NLP Researcher
Google Brain
June 2017 - Feb 2020
Mountain View, CA
- •Worked on the original BERT team, contributing to architecture design and pre-training methodology for transformer-based models.
- •Developed techniques for knowledge extraction and commonsense reasoning that improved model performance on multiple NLP benchmarks.
- •Published research on attention mechanisms that has been cited over 1,500 times.
Education
Ph.D. in Machine Learning
Stanford University
2014 - 2017
Dissertation: "Attention Mechanisms for Neural Network Architectures" Advisor: Dr. Andrew Ng
Master of Science in Computer Science
Massachusetts Institute of Technology
2012 - 2014
Specialized in artificial intelligence and natural language processing. Research focus on neural network architectures for language understanding.
Bachelor of Science in Computer Science
University of California, Berkeley
2008 - 2012
Graduated summa cum laude. Honors thesis on statistical methods for machine translation. Minor in Mathematics.
Certifications
NVIDIA Deep Learning Institute Certified Instructor
NVIDIA • 2023
TensorFlow Certified Developer
Google • 2022
AWS Machine Learning Specialty
Amazon Web Services • 2021
Professional Certificate in AI Ethics
Harvard University • 2020