Capital One’s Summer 2024 Applied Research PhD Internship Program
Students interested in the Summer 2024 Applied Research PhD Internship Program can apply here.
This is a paid internship. This is a limited-time internship position, and Capital One will not sponsor a new applicant for employment authorization for this position. However, a full-timeApplied Researchrole, for which you may be considered upon completion of the internship (subject to business need, market conditions, and other factors) is eligible for employer immigration sponsorship.
Team Description:
The AI Foundations team is at the center of bringing our vision for AI at Capital One to life. Our work touches every aspect of the research life cycle, from partnering with Academia to building production systems. We work with product, technology and business leaders to apply the state of the art in AI to our business.
In this role, you will:
- Join Capital One for a full-time, 12 week, summer applied research experience, discovering solutions to real world, large-scale problems.
- Engage in high impact applied research with the goal of taking the latest AI developments and pushing them into the next generation of customer experiences, or contributing to publications in this field.
- Partner with a cross-functional team of applied researchers, data scientists, software engineers, machine learning engineers and product managers to test and design AI- powered products that change how customers interact with their money.
- Leverage a broad stack of technologies — Pytorch, AWS Ultraclusters, Huggingface, Lightning, VectorDBs, and more — to reveal the insights hidden within huge volumes of numeric and textual data.
- Flex your interpersonal skills to translate the complexity of your work into tangible business goals.
- Partner with leading researchers to publish papers at top academic conferences.
- Develop Professionally through networking sessions, technical deep dives and executive speaker sessions from across Capital One.
The Ideal Candidate:
- You love the process of analyzing and creating, but also share our passion to do the right thing. You want to work on problems that will help change banking for good.
- Innovative. You continually research and evaluate emerging technologies. You stay current on published state-of-the-art methods, technologies, and applications and seek out opportunities to apply them.
- Creative. You thrive on bringing definition to big, undefined problems. You love asking questions and pushing hard to find answers. You’re not afraid to share a new idea.
- Technical. You possess a strong foundation in mathematics, deep learning theory, and the engineering required for contributing to the development of AI.
- Determined. Strengthen your field of study by applying theory to practice. Bring your ideas to life in industry.
Basic Qualifications:
- Currently enrolled in an accredited PhD Program
- Completed 2nd year of PhD coursework by program start date
Preferred Qualifications:
- Completed 3rd or 4th year of PhD Program
- PhD in Computer Science, Machine Learning, Computer Engineering, Applied Mathematics, Electrical Engineering or related fields
- Programming experience in Python, PyTorch, C++, and other deep learning frameworks
- Publications in leading conferences such as KDD, ICML, NeurIPs, ICLR, ACL, NAACL and EMNLP, or ICLR
- Focused area of researchin one of the following areas:
- LLM Pre-training
- PhD focus on Natural Language Processing
- Publications on topics related to the pre-training of large language models (e.g. technical reports of pre-trained LLMs, SSL techniques, model pre-training optimization)
- Publications in deep learning theory
- LLM Finetuning
- PhD focused on topics related to guiding LLMs with further tasks (Supervised Finetuning, Instruction-Tuning, Dialogue-Finetuning, Parameter Tuning)
- Demonstrated knowledge of principles of transfer learning, model adaptation and model guidance
- Behavioral Models
- PhD focus on topics in geometric deep learning (Graph Neural Networks, Sequential Models, Multivariate Time Series)
- Contributions to common open source frameworks (pytorch-geometric, DGL)
- Proposed new methods for inference or representation learning on graphs or sequences
- Optimization (Training & Inference)
- PhD focused on topics related to optimizing training of very large deep learning models
- Multiple years of experience and/or publications on one of the following topics: Model Sparsification, Quantization, Training Parallelism/Partitioning Design, Gradient Checkpointing, Model Compression
- Deep knowledge of deep learning algorithmic and/or optimizer design
- Large Scale Data Preparation
- Publications studying tokenization, data quality, dataset curation, or labeling