Reddit, Inc.
Dive into anything
Staff Research Engineer, Pre-training Science
Research EngineerResearch EngineerFull TimeRemoteTeam 501-1,000Since 2005H1B No SponsorCompany SiteLinkedIn
Location
United States
Posted
48 days ago
Salary
$230K - $322K / year
Postgraduate Degree7 yrs expEnglishAWSDistributed SystemsPythonPy Torch
Job Description
• Architect and validate rigorous Continual Pre-Training (CPT) frameworks, focusing on domain adaptation techniques that effectively transfer Reddit’s knowledge into licensed frontier models.
• Design the "Science of Multimodality": Lead research into fusing vision and language encoders to process Reddit’s rich media (images, video) alongside conversational text threads.
• Formulate data curriculum strategies: scientifically determining the optimal ratio of "Reddit data" vs. "General data" to maximize community understanding while maintaining safety and reasoning capabilities.
• Conduct deep-dive research into Scaling Laws for Graph-based data: investigating how Reddit’s tree-structured conversations impact model convergence compared to flat text.
• Design and scale continuous evaluation pipelines (the "Reddit Gym") that monitor model reasoning and safety capabilities in real-time, enabling dynamic adjustments to training recipes.
• Drive high-stakes architectural decisions regarding compute allocation, distributed training strategies (3D parallelism), and checkpointing mechanisms on AWS Trainium/Nova clusters.
• Serve as a force multiplier for the engineering team by setting coding standards, conducting high-level design reviews, and mentoring senior engineers on distributed systems and ML fundamentals.
Job Requirements
- 7+ years of experience in Machine Learning engineering or research, with a specific focus on LLM Pre-training, Domain Adaptation, or Transfer Learning.
- Expert-level proficiency in Python and deep learning frameworks (PyTorch or JAX), with a track record of debugging complex training instabilities at scale.
- Deep theoretical understanding of Transformer architectures and Pre-training dynamics—specifically regarding Catastrophic Forgetting and Knowledge Injection.
- Experience with Multimodal models (VLM): understanding how to align image/video encoders (e.g., CLIP, SigLIP) with language decoders.
- Experience implementing continuous integration/evaluation systems for ML models, measuring generalization and reasoning performance.
- Demonstrated ability to communicate complex technical concepts (like loss spikes or convergence issues) to leadership and coordinate efforts across Infrastructure and Data teams.
Benefits
- Comprehensive Healthcare Benefits and Income Replacement Programs
- 401k with Employer Match
- Global Benefit programs that fit your lifestyle, from workspace to professional development to caregiving support
- Family Planning Support
- Gender-Affirming Care
- Mental Health & Coaching Benefits
- Flexible Vacation & Paid Volunteer Time Off
- Generous Paid Parental Leave
Related Guides
Related Categories
Related Job Pages
More Research Engineer Jobs
Senior Operations Research Engineer
Lawrence Livermore National LaboratoryScience and technology on a mission
Research Engineer52 days ago
Full TimeRemoteTeam 5,001-10,000Since 1952H1B No Sponsor
Senior Operations Research Engineer developing optimization models for energy systems
Research Engineer53 days ago
Full TimeRemoteTeam 11-50Since 2023H1B No Sponsor
Research Engineer developing verification tools for Code Metal
PythonRust
Massachusetts
Research Engineer54 days ago
Part TimeRemote
We are seeking a detail-oriented intern to provide administrative, content, and quality improvement support for grant-funded educational events. This role is ideal for a highly organized individual with strong proofreading skills who is comfortable working independently in a remo...
United States
Research Engineer61 days ago
Full TimeRemoteTeam 201-500Since 2012H1B No Sponsor
Exploit Development Specialist designing malware defenses for Bugcrowd's security platform.
AssemblyPythonTypeScript