Position Expired
This job is no longer accepting applications.
AI LLM Scientist
RxGPT
We’re looking for an AI LLM Scientist who can push the boundaries of model intelligence, reasoning, and agentic capabilities. If you enjoy diving deep into transformer architectures, experimenting with frontier models, optimizing training pipelines, and prototyping novel ways LLMs can plan, act, and collaborate—this role will feel like home.
You’ll play a key part in shaping our next generation of AI-driven products: smarter agents, efficient pipelines, adaptive workflows, and robust AI behaviors built on top of cutting-edge research.
What You’ll Work On
- Core LLM Research & Development
- Experiment with foundational models (e.g., GPT, Claude, Llama, Mixtral, Phi).
- Fine-tune, supervise, distill, and reinforce models for specialized tasks.
- Improve reasoning, planning, and tool-use capabilities of agentic workflows.
- Build evaluation frameworks for multi-step logic, safety, and reliability.
- Research and deploy optimization techniques: LoRA, QLoRA, 4-bit quantization, SFT, RLHF/RLAIF, preference modeling, chain-of-thought optimization.
- Agentic AI & System Design
- Architect scalable pipelines for multi-agent or tool-augmented systems.
- Prototype autonomous behaviors, memory systems, retrieval workflows, and reasoning chains.
- Integrate external tools/APIs, knowledge bases, and embeddings for enhanced context.
- LLM Engineering
- Deploy models efficiently using GPU/TPU stacks.
- Optimize latency, throughput, and cost for production workloads.
- Experiment with model compression, routing, and hybrid LLM ensembles.
- Research + Product Collaboration
- Translate complex AI research into practical, reliable features.
- Work closely with engineers, product teams, and designers to ship experimental ideas.
- Stay ahead of the AI frontier and propose new directions for the team.
What You Should Bring
Strong understanding of LLMs, transformers, embeddings, and NLP fundamentals.
Hands-on experience with training/fine-tuning frameworks: PyTorch, JAX, Hugging Face, DeepSpeed, Ray.
Experience building or optimizing LLM-powered applications and agents.
Ability to read, interpret, and apply research papers (NeurIPS, ICML, ICLR, ACL, etc.).
Solid grasp of distributed systems, model optimization, or inference engineering.
Bonus points for
Publications or open-source research contributions
Experience training smaller foundation models
Knowledge of RLHF pipelines
Exposure to retrieval systems, vector DBs, or tool-use frameworks
Familiarity with frontier safety research
Note: This is an unpaid opportunity focused on skill-building, experience, and professional growth.
Job Alerts
Get notified when new positions matching your interests become available at {organizationName}.
Need Help?
Questions about our hiring process or want to learn more about working with us?