Job Url: https://www.linkedin.com/jobs/search/?currentJobId=4359024530&distance=25.0&f_AL=true&f_TPR=r86400&f_WT=2&geoId=103644278&keywords=software%20engineer&origin=JOB_SEARCH_PAGE_JOB_FILTER&start=300 Job Description: Founding AI/ML Research Engineer United States · 6 hours ago · 40 applicants Promoted by hirer · No response insights available yet Remote Matches your job preferences, workplace type is Remote. Full-time Matches your job preferences, job type is Full-time. Easy Apply Save Save Founding AI/ML Research Engineer at A1 Your profile is missing required qualifications Show match details Help me update my profile BETA Is this information helpful? Get personalized tips to stand out to hirers Find jobs where you’re a top applicant and tailor your resume with the help of AI. Try Premium for PKR0 About the job About A1 A1 is a self-funded, independent AI group, focused on building a new consumer AI application with global impact. We’re assembling a small, elite team of ML, engineering and product builders who want to work on meaningful, high-impact problems. About The Role You will shape the core technical direction of A1 - model selection, training strategy, infrastructure, and long-term architecture. This is a founding technical role: your decisions will define our model stack, our data strategy, and our product capabilities for years ahead. You won’t just fine-tune models - you’ll design systems: training pipelines, evaluation frameworks, inference stacks, and scalable deployment architectures. You will have full autonomy to experiment with frontier models (LLaMA, Mistral, Qwen, Claude-compatible architectures) and build new approaches where existing ones fall short. What You’ll be Doing Build end-to-end training pipelines: data → training → eval → inference Design new model architectures or adapt open-source frontier models Fine-tune models using state-of-the-art methods (LoRA/QLoRA, SFT, DPO, distillation) Architect scalable inference systems using vLLM / TensorRT-LLM / DeepSpeed Build data systems for high-quality synthetic and real-world training data Develop alignment, safety, and guardrail strategies Design evaluation frameworks across performance, robustness, safety, and bias Own deployment: GPU optimization, latency reduction, scaling policies Shape early product direction, experiment with new use cases, and build AI-powered experiences from zero Explore frontier techniques: retrieval-augmented training, mixture-of-experts, distillation, multi-agent orchestration, multimodal models What You'll Need Strong background in deep learning and transformer architectures Hands-on experience training or fine-tuning large models (LLMs or vision models) Proficiency with PyTorch, JAX, or TensorFlow Experience with distributed training frameworks (DeepSpeed, FSDP, Megatron, ZeRO, Ray) Strong software engineering skills — writing robust, production-grade systems Experience with GPU optimization: memory efficiency, quantization, mixed precision Comfortable owning ambiguous, zero-to-one technical problems end-to-end Nice to Have Experience with LLM inference frameworks (vLLM, TensorRT-LLM, FasterTransformer) Contributions to open-source ML libraries Background in scientific computing, compilers, or GPU kernels Experience with RLHF pipelines (PPO, DPO, ORPO) Experience training or deploying multimodal or diffusion models Experience in large-scale data processing (Apache Arrow, Spark, Ray) Prior work in a research lab (Google Brain, DeepMind, FAIR, Anthropic, OpenAI)