Job Url: https://www.remoterocketship.com/company/spiralscout/jobs/ai-data-engineer-united-states-remote/ Job Description: Spiral Scout Website LinkedIn All Job Openings IT Outstaffing β€’ Outsourcing β€’ Staff Augmentation β€’ Software Development β€’ Design Spiral Scout is a technology partner that specializes in AI-powered software development and digital transformation. They build agile teams to deliver industry-leading AI solutions for startups and enterprises, focusing on AI-driven software innovation, agile development, and on-demand tech expertise. With a history of successful projects, Spiral Scout helps businesses improve efficiency, optimize legacy systems, and streamline operations through custom AI frameworks such as Wippy. ai. Their services extend to enhancing eCommerce platforms, educational tech development, and integrating payment systems. Serving over 300 companies, Spiral Scout positions itself as a reliable partner for mid-sized companies and startups looking to scale their technology capabilities, modernize their systems, or require premium tech talent quickly. 51 - 200 employees πŸ€– Artificial Intelligence ☁️ SaaS 🏒 Enterprise AI Data Engineer October 8 πŸ‡ΊπŸ‡Έ United States – Remote ⏳ Contract/Temporary 🟑 Mid-level 🟠 Senior 🚰 Data Engineer AWS Azure Cloud ETL Google Cloud Platform Python SQL Apply Now Receive Emails with Similar Jobs Report problem πŸ“‹ Description β€’ Build and maintain data pipelines (Python, SQL, ETL, APIs) to prepare structured and unstructured data for AI workflows. β€’ Translate client problems into orchestrated AI workflows, balancing automation and human-in-the-loop design. β€’ Configure multi-agent logic (planner/worker, feedback loops) using LangChain, Wippy, n8n, Zapier, Make, or custom Python code. β€’ Prototype and ship proof-of-concepts: onboarding bots, quoting assistants, presales flows, project management helpers. β€’ Facilitate scoping workshops with stakeholders to clarify requirements and design workflows. β€’ Collaborate with engineers and product leads to create reusable AI workflow templates and automation patterns. 🎯 Requirements β€’ Strong data engineering experience: Python + SQL, pipelines, ETL, API integrations. β€’ Experience working with data ahead of AI: cleaning, structuring, connecting multiple sources. β€’ Hands-on with AI tools: ChatGPT, Claude, LangChain, n8n, Zapier, Make, Autogen, or similar. β€’ Understanding of LLM orchestration beyond prompt engineering. β€’ Systems thinking: ability to design workflows with multiple agents, branching logic, loops, and state. β€’ Strong communication skills: able to explain AI/data concepts and lead workshops with technical and non-technical stakeholders. β€’ Experience with agent-based frameworks (LangGraph, CrewAI, AutoGen). β€’ Designed human-in-the-loop workflows (customer support, onboarding, quoting, project management). β€’ Prototyping with low-code/no-code platforms (Zapier, Airtable, Streamlit, custom GPTs). β€’ Familiarity with cloud platforms (AWS, GCP, Azure). β€’ Exposure to ML practices (fine-tuning, RAG, evaluation, multimodal inputs). πŸ–οΈ Benefits β€’ Start with a paid pilot project to evaluate collaboration. β€’ Work at the cutting edge of AI workflows and data-driven automation. β€’ Fully remote, flexible schedule. β€’ Fast-moving, innovation-driven culture where ideas quickly turn into practice. β€’ Pathway to extended part-time or full-time engagement.