← Back to all jobs
I

Senior Data AI Engineer

Intellitech llc

7h ago

0DevRemote, USjobspy_indeed
remoteindeed

Job Description

**Location:** Remote **Clearance:** Active DoD Secret clearance required **Employment Type:** Full\-Time (W\-2\) **Citizenship:** U.S. Citizenship required IntelliTech is seeking a **Senior Data / AI Engineer** to support a Department of War program focused on operationalizing a Government\-owned digital twin application for ammunition industrial base readiness. The platform is a supply chain simulation solution built on **Python, FastAPI, and React** that enables analysts to model production timelines, identify bottlenecks, assess supply chain risk, and evaluate surge and modernization scenarios. This role will own the data lifecycle end\-to\-end—from raw file ingestion through validation, normalization, versioning, and delivery of run\-ready artifacts to the simulation engine. The engineer will also help design and implement the **AI\-enabled decision\-support layer**, supporting natural\-language analysis of scenario outputs, automated comparison and briefing generation, and guided scenario creation. This is a hands\-on role on a lean, senior team. The ideal candidate is comfortable writing production code daily, designing scalable data pipelines, and working directly with Government analysts and data stakeholders to deliver mission\-focused solutions. ### **Key Responsibilities** #### **Data Ingestion and Automation** * Design and implement governed ingestion pipelines for complex defense supply chain datasets, including Bills of Materials (BOM), demand and order backlogs, facility and production line capacity, supplier risk, and acquisition planning data. * Build validation services that enforce schema conformance, referential integrity across linked datasets, circular reference detection, and business\-rule validation with actionable row\- and column\-level feedback. * Implement raw data preservation in object storage such as **Amazon S3**, including metadata capture for source type, upload timestamp, uploader identity, file checksum, and dataset version. * Develop canonical data transformation workflows that convert validated source inputs into normalized, run\-ready artifacts aligned to the simulation engine’s entity model. * Implement dataset versioning and lineage tracking so each scenario run is tied to explicit input versions and assumptions. #### **Automated Data Refresh** * Work with Government stakeholders and source\-system owners to identify, prioritize, and implement automated or semi\-automated data refresh paths. * Participate in Technical Exchange Meetings (TEMs) to help define data contracts, including source format, semantics, refresh cadence, and validation requirements. * Implement approved connection patterns such as scheduled file landing, secure file exchange (SFTP), API\-based retrieval, and cloud\-to\-cloud transfer mechanisms. * Maintain hardened controlled upload workflows in parallel so mission operations are not dependent solely on external integrations or approvals. #### **AI\-Enabled Decisi