i4DM Logo

i4DM

C2 SMART Integration Engineer (Data Tagging and Availability)

Posted Yesterday
Be an Early Applicant
Remote
Hiring Remotely in USA
Senior level
Remote
Hiring Remotely in USA
Senior level
The Integration Engineer focuses on building data tagging frameworks and integrating workflows for data governance and management in complex environments.
The summary above was generated by AI

About Our Team

Our employees thrive in a culture that's fast-paced and ego-free, where innovation and collaboration are encouraged at every turn. We are an organization that provides federal agencies and commercial clients instant access to experienced and talented professionals who understand their unique challenges and know the most efficient ways to address them. We are continually investing in resources and talent, so we stay prepared with specialized teams in the place who are experts in creating tailored technologies. Our solutions empower our clients to grow, modernize, and succeed in a rapidly evolving landscape.

We value all voices and want to attract talent from all backgrounds. We are on the lookout for individuals who are passionate about technology and thrive in environments where problem-solving is approached with creativity and enthusiasm. If you are someone who enjoys continuously expanding your skill set while tackling real-world business problems, you will feel right at home with us. Veterans and military spouses are especially encouraged to bring your unique and valuable experience to our team.

About the Role

Are you ready to help enable data‑driven operations at the Army’s tactical edge? We are seeking an Integration Engineer with strong expertise in data tagging, metadata management, workflow orchestration, Apache Kafka, and cross‑domain integration. In this role, you’ll build and maintain the integration layer that supports automated and AI‑assisted data workflows—ensuring data is properly tagged, governed, and reliably shared across domains, including disconnected and bandwidth‑constrained environments. Your work will support a data mesh or data fabric architecture where data products are discoverable, secure, and interoperable, helping operators and analysts access trusted information when and where it’s needed.

Responsibilities

  • Design and implement automated data tagging frameworks that attach business, technical, and operational metadata to data assets at ingestion time.
  • Integrate with data catalogs to programmatically populate and maintain metadata, including:
  • Business glossaries and term definitions
  • Data classification (PII, sensitive, confidential)
  • Domain ownership and stewardship
  • Data quality scores and lineage
  • Build pipelines that extract metadata from source systems (databases, Kafka schemas, file formats) and synchronize with enterprise metadata repositories.
  • Implement tagging policies that propagate across domains, ensuring data assets are consistently labeled for discoverability, access control, and retention.
  • Design and implement governed data workflows that enforce approval gates, validation checks, and compliance requirements before data is published to consumers.
  • Build workflow automation using tools like Apache Airflow, Prefect, Dagster, or cloud-native workflow services (AWS Step Functions, Azure Logic Apps).
  • Integrate workflow engines with data catalogs and tagging systems to trigger actions based on metadata changes (e.g., when a dataset is tagged as "sensitive," automatically apply encryption and restrict access).
  • Implement SLA monitoring and alerting for workflow completion, data freshness, and compliance checks.
  • Design and implement integration patterns that enable secure, governed data flow across multiple domains.
  • Implement cross-domain service architectures using APIs, event streaming, and data virtualization.
  • Define and manage data contracts between domains, specifying schemas, SLAs, quality requirements, and tagging expectations.
  • Architect, deploy, and manage Apache Kafka clusters across multiple domains and environments (on-premise, cloud, hybrid).
  • Implement streaming workflows where Kafka messages trigger governed workflows (e.g., new data arrival initiates validation and tagging pipeline).
  • Design canonical data models that serve as the standard for cross-domain data exchange, embedding tags and metadata into the model structure.
  • Collaborate with domain experts to define business terms, hierarchies, and metrics that are consistently tagged and governed across domains.
  • Implement tag-based access control (TBAC) where data access policies are enforced based on tags applied to datasets.
  • Ensure compliance with regulatory requirements (GDPR, CCPA, SOX) through automated tagging of sensitive data and workflow-enforced retention/deletion policies.
  • Build data lineage that captures tagging events and workflow approvals, providing end-to-end visibility into how data is governed.
  • Implement data quality workflows where datasets must pass quality checks before being tagged as "certified" or "trusted."
  • Set up alerting for workflow failures, tagging inconsistencies, schema drift, and cross-domain connectivity issues.
  • Document tagging schemas, workflow definitions, and integration patterns for operational handoff.

TAG: #LI-I4DM

Required Qualifications:

  • Bachelor’s or Master’s degree in Computer Science, Information Systems, Engineering, or a related field.
  • 5+ years of experience in data engineering, data integration, or software engineering with a focus on data governance, metadata management, and enterprise integration.
  • Proven experience implementing automated data tagging frameworks at enterprise scale.
  • Hands-on experience with data catalog tools (Alation, Collibra, DataHub, or Amundsen) including API integration and metadata synchronization.
  • Understanding of metadata standards (DCAT, W3C PROV, OpenLineage) and semantic web concepts.
  • Experience with data classification and sensitive data detection (PII, PHI, PCI).
  • Experience with workflow orchestration tools:
  • Apache Airflow with custom operators, sensors, and DAG design
  • Alternatives: Prefect, Dagster, AWS Step Functions, Azure Logic Apps
  • Experience building governed workflows with approval gates, validation steps, and audit trails.
  • Familiarity with event-driven workflows triggered by Kafka messages or metadata changes.
  • Proven experience designing and operating cross-domain data integration architectures in large enterprises.
  • Understanding of data mesh principles and data product orientation with tagging as a core component.
  • Production experience with Apache Kafka, including:
  • Kafka cluster administration
  • Kafka Streams or ksqlDB
  • Experience embedding metadata and tags within Kafka messages or schema annotations.
  • Deep experience with enterprise data modeling across multiple domains.
  • Proficiency with data modeling tools (ERwin, ER/Studio, SAP PowerDesigner).
  • Experience embedding business tags, classifications, and governance attributes into physical and logical data models.
  • Advanced proficiency for custom tagging scripts, workflow automation, and Kafka integration.
  • Experience with Kafka client libraries and stream processing applications.
  • Expert-level for metadata querying, validation, and lineage extraction.
  • Deep experience with AWS (MSK, ECS, Lambda, S3, Glue, IAM) or Azure (Event Hubs, Data Factory, Synapse, Purview).
  • Docker, Kubernetes, Helm for deploying workflow and streaming applications.
  • Git and CI/CD pipelines (GitHub Actions, GitLab CI, Jenkins).     
  • Infrastructure as Code: Terraform, AWS CloudFormation, or Azure Resource Manager.

Top Skills

Apache Airflow
Apache Kafka
Aws Cloudformation
Aws Step Functions
Azure Logic Apps
Ci/Cd
Dagster
Docker
Git
Kubernetes
Prefect
Terraform

Similar Jobs

An Hour Ago
Remote or Hybrid
105K-130K Annually
Senior level
105K-130K Annually
Senior level
eCommerce • Fashion • Other • Retail • Sales • Wearables • Design
Leverage HR proficiency to manage employee lifecycle, provide guidance on employee relations, support talent management, lead organizational change, and ensure operational HR excellence.
Top Skills: ExcelMicrosoft PowerpointMicrosoft TeamsMicrosoft VisioMicrosoft WordWorkday
An Hour Ago
Remote or Hybrid
South Coast, CA, USA
19-38 Hourly
Junior
19-38 Hourly
Junior
eCommerce • Fashion • Other • Retail • Sales • Wearables • Design
As an Assistant Store Manager, you'll support sales goals, deliver exceptional customer service, lead team development, and ensure operational excellence.
Top Skills: MS OfficeRetail Systems
13 Hours Ago
Remote or Hybrid
209K-286K Annually
Senior level
209K-286K Annually
Senior level
Fintech • Machine Learning • Payments • Software • Financial Services
The Premium Support Engineer will serve as a technical authority for enterprise customers, ensuring optimal performance of air-gapped systems, advocating for feature requests, and managing incident responses.
Top Skills: AWSAws CloudformationDatabricksDockerEksGoHelmJavaKubernetesPrometheusPysparkPythonSnowflakeTerraform

What you need to know about the Colorado Tech Scene

With a business-friendly climate and research universities like CU Boulder and Colorado State, Colorado has made a name for itself as a startup ecosystem. The state boasts a skilled workforce and high quality of life thanks to its affordable housing, vibrant cultural scene and unparalleled opportunities for outdoor recreation. Colorado is also home to the National Renewable Energy Laboratory, helping cement its status as a hub for renewable energy innovation.

Key Facts About Colorado Tech

  • Number of Tech Workers: 260,000; 8.5% of overall workforce (2024 CompTIA survey)
  • Major Tech Employers: Lockheed Martin, Century Link, Comcast, BAE Systems, Level 3
  • Key Industries: Software, artificial intelligence, aerospace, e-commerce, fintech, healthtech
  • Funding Landscape: $4.9 billion in VC funding in 2024 (Pitchbook)
  • Notable Investors: Access Venture Partners, Ridgeline Ventures, Techstars, Blackhorn Ventures
  • Research Centers and Universities: Colorado School of Mines, University of Colorado Boulder, University of Denver, Colorado State University, Mesa Laboratory, Space Science Institute, National Center for Atmospheric Research, National Renewable Energy Laboratory, Gottlieb Institute

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account