Search

Data Governance Engineer

Aleknagik Technology
locationWashington, DC, USA
PublishedPublished: 6/14/2022
Technology
Full Time

Job Description

Job Description

Aleknagik Technology, LLC is searching for a qualified candidate to fill the role Data Governance Engineer

The Data Governance Engineer is responsible for designing, building, and optimizing the pipelines, workflows, and technical infrastructure required to support DHA's enterprise data governance program. The Data Governance Engineer ensures that metadata, data products, and discovery services operate seamlessly within DHA's federated data mesh environment. This role directly supports DHA's Chief Data and Analytics Officer (CDAO) mission by enabling governed, automated, and scalable data flows that adhere to VAULTIS principles (Visible, Accessible, Understandable, Linked, Trusted, Interoperable, Secure).

Required Skills and Experience

  • Clearance: DoD SECRET
  • Education: Bachelor's degree in Computer Science, Information Systems, Data Engineering, or related field. Master's degree preferred.
  • Certifications: AWS Certified Data Analytics, Microsoft Azure Data Engineer, or similar certifications; DAMA CDMP desirable.
  • Experience:
    • 5+ years of experience in data engineering, ETL/ELT pipeline development, and data integration.
    • Demonstrated expertise with Microsoft Purview, AWS Glue, Redshift, and Oracle Analytics Cloud.
    • Proven experience implementing metadata harvesting and lifecycle governance solutions.
    • Familiarity with DoD cybersecurity standards, HIPAA, FISMA, and Zero Trust models.
  • Technical Skills: Strong proficiency in Python, SQL, Spark, and pipeline orchestration tools (Airflow, AWS Step Functions, or Azure Data Factory).
  • Soft Skills: Strong collaboration and problem-solving skills; ability to translate architectural requirements into engineered solutions.

Key Roles and Responsibilities

Metadata Management (SOO Obj. 3.1 / SOW Task 1)

  • Develop and maintain ETL/ELT pipelines that automatically capture, store, and update technical, business, and operational metadata.
  • Support the Enterprise Metadata Framework and Dictionary, ensuring quarterly updates meet ≥95% accuracy requirements.
  • Integrate Purview and AWS Glue catalog metadata into DHA's federated governance repository.

Business Process Automation (SOO Obj. 3.10 / SOW Task 2)

  • Engineer automated workflows for metadata ingestion, validation, approval, and publication.
  • Implement SLA-based monitoring ensuring ≥98% compliance for automated workflows.
  • Collaborate with Data Architects and Automation Engineers to ensure pipeline reliability, scalability, and interoperability across DHA domains.

Federated Governance Enablement (SOO Obj. 4 / SOW Task 3)

  • Build and sustain domain-level data pipelines that feed into the Federated Governance Playbook framework.
  • Ensure pipelines enforce access controls and comply with DoD Zero Trust and HIPAA/FISMA directives.
  • Provide technical input for workforce training on pipeline and workflow automation.

Metadata Harvesting and Discovery Services (SOO Obj. 5 / SOW Task 4)

  • Develop automated metadata harvesting pipelines across structured, unstructured, and streaming data environments.
  • Normalize harvested metadata into ISO/IEC 11179-compliant structures for discoverability.
  • Ensure discovery portals (Purview, Redshift, and integrated tools) are continuously updated with trusted, validated datasets.

Data Product Lifecycle Governance (SOO Obj. 6 / SOW Task 5)

  • Engineer solutions for data product lifecycle management including creation, transformation, versioning, and archival.
  • Build audit trails and lineage tracking pipelines to ensure transparency and compliance.
  • Support lifecycle playbooks with operational metadata metrics on performance and utilization.

Self-Service & End User Enablement (SOO Obj. 10 / SOW Task 6)

  • Deliver secure, role-based pipelines and APIs that make curated data products accessible for self-service portals.
  • Ensure ad-hoc queries and dashboards are powered by continuously refreshed and validated pipelines.
  • Support user feedback-driven iterations to enhance discoverability, performance, and security of self-service capabilities.

Contribution to DHA Deliverables (CDRL Alignment)

  • Enterprise Metadata Framework & Dictionary – Builds and maintains pipelines that capture and update metadata.
  • Automation Implementation Plan & Workflow Demonstration – Engineers pipelines and validates workflow automation performance.
  • Federated Governance Playbook – Provides pipeline architecture supporting federated metadata synchronization.
  • Quarterly Metadata Harvesting & Discovery Reports – Supplies technical input and operational metrics from harvesting pipelines.
  • Data Product Lifecycle Playbook – Delivers lineage, version control, and archival pipeline support.
  • Semi-Annual Self-Service Evaluation Reports – Ensures curated data pipelines enable discoverability and accessibility metrics.
  • Final Executive Report & Briefing – Provides technical evidence of pipeline performance, SLA adherence, and governance compliance.
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...