Search

Solutions Architect-L3

ZAPCOM GROUP INC
locationIrving, TX, USA
PublishedPublished: 6/14/2022
Technology
Full Time

Job Description

Job DescriptionAs a Solution Architect (L3), you will be the technical authority bridging business requirements and engineering execution. You will design end-to-end data and application architectures on Microsoft Azure, leveraging Databricks for large-scale data engineering and Power BI for enterprise analytics and reporting. You will lead cross-functional technical discussions, define data platform strategy, and ensure that solutions are scalable, resilient, secure, and aligned with organisational technology goals. This is a high-impact senior individual contributor role requiring deep cloud and data engineering expertise alongside strong stakeholder communication skills.



Key Responsibilities

Solution Architecture & Design

  • Define and own end-to-end solution architecture for complex, data-intensive, multi-service systems on Microsoft Azure.
  • Create and maintain architecture blueprints, HLDs, LLDs, data flow diagrams, and Architecture Decision Records (ADRs).
  • Evaluate build-vs-buy decisions and drive technology selection with clear trade-off analysis across the Azure ecosystem.
  • Ensure solutions satisfy non-functional requirements: performance, scalability, reliability, data freshness SLAs, and security.

Data Engineering & Platform Architecture

  • Architect end-to-end data pipelines using Azure Databricks spanning ingestion, transformation (Bronze / Silver / Gold layers), and serving.
  • Design and govern data lakehouse solutions using Azure Data Lake Storage Gen2 (ADLS Gen2) and Delta Lake.
  • Define data modelling standards (star schema, data vault, OBT) aligned to business reporting requirements in Power BI.
  • Architect real-time and batch ingestion patterns using Azure Event Hubs, Azure Data Factory (ADF), and Databricks Structured Streaming.
  • Drive adoption of Databricks Unity Catalog for data governance, lineage tracking, and fine-grained access control.
  • Design cost-efficient Databricks cluster strategies (job clusters, instance pools, spot instances) and monitor DBU consumption.
  • Establish data quality frameworks using Great Expectations, Databricks Lakehouse Monitoring, or equivalent tooling.

Analytics & Reporting Architecture

  • Design scalable semantic layer and data models powering Power BI enterprise reports and dashboards.
  • Define DirectQuery vs Import vs Composite model strategies for optimal Power BI performance and data freshness.
  • Architect Power BI Premium / Fabric workspace strategies, including capacity planning and deployment pipelines (Dev / Test / Prod).
  • Guide teams on DAX optimisation, row-level security (RLS), and Power BI paginated reports for operational reporting.
  • Integrate Power BI with Azure Databricks via Partner Connect or Lakehouse connector for live analytics on Delta tables.

Azure Cloud & Infrastructure

  • Architect cloud-native solutions leveraging Azure services: Azure Synapse Analytics, Azure Purview, Azure Key Vault, Azure Monitor, and APIM.
  • Define Infrastructure-as-Code (IaC) standards using Terraform or Bicep for repeatable, auditable deployments.
  • Design Azure networking topology: VNets, Private Endpoints, NSGs, and hub-spoke patterns to secure data flows.
  • Guide teams on containerisation (Docker, Azure Kubernetes Service) and Databricks-native MLflow for ML model management.
  • Define disaster recovery, backup, and high-availability strategies across the Azure data platform stack.

Technical Leadership

  • Serve as the primary technical point of contact across data engineering, analytics, product, and business teams.
  • Provide technical mentorship to senior data engineers, analytics engineers, and tech leads.
  • Lead architecture reviews, proof-of-concepts, and data platform migration spikes.
  • Drive engineering best practices: modular pipeline design, idempotency, schema evolution, and CI/CD for data workflows.

Stakeholder Engagement & Governance

  • Partner with data owners, business analysts, and C-suite stakeholders to translate analytical requirements into robust data products.
  • Present platform architecture, roadmaps, and data strategy to executive leadership with clarity and confidence.
  • Establish data governance frameworks covering cataloguing (Azure Purview), data classification, and lineage.
  • Ensure compliance with data privacy regulations (PDPB, GDPR) and security standards (ISO 27001, SOC 2) across the data platform.



Required Qualifications

Experience

  • 8 12 years of overall IT experience, with at least 4 years in a Solution Architect, Principal Data Engineer, or Senior Data Architect role.
  • Proven track record of architecting and delivering large-scale data platforms on Microsoft Azure in production.
  • Hands-on experience with Azure Databricks including workspace administration, cluster management, Delta Live Tables, and Unity Catalog.
  • Strong experience delivering Power BI enterprise analytics solutions semantic modelling, RLS, Premium capacities, and deployment pipelines.
  • Experience in Agile / SAFe delivery environments with strong DataOps and CI/CD practices for data pipelines.

Azure Data Platform Proficiency

  • Deep expertise in Azure Data Lake Storage Gen2, Delta Lake, and the Medallion (Bronze / Silver / Gold) architecture pattern.
  • Proficient in Azure Data Factory for pipeline orchestration, and Azure Event Hubs / Azure Service Bus for streaming ingestion.
  • Strong knowledge of Azure Synapse Analytics, Azure SQL Database / Managed Instance, and Cosmos DB.
  • Hands-on with Azure Purview (Microsoft Purview) for data cataloguing, classification, and lineage.
  • Experience with Azure Key Vault, Azure Active Directory / Entra ID, and Azure Private Link for platform security.
  • Familiarity with Azure Monitor, Log Analytics, and Databricks monitoring integrations for observability.

Databricks & Data Engineering

  • Expert-level Spark programming in PySpark and/or Scala for large-scale batch and streaming transformations.
  • Deep knowledge of Delta Lake features: ACID transactions, time travel, Z-ordering, liquid clustering, and schema enforcement.
  • Experience with Databricks Workflows (formerly Jobs), Delta Live Tables (DLT), and Databricks Asset Bundles for CI/CD.
  • Familiarity with Databricks MLflow for experiment tracking and model registry (ML workloads are a plus, not mandatory).
  • Ability to design cost-optimised cluster policies, autoscaling strategies, and spot-instance configurations.

Power BI & Analytics

  • Strong Power BI development skills: data modelling (star schema), DAX measures, calculated columns, and report/dashboard design.
  • Experience with Power BI Premium / Fabric capacity management, paginated reports, and dataflows Gen2.
  • Proficiency in defining composite model strategies (DirectQuery + Import) and optimising query performance with aggregations.
  • Knowledge of Power BI REST APIs, XMLA endpoints, and Tabular Editor for advanced semantic model management.

Soft Skills

  • Exceptional written and verbal communication ability to author crisp architecture documents and present confidently to diverse audiences.
  • Strong analytical and problem-solving mindset with the ability to navigate ambiguity and competing priorities.
  • Collaborative leader who can influence without authority across data engineering, analytics, product, and business stakeholders.

Education & Certifications

  • E. / B.Tech / M.Tech in Computer Science, Information Technology, or a related field (or equivalent industry experience).
  • Microsoft Certified: Azure Data Engineer Associate (DP-203) required.
  • Databricks Certified Associate / Professional Data Engineer strongly preferred.
  • Microsoft Certified: Azure Solutions Architect Expert (AZ-305) or Power BI Data Analyst (PL-300) a strong plus.



Preferred Qualifications

  • Experience with Microsoft Fabric (OneLake, Fabric Data Engineering, Real-Time Analytics) the strategic evolution of the Azure data platform.
  • Exposure to dbt (data build tool) for SQL-based transformation layers and analytics engineering workflows.
  • Familiarity with Apache Iceberg or Apache Hudi as alternative open table formats alongside Delta Lake.
  • Prior experience with Azure Machine Learning or Databricks Model Serving for operationalising ML models.
  • Background in FinTech, HealthTech, Retail, or Manufacturing verticals with complex regulatory data requirements.
  • Contributions to open-source data projects or published technical content (blogs, conference talks, white papers).
  • Exposure to data mesh principles and federated data ownership models.

Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...
Loading...