Software Engineer
Aptonet
Job Description
Software Engineering III – Databricks Data Engineer (Remote | U.S. Citizenship Required) Remote – quarterly travel to Gaithersburg, MD We are seeking a highly skilled Software Engineering III – Databricks Data Engineer to join our growing team and play a key role in building scalable, secure, and high-performing data solutions within an enterprise Microsoft Azure cloud environment . This role is ideal for a hands-on Data Engineer with strong expertise in Databricks, Apache Spark, Azure data services, and modern data lake architecture .
You will be responsible for designing, developing, and optimizing enterprise-grade data pipelines, supporting advanced analytics and AI-driven workflows, and collaborating with both technical and business stakeholders to deliver impactful data solutions. If you enjoy solving complex data engineering challenges, working with large-scale datasets, and contributing to mission-critical cloud analytics platforms, this is a great opportunity to make a meaningful impact. Key Responsibilities Design, develop, and optimize scalable data pipelines and analytics solutions using Databricks and Medallion Architecture Build and maintain data ingestion pipelines for structured, streaming, and unstructured data Manage and support ingestion routines for multi-terabyte datasets across multiple Databricks workspaces Translate business and technical requirements into robust Azure-based data engineering solutions Integrate data from multiple enterprise sources to enable high-quality analytics, reporting, and insights Maintain and enhance ETL/ELT processes with a strong focus on data quality, reliability, and performance Monitor platform health, troubleshoot production issues, and optimize Spark workloads, cluster performance, and platform stability Implement data management strategies to ensure data integrity, availability, accessibility, and governance Identify and drive cost optimization opportunities across data storage, compute, and analytics workloads Integrate and work with key Azure services , including: Provision and manage cloud infrastructure using Infrastructure-as-Code (IaC) Apply best practices for data security, compliance, governance, and federal/Public Trust standards Partner with cross-functional stakeholders, including business users, architects, analysts, and engineering teams Support AI/ML and advanced analytics workflows by enabling reliable and scalable data foundations Provide client-facing support , technical consultation, and clear solution documentation Create and maintain architecture diagrams, interface diagrams, and technical documentation Mentor and review the work of junior technical team members as needed Required Technical Skills Strong hands-on experience with Databricks platform development, administration, and maintenance Expertise in Apache Spark Proficiency in Python and Spark (R experience is a plus) Strong experience with Azure cloud services , including: User / Identity Management Experience building data pipelines for structured, streaming, and unstructured data Strong understanding of ETL/ELT development , orchestration, and data quality management Deep knowledge of Medallion Architecture and modern data lake / lakehouse design patterns Experience with Infrastructure-as-Code (IaC) Knowledge of data governance , including: Metadata management Data security / access controls Experience with CI/CD automation for data engineering workflows Familiarity with Agile / Scrum methodologies Exposure to cloud platforms such as Azure (required) and AWS (preferred) Preferred / Nice-to-Have Skills Experience with .NET development Azure cloud certifications Understanding of FinOps / cloud cost management principles Exposure to AI/ML platforms, advanced analytics, or modern AI technologies Qualifications & Experience Bachelor’s degree in Computer Science, Engineering, or related field with 3+ years of relevant experience , OR Master’s degree in a related field with 2+ years of relevant experience Databricks Data Engineer Certification strongly preferred / required 2+ years of hands-on experience supporting and maintaining the Databricks platform and Spark-based development 3+ years of experience designing and developing cloud-based data ingestion and processing pipelines Proven experience working directly with clients / end users and providing front-line technical support Strong ability to create and maintain technical architecture and interface documentation Work Authorization & Clearance Requirements U.S.
Citizenship is required Must be eligible to obtain or currently hold a Public Trust clearance Work Location / Travel Remote position Must be comfortable with quarterly travel to Gaithersburg, MD Candidates located in the Greater DC / Maryland / Virginia area are highly preferred #J-18808-Ljbffr