AWS Data Engineer
Job Description
The AWS Data Engineer will play a key role in developing and scaling a modern cloud-based data platform. This individual will architect, implement, and support high-performing data pipelines and lakehouse environments within AWS, ensuring reliable, efficient, and secure data delivery for analytics and reporting initiatives.
This role centers on designing scalable data structures, enabling seamless data integration across varied sources, and leveraging AWS-native technologies to build a resilient and future-ready data ecosystem. The position reports to the Engineering Manager.
Core Responsibilities
-
Data Architecture & Modeling
Develop and maintain dimensional and relational data models tailored for analytical performance within a lakehouse framework.
-
Cloud Lakehouse Development
Design, implement, and administer AWS-based lakehouse solutions utilizing services such as Amazon S3, AWS Glue, Amazon Redshift Spectrum, and AWS Lake Formation to support governance, cataloging, and secure data access.
-
Pipeline Engineering
Build and enhance scalable batch and near-real-time data workflows using AWS Glue, Lambda, Kinesis, and Step Functions. Ensure reliability through monitoring and performance optimization.
-
Data Integration
Consolidate structured and unstructured data from APIs, databases, streaming platforms, and third-party systems into centralized storage layers.
-
Automation & Infrastructure Management
Implement Infrastructure-as-Code practices using CloudFormation, CDK, or similar tools, supported by CI/CD processes for deployment and lifecycle management.
-
Analytics Enablement
Prepare curated datasets for analytics and machine learning initiatives, conduct detailed data analysis, and support visualization efforts through tools such as AWS QuickSight.
-
Perform additional related duties as needed.
Performance Expectations
Success in this role will be demonstrated by:
-
Delivering adaptable and well-documented data models that evolve with business requirements.
-
Establishing secure, governed lakehouse environments with strong metadata management and access controls.
-
Creating resilient, high-volume pipelines equipped with monitoring, alerting, and automated failover mechanisms.
-
Improving cost efficiency and query performance through optimization strategies across Redshift, Athena, and S3 storage design.
-
Maintaining high standards of data integrity, lineage tracking, and documentation.
-
Partnering effectively with data scientists, analytics engineers, and business stakeholders to align technical solutions with organizational goals.
Qualifications
To thrive in this role, candidates should bring:
-
3+ years of hands-on experience with AWS data services, including S3, Glue, Lake Formation, Redshift, Athena, and IAM.
-
4+ years of experience designing data models for warehouse and lakehouse environments.
-
4+ years building and orchestrating ETL/ELT pipelines using AWS-native tools such as Glue, Lambda, and Step Functions.
-
Strong proficiency in SQL and Python.
-
Experience implementing Infrastructure-as-Code (CloudFormation, CDK, Terraform) along with CI/CD practices.
-
Solid understanding of cloud data governance, security frameworks, and compliance standards.
-
Strong communication skills with the ability to collaborate across technical and business teams.
-
Experience working effectively in remote or distributed environments.
-
Familiarity with BI platforms (e.g., Tableau or Power BI) and project collaboration tools such as Jira and Confluence.
The Phoenix Group Advisors is an equal opportunity employer. We are committed to creating a diverse and inclusive workplace and prohibit discrimination and harassment of any kind based on race, color, religion, gender, sexual orientation, gender identity or expression, national origin, age, genetic information, disability, or veteran status. We strive to attract talented individuals from all backgrounds and provide equal employment opportunities to all employees and applicants for employment.