AWS Data Architect Senior Developer
SCOPE: As an AWS Data Architect Senior Developer, you will be responsible for designing, implementing, and maintaining AWS data pipelines, databases, and data warehousing solutions. You will collaborate closely with our cross-functional teams to ensure efficient data extraction, transformation, loading (ETL), and reporting processes. Your expertise in AWS Analytics will be crucial in optimizing our data ecosystem and enabling data-driven decision-making.
DUTIES: As part of the Data Analytics team on Experience Transformation (XT) Program, the individual is responsible for delivering and maintaining a solution that best suits business requirements. XT Program is focused on transforming and upgrading our SAP footprint to support Finance, HR, Procurement, Order to Cash and CRM capabilities. The successful candidate will develop application code in accordance with application standards and best practices working on projects, change requests and minor enhancements. This individual is responsible for analyzing, developing and delivering the technical solution that satisfies the functional design documents and other useful requirements.
Key Responsibilities:
- Design, develop, and maintain robust and scalable data pipelines using Python Scripts.
- Implement ETL processes to extract, transform, and load data from various sources into our data warehouse (Amazon RedShift).
- Collaborate with Subject Matter Experts, data analysts, and business stakeholders to understand data requirements and ensure data accuracy and integrity.
- Build and maintain data models and schemas for efficient data storage and retrieval.
- Monitor and troubleshoot data pipelines and ETL processes to ensure smooth data flow and minimal downtime.
- Utilize Python programming for scripting, automation, and data manipulation tasks.
- Stay updated with the latest AWS services, data engineering best practices, and emerging technologies.
- Meet with Data Architects and Project Leads to understand business, financial, and operations requirements and systems goals, and identify and resolve systems issues.
- Load target data bases using complex transformations that come from many systems.
- Monitor loads and troubleshoot any issues that arise.
SKILLS AND EXPERIENCES
Must Have:
- 4+ years of relevant experience in a similar role.
- 4+ years of proven experience as a Data Engineer, with a strong focus on AWS technologies (S3, AWS Glue, Lambda, Athena) and developing ETL/ELT pipelines using these AWS services.
- 4+ years of hands On with AWS managed Apache Airflow / Apache Airflow.
- 4+ year of proficiency in establishing cost-effective data centered solutions in AWS
- Proficiency in SQL for data querying, manipulation, and optimization
- Familiarity with consuming data from SAP S/4 entities, models, and schemas.
- Understanding of data integration, ETL processes, and data quality management.
- Excellent problem-solving skills and ability to work collaboratively in a team environment.
Nice to have:
- Development experience using data programming languages such as Python, Pyspark.
- Development with database technologies and experience with data warehousing methodologies and technologies.
- Experience with Data Lakehouse and Delta Lakehouse.
- Experience with real-time data processing and stream processing frameworks.
- Understanding of BI tools such as Power BI and QuickSight.
- Hands-on experience with Amazon RedShift or similar data warehousing solutions.
- Strong programming skills in Python for scripting and automation.
- Understanding of data integration, ETL processes, and data quality management.
- Good to have AWS certifications (e.g., AWS Certified Data Analytics – Specialty).
Competencies
Be Courageous
Innovate
Be Inclusive
Inspire
Be Curious
Technical Knowledge