Home

Data Engineer Pipeline/ETL (AWS) || 9+ years at Remote, Remote, USA
Email: [email protected]
Position: Data Engineer Pipeline/ETL (AWS)

Duration: 6+ Months (with high possibility of an extension)

Location: Remote
Job Summary:

Looking for a Data Pipeline Engineer to join our growing Data Engineering team and participate in design and build of data ingestion and transformation pipelines based on the specific needs driven by Product Owners and Analytics consumers. The candidate should possess strong knowledge, interest in data processing, and have a background in data engineering. Candidate will also have to work directly with senior data engineers, solution architects, DevOps engineers, product owners and data consumers to deliver data products in a collaborative and agile environment. They will also have to continuously integrate and push code into our cloud production environments.

Job Description:

As a key contributor to the data engineering team, the candidate is expected to:

Build and deploy modular data pipeline components such as
 Apache Airflow DAGs, AWS Glue jobs, AWS Glue crawlers through a CI/CD process.

Translate Business or Functional Requirements to actionable technical build specifications.

Collaborate with other technology teams to extract, transform, and load data from a wide variety of data sources.

Work closely with product teams to deliver data products in a collaborative and agile environment.

Perform data analysis and onboarding activities as new data sources are added to the platform.

Proficient in data modeling techniques and concepts to support data consumers in designing the most efficient method of storage and retrieval of data.

Evaluate innovative technologies and tools while establishing standard design patterns and best practices for the team.

Qualifications:

Required:

Experience in AWS Data processing, Analytics, and storage Services such as
 Simple Storage Service (s3), Glue, Athena and Lake Formation

Experience in extracting and delivering data from various databases such as
 MongoDB, DynamoDB, SnowFlake, Redshift, Postgres, RDS

Coding experience with
 Python, SQL, yaml, spark programming (pyspark)

Hands on experience with
 Apache Airflow
 as a pipeline orchestration tool

Experience in AWS Serverless services such as
 Fargate, SNS, SQS, Lambda

Experience in Containerized Workloads and using cloud services such as
 AWS ECS, ECR and Fargate to scale and organize these workloads.

Experience in data modeling and working with analytics teams to design efficient data structures.

Applied knowledge of working in agile, scrum, or DevOps environments and teams

Applied knowledge of modern software delivery methods like
 TDD, BDD, CI/CD

Applied knowledge of Infrastructure as Code (IAC)

Experience with development lifecycle (development, testing, documentation, and versioning)

Preferred:

AWS Certified Developer Associate

AWS Certified Big Data Specialty

Gitlab CI/CD

--

Keywords: continuous integration continuous deployment sthree information technology
[email protected]
View all
Wed Oct 18 22:04:00 UTC 2023

To remove this job post send "job_kill 765999" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 0

Location: ,