Home

AWS Data Engineer - Cincinnati, OH - REMOTE at Cincinnati, Ohio, USA
Email: [email protected]
From:

Ashwini Lokhande,

ChabezTech LLC

[email protected]

Reply to:   [email protected]

Hello,

Please review the JD below and let me know your interest.

Role - AWS Data Engineer

Location: Cincinnati, OH - REMOTE

Experience : 12+ years

Data Engineer with Data Stage and  Snowflake MUST

JOB DESCRIPTION:

Key Responsibilities:

Purpose of a Stream: Understand the concept of data streams and design real-time data processing

pipelines to enable near real-time analytics and decision-making.

Language Choice: Utilize Python as the primary programming language for data engineering tasks,

leveraging its rich ecosystem of libraries and frameworks for data manipulation, transformation, and

integration.

Loading Data into Snowflake Tables: Implement data loading strategies and techniques that ensure data

integrity and eliminate duplicates when loading data into Snowflake tables.

Airflow-Snowflake Integration: Configure Airflow to seamlessly invoke and interact with Snowflake,

ensuring smooth data pipeline execution and monitoring.

Big Data and Hadoop: Demonstrate experience and knowledge in working with Big Data technologies

and Hadoop ecosystem, including data ingestion, processing, and storage.

Big Data File Processing: Experience in handling and processing large-scale Big Data files, such as

Parquet, Avro, or ORC, optimizing performance and data retrieval.

Excel File in Snowflake: Proficiently load and query Excel files in Snowflake, leveraging appropriate data

connectors or integration tools.

SQS and SNS Experience: Utilize Amazon Simple Queue Service (SQS) and Simple Notification Service

(SNS) to build scalable and reliable event-driven data processing architectures.

Event Tables: Implement event-based data architectures using event tables, capturing and processing

real-time events for timely data updates and analytics.

Dataframes: Work with various data manipulation frameworks and libraries such as Pandas, PySpark, or

Dask, leveraging their capabilities to process and transform data efficiently.

Qualifications:
Bachelor's degree in Computer Science, Data Engineering, or a related field.
Proven experience as a Data Engineer or similar role, with a strong focus on data pipeline design,
implementation, and optimization.
Proficient in Python and its data manipulation libraries (e.g., Pandas, NumPy).
Familiarity with Apache Airflow or similar workflow orchestration tools.
Experience with Snowflake data warehouse and associated functionalities.
Strong SQL skills and knowledge of database concepts.
Familiarity with Big Data technologies and Hadoop ecosystem.
Excellent problem-solving and analytical skills.
Strong attention to detail and ability to deliver high-quality work.
Strong communication and collaboration skills to work effectively with cross-functional teams.
Join our team and contribute to building a scalable and efficient data infrastructure that empowers

data-driven decision-making across our organization.

Keywords: Ohio
[email protected]
View all
Wed Jan 31 23:58:00 UTC 2024

To remove this job post send "job_kill 1071867" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 7

Location: Cincinnati, Ohio