Home

Hiring!!! - Data engineer - eCommerce domain preferred - Need Locals at Remote, Remote, USA
Email: [email protected]
Position : Data
engineer - eCommerce domain preferred

Location :
Sunnyvale, CA Need Locals

Duration : 12
Months

Must
to have Experience :

Experience
in eCommerce domain preferred

Experience
with in big data technologies like Hadoop, Apache Spark (Scala preferred),
Apache Hive,

Experience
in building idempotent workflows using orchestrators like Automic, Airflow,
Luigi etc.

Experience
in writing SQL to analyze, optimize, profile data preferably in BigQuery or
SPARK SQL

Experience
building complex near real time (NRT) streaming data pipelines using Apache
Kafka, Spark streaming, Kafka Connect with a strong focus on stability,
scalability and SLA adherence

What
you'll do:

You
will use cutting edge data engineering techniques to create critical datasets
and dig into our mammoth scale of data to help unleash the power of data
science by imagining, developing, and maintaining data pipelines that our Data
Scientists and Analysts can rely on.

You
will be responsible for contributing to an orchestration layer of complex data
transformations, refining raw data from source into targeted, valuable data
assets for consumption in a governed way.

You
will partner with Data Scientists, Analysts, other engineers, and business
stakeholders to solve complex and exciting challenges so that we can build out
capabilities that evolve the marketplace business model while making a positive
impact on our customers' and sellers lives.

You
will participate with limited help in small to large sized projects by
reviewing project requirements; gather requested information; write and develop
code; conduct unit testing; communicate status and issues to team members and
stakeholders; collaborate with project team and cross functional teams;
troubleshoot open issues and bug-fixes; and ensure on-time delivery and
hand-offs.

You
will design, develop and maintain highly scalable and fault-tolerant real time,
near real time and batch data systems/pipelines that process, store, and serve
large volumes of data with optimal performance.

You
will ensure data ingested and processed is accurate and of high quality by
implementing data quality checks, data validation, and data cleaning processes.

You
will identify possible options to address business problems within one's
discipline through analytics, big data analytics, and automation.

You
will build business domain knowledge to support the data need for product
teams, analytics, data scientists and other data consumers.

What
you'll bring:

At
least 8+ years of experience development of big data technologies/data
pipelines

Experience
in managing and manipulating huge datasets in the order of terabytes (TB) is
essential.

Experience
with in big data technologies like Hadoop, Apache Spark (Scala preferred),
Apache Hive, or similar frameworks on the cloud (GCP preferred, AWS, Azure
etc.) to build batch data pipelines with strong focus on optimization, SLA
adherence and fault tolerance.

Experience
in building idempotent workflows using orchestrators like Automic, Airflow,
Luigi etc.

Experience
in writing SQL to analyze, optimize, profile data preferably in BigQuery or
SPARK SQL

Strong
data modeling skills are necessary for designing a schema that can accommodate
the evolution of data sources and facilitate seamless data joins across various
datasets.

Ability
to work directly with stakeholders to understand data requirements and translate
that to pipeline development / data solution work.

Strong
analytical and problem-solving skills are crucial for identifying and resolving
issues that may arise during the data integration and schema evolution
process.

Ability
to move at a rapid pace with quality and start delivering with minimal ramp up
time will be crucial to succeed in this initiative.

Effective
communication and collaboration skills are necessary for working in a team
environment and coordinating efforts between different stakeholders involved in
the project.

Nice
to have from you:

Experience
building complex near real time (NRT) streaming data pipelines using Apache
Kafka, Spark streaming, Kafka Connect with a strong focus on stability,
scalability and SLA adherence.

Good
understanding of REST APIs working knowledge on Apache Druid, Redis, Elastic
search, GraphQL or similar technologies. Understanding of API contracts,
building telemetry, stress testing etc.

Exposure
in developing reports/dashboards using Looker/Tableau

Experience
in eCommerce domain preferred.

--

Keywords: information technology California
Hiring!!! - Data engineer - eCommerce domain preferred - Need Locals
[email protected]
[email protected]
View all
Thu Sep 19 21:14:00 UTC 2024

To remove this job post send "job_kill 1766050" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 11

Location: Sunnyvale, California