Home

Senior Data Engineer with Databricks, Python, Pyspark experience at Remote, Remote, USA
Email: [email protected]
From:

Sanjeev Kumar Singh,

Tek Inspirations LLC

[email protected]

Reply to:   [email protected]

Job Description -

Senior Data Engineer with Databricks, Python, Pyspark experience

Location: Remote

Duration: 6+ months

GC/USC only

Need Senior Resource 

Project Description

As a Senior Engineer in the Data Engineering & Analytics team, you will develop data & analytics solutions that sit atop vast datasets gathered by retail stores, restaurants, banks, and other consumer-focused companies. The challenge will be to create high-performance algorithms, cutting-edge analytical techniques including machine learning and artificial intelligence, and intuitive workflows that allow our users to derive insights from big data that in turn drive their businesses. You will have the opportunity to create high-performance analytic solutions based on data sets measured in the billions of transactions and front-end visualizations to unleash the value of big data. You will have the opportunity to develop data-driven innovative analytical solutions and identify opportunities to support business and client needs in a quantitative manner and facilitate informed recommendations/decisions through activities like building ML models, automated data pipelines, designing data architecture/schema, performing jobs in big data cluster by using different execution engines and program languages such as Hive/Impala, Python, Java, Kafka, Spark, R, etc.

Must have

Working proficiency in using Python/Scala, Spark (tuning jobs), SQL, Hadoop platforms to build Big Data products & platforms.

Good understanding of Databricks security, clusters, user management, deployment and performance tuning.

Strong PySpark skills

Good programming skills in Java and spring boot and Junit.

Knowledge in software development test approaches & frameworks

Familiarity with RESTful APIs and micro-services architectures

Experience in working with CI/CD

Experience in working with SQL database like Postgres, Oracle

Preferably with hands-on experience with Hadoop big data tools (Hive, Impala, Spark)

Experience with data pipeline and workflow management tools: NIFI, Airflow.

Comfortable in developing shell scripts for automation.

Good troubleshooting and debugging skills.

Proficient in standard software development, such as version control, testing, and deployment

Demonstrated basic knowledge of statistical analytical techniques, coding, and data engineering

Ability to quickly learn and implement new technologies

Ability to Solve complex problems with multi-layered data sets

Ability to innovate and determine new approaches & technologies to solve business problems and generate business insights & recommendations.

Ability to multi-task and strong attention to detail

Flexibility to work as a member of a matrix based diverse and geographically distributed project teams

Good communication skills - both verbal and written - and strong relationship, collaboration skills, and organizational skills

Nice to have

Experience with performance Tuning of Database Schemas, Databases, SQL, ETL Jobs, and related scripts

Experience in working with Cloud APIs (e.g., Azure, AWS)

Experience participating in complex engineering projects in an Agile setting e.g. Scrum

Keywords: continuous integration continuous deployment machine learning rlang green card
[email protected]
View all
Mon Nov 20 22:58:00 UTC 2023

To remove this job post send "job_kill 872244" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 0

Location: ,