Home

bigdata engineers: onsite Pheonix: DOE at Remote, Remote, USA
Email: [email protected]
From:

Jay,

Brillius

[email protected]

Reply to:   [email protected]

Bigdata engineer- Phoenix

Key Responsibilities:

Responsible for designing system solutions, developing custom applications, and modifying existing applications to meet distinct and changing business requirements. Handle coding, debugging, and documentation, as well working closely with SRE team. Provide post implementation and ongoing production support

Develop and design software applications, translating user needs into system architecture. Assess and validate application performance and integration of component systems and provide process flow diagrams. Test the engineering resilience of software and automation tools.

You will be challenged with identifying innovative ideas and proof of concept to deliver against the existing and future needs of our customers. Software Engineers who join our Loyalty Technology team will be assigned to one of several exciting teams that are developing a new, nimble, and modern loyalty platform which will support the key element of connecting with our customers where they are and how they choose to interact with American Express.

Be part of an enthusiastic, high performing technology team developing solutions to drive engagement and loyalty within our existing cardmember base and attract new customers to the Amex brand.

The position will also play a critical role partnering with other development teams, testing and quality, and production support, to meet implementation dates and allow smooth transition throughout the development life-cycle.

The successful candidate will be focused on building and executing against a strategy and roadmap focused on moving from monolithic, tightly coupled, batch-based legacy platforms to a loosely coupled, event-driven, microservices-based architecture to meet our long-term business goals.

Must Have Qualifications:

Bachelors degree in Engineering or Computer Science or equivalent OR Masters in Computer Applications or equivalent.

5+ years of software development experience and leading teams of engineers and scrum teams

3+ years of hands-on experience of working with Map-Reduce, Hive, Spark (core, SQL and PySpark)

Hands-on experience on writing and understanding complex

SQL(Hive/PySpark-dataframes), optimizing joins while processing huge amount of data

Experience in UNIX shell scripting

Additional Good to have requirements:

Solid Datawarehousing concepts

Knowledge of Financial reporting ecosystem will be a plus

Experience with Data Visualization tools like Tableau, Sisense, Looker

Expert on Distributed ecosystem

Hands-on experience with programming using Python/Scala

Expert on Hadoop and Spark Architecture and its working principle

Ability to design and develop optimized Data pipelines for batch and real time data processing

Should have experience in analysis, design, development, testing, and implementation of system applications

Demonstrated ability to develop and document technical and functional specifications and analyze software and system processing flows

Aptitude for learning and applying programming concepts.

Ability to effectively communicate with internal and external business partners. Preferred Qualifications:

Knowledge of cloud platforms like GCP/AWS, building Microservices and scalable solutions, will be preferred

2+ years of experience in designing and building solutions using Kafka streams or queues

Experience with GitHub and leveraging CI/CD pipelines

Experience with NoSQL i.e., HBase, Couchbase, MongoDB

Bigdata engineers for finance

Minimum Qualifications: Bachelors degree in engineering or computer science or equivalent OR masters in computer applications or equivalent. 10+ years of software development experience and leading teams of engineers and scrum teams 5+ years of hands-on experience of working with Map-Reduce, Hive, Spark (core, SQL and PySpark) Solid Data warehousing concepts.

Knowledge of Financial reporting ecosystem will be a plus .5+ years of experience within Data Engineering/ Data Warehousing using Big Data technologies will be a addon Expert on Distributed ecosystem Hands-on experience with programming using Core Java or Python/Scala . Expert on Hadoop and Spark Architecture and its working principle Hands-on experience on writing and understanding complex SQL(Hive/Py Spark-dataframes), optimizing joins while processing huge amount of data .Experience in UNIX shell scripting .Ability to design and develop optimized Data pipelines for batch and real time data processing Should have experience in analysis, design, development, testing, and implementation of system applications Demonstrated ability to develop and document technical and functional specifications and analyze software and system processing flows.

Preferred Qualifications: Knowledge of cloud platforms like

GCP/AWS, building Microservices and scalable solutions, will be an advantage 1 + years of experience in designing and building solutions using Kafka streams or queues Experience with GitHub/Bitbucket and leveraging CI/CD pipelines .Experience with NoSQL i.e., HBase, Couchbase, MongoDB is good to have Excellent technical and analytical aptitude Good communication skills .Excellent Project management skills. Results driven.

Keywords: continuous integration continuous deployment
[email protected]
View all
Thu Oct 19 00:42:00 UTC 2023

To remove this job post send "job_kill 767365" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 0

Location: ,