Data Engineer with expert at building Data Pipelines in AWS and validating Data Quality Exp ( 9+ Exp in need ) at Remote, Remote, USA |
Email: [email protected] |
From: Steve, 3mkllc [email protected] Reply to: [email protected] Hello, Greetings for the day!!! Please review the below role and advise the best time to connect with you. If you are interested, You can call me at +1 801-701-7848or reach me on linkedin.com/in/saumen-kumar-porel-97a0696b and send resumes to [email protected] Hiring : Data Engineer with expert at building Data Pipelines in AWS and validating Data Quality Exp ( 9+ Exp in need ) Loc: Malvern, PA (Remote Now) Start Date: Immediate Contract: Long Term Feedback from the manager- The candidate should be an expert at building data pipelines in AWS and validating data quality. The following core technologies are most relevant: AWS (including S3, Athena, Glue, CloudFormation, IAM, and SageMaker as must haves) Python + Data Analytics Libraries (NumPy, Pandas, Matplotlib) Git SQL Familiarity with software engineering principles like Object-Oriented Design, Functional Programming, APIs, Libraries, etc. Familiarity with Spark ***Exclusive backfill We are looking for someone who has a data engineering skillset who will be part of an AI/ML (artificial intelligence / machine learning) team. Below are the core responsibilities and qualifications: Core Responsibilities: Writes ETL (Extract / Transform / Load) processes, designs database systems, and develops tools for real-time and offline analytic processing. Troubleshoots software and processes for data consistency and integrity. Integrates data from a variety of sources for business partners to generate insight and make decisions. Translates business specifications into design specifications and code. Responsible for writing programs, ad hoc queries, and reports. Ensures that all code is well structured, includes sufficient documentation, and is easy to maintain and reuse. Partners with internal clients to gain a basic understanding of business functions and informational needs. Gains working knowledge in tools, technologies, and applications/databases in specific business areas and company-wide systems. Participates in all phases of solution development. Explains technical considerations at related meetings. Tests code thoroughly for accuracy of intended purpose. Reviews end product with the internal stakeholders to ensure adequate understanding. Provides data analysis guidance as required. Provides tool and data support to business users and fellow team members. Tests and implements new software releases through regression testing. Identifies issues and engages with vendors to resolve and elevate software into production. Qualifications: Experience and expertise in engineering data ingestion and transformation products using modern cloud-based services and platforms such as Glue ETL, Glue Catalog, Glue Crawlers, S3, Scala/Python, Spark/Pyspark, SQL, Attunity and other adjacent tools used for data exploration such as AWS Athena and Presto. Experience and expertise in Python and SQL. Minimum of five years data analytics, programming, database administration, or data management experience. Undergraduate degree or equivalent combination of training or experience. Additional/Preferred Qualifications: Strong domain knowledge in cybersecurity I know that it is difficult to find an experienced data engineer with domain knowledge in cybersecurity, but thought it couldnt hurt to mention the preference in case there is someone available. |
[email protected] View all |
Sat Oct 22 18:59:00 UTC 2022 |