DATA ENGINEER - Remote - Immediate requirement at Remote, Remote, USA |
Email: harry@dvgts.com |
https://jobs.nvoids.com/job_details.jsp?id=271166&uid= From: Harry Mino, DVG TECH SOLUTIONS harry@dvgts.com Reply to: harry@dvgts.com Position: Data Engineer (AWS, Python, PySpark) Location: Candidate can be anywhere in USA/Canada Terms: W2/C2C/1099 Job Description: Primary: Databricks/Lakes+SQL+Data Science MLops (Python or Pyspark related) Develop, implement, support, and operationalize AWS data lake infrastructure & services Create and maintain optimal data pipeline architecture, Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL, ETL (e.g., Informatica Cloud) and AWS Data Lake technologies. Build analytics tools that utilize the data pipeline to provide actionable insights into patient care, operational efficiency and other key business performance metrics. Develop a deep understanding of AWSs vast data sources and know exactly how, when, and which data to use to solve business problems. Monitor and maintain data lake security and data lake services. Manage numerous requests concurrently and strategically, prioritizing when necessary Troubleshoot technical issues and provide solutions and fixes using various tools and information such as server logs and report debug logs. General and administrative tasks Job Requirement Bachelors degree in Computer Science, Information Systems, Mathematics, or a related discipline. 10+ Years experience in Information Technology within a complex, matrixed, and global business environment. 5+ Years Experienced as a data engineering/Science with AWS Data Lake Technologies and Services Building and optimizing AWS Data Lake data pipelines, architectures, and data sets. Strong analytic skills related to working with unstructured datasets. Build processes supporting data transformation, data structures, metadata, dependency, and workload management. A successful history of manipulating, processing, and extracting value from large, disconnected datasets. Understanding of message queuing, stream processing, and highly scalable AWS Data Lake data stores. Understanding of database and analytical technologies in the industry including MPP and NoSQL databases (e.g., Snowflake), Data Warehouse design, ETL, BI reporting and Dashboard development. Experience with Agile framework and DevOps. 3+yrs of experience in building ETL data pipelines using AWS Glue and Pyspark. Efficient in developing Spark scripts for data ingestion, aggregation and transformation Exception Handling and performance optimization techniques on python/pyspark scripts Preferred Qualifications: Professional certifications e.g., AWS Certified solution architect, etc. AWS data lake administrations Excellent documentation and interpersonal relationship skills with ability to drive achievement of objectives. Strong interpersonal and leadership skills. Strong written and verbal communication skills including the ability to communicate at various levels within an organization and to explain complex or technical matters in a manner suitable for a non-technical audience. Knowledge of best practices related to Data Lake, Data lake governance, Data Security (e.g., HITRUST), and Data Integration & Interoperability. Email is the best way to reach me.! Thanks & Regards, Harry - IT RECRUITER DVG Tech Solutions LLC 666 Plainsboro Rd, Suite 1010,Plainsboro, NJ 08536 Desk: 732-807-6258 Email : harry@dvgts.com|| http://www.dvgts.com LinkedIn: https://www.linkedin.com/in/harinadha-munelli-33295a146/ Keywords: business intelligence information technology wtwo New Jersey https://jobs.nvoids.com/job_details.jsp?id=271166&uid= |
harry@dvgts.com View All |
09:33 PM 12-Jan-23 |