Need 11+ years :: Hadoop withETL Informatica Pyspark Developer, Columbus OH at Columbus, Ohio, USA |
Email: [email protected] |
From: Anil, Maxpro Inc,. [email protected] Reply to: [email protected] Hi, Hope you are doing great!!! Please revert me resume and details ASAP. Title: Hadoop ETL Informatica Developer with Pyspark Location: Columbus OH Duration: Longterm Need 11+ years of experience candidates. Need than 3 years of experience in the PySpark ETL Below are the JD and key requirements of tolls and technologies for 2 positions - 5-6 years of development experience with Oracle and Big Data Hadoop platform on Data Warehousing and/or Data Integration projects in an agile environment. Understanding of business requirements as well as technical aspects. Good knowledge of Big Data, Hadoop, Hive, Impala database, data security, and dimensional model design. 6-8 Strong Exprriend in the - Sqoop, PySpark, Spark, HDFS, Hive, Impala, StreamSets, and Kudu technologies. Strong knowledge in analyzing data in data warehouse environment with Cloudera Bigdata Technologies (Hadoop, MapReduce, Sqoop, PySpark, Spark, HDFS, Hive, Impala, StreamSets, Kudu, Oozie, Hue, Kafka, Yarn, Python, Flume, Zookeeper, Sentry, Cloudera Navigator) and Oracle SQL/PL-SQL. Strong knowledge in writing complex SQL queries (Oracle and Hadoop (Hive/Impala etc.)). Knowledge in analyzing the log files, error files for any data ingestion failures. Experience in writing Python/Impala scripts Tokenization or Data masking knowledge Experience working in Medicaid and healthcare domain is preferred. Participate in Team activities, Design discussions, stand ups, sprint planning and execution meetings with team. Perform data analysis, data profiling and data quality assessment in various layers using big data/Hadoop/Hive/Impala and Oracle SQL queries. Thanks & Regards, Anil Maxpro Inc,. E-mail: [email protected] Web: www.maxprous.com Keywords: procedural language Ohio |
[email protected] View all |
Fri Nov 17 22:12:00 UTC 2023 |