Home

Urgently Hiring ::Technical Specialist::OH:: at Columbus, Ohio, USA
Email: [email protected]
From:

Priyabrata Pradhan,

Vyze Inc

[email protected]

Reply to:   [email protected]

Job Description -

Title: Technical Specialist

Location: Columbus, Ohio (Hybrid)              

Visa: USC, GC, GC-EAD, H4-EAD  

Duration: 6+ Months

MOI: Video

Note: This resource would be required to be on-site Tuesdays and Thursdays and remote remaining days. Candidates must be EST Zone.

About The Job:

The Technical Specialist will be responsible for Enterprise Data Warehouse (EDW) design, development, implementation, migration, maintenance and operation activities. Works closely with Data Governance and Analytics team. The candidate will closely with Data Governance and Analytics team. 

Will be one of the key technical resource for data warehouse projects for various Enterprise Data Warehouse projects and building critical Data Marts, data ingestion to Big Data platform for data analytics and exchange with State and Medicaid partners. Participate in Team activities, Design discussions, Stand up meetings and planning Review with team.

Perform data analysis, data profiling, data quality and data ingestion in various layers using big data/Hadoop/Hive/Impala queries, Spark programs and UNIX shell scripts.

Follow the organization coding standard document, Create mappings, sessions and workflows as per the mapping specification document.

Perform Gap and impact analysis of ETL and IOP jobs for the new requirement and enhancements.

Create jobs in Hadoop using SQOOP, PYSPARK and Stream Sets to meet the business user needs.

Create mock-up data, perform Unit testing and capture the result sets against the jobs developed in lower environment.

Updating the production support Run book, Control M schedule document as per the production release.

Create and update design documents, provide detail description about workflows after every production release.

Continuously monitor the production data loads, fix the issues, update the tracker document with the issues, and identify the performance issues.

Performance tuning long running ETL/ELT jobs by creating partitions, enabling full load and other standard approaches.

Perform Quality assurance check, Reconciliation post data loads and communicate to vendor for receiving fixed data.

Participate in ETL/ELT code review and design re-usable frameworks.

Create Remedy/Service Now tickets to fix production issues, create Support Requests to deploy Database, Hadoop, Hive, Impala, UNIX, ETL/ELT and SAS code to UAT environment.

Create Remedy/Service Now tickets and/or incidents to trigger Control M jobs for FTP and ETL/ELT jobs on ADHOC, daily, weekly, monthly and quarterly basis as needed.

Model and create STAGE / ODS / Data warehouse Hive and Impala tables as and when needed.

Create Change requests, work plan, Test results, BCAB checklist documents for the code deployment to production environment and perform the code validation post deployment.

Work with Hadoop Admin, ETL and SAS admin teams for code deployments and health checks.

Create re-usable UNIX shell scripts for file archival, file validations and Hadoop workflow looping.

Create re-usable framework for Audit Balance Control to capture Reconciliation, mapping parameters and variables, serves as single point of reference for workflows.

Create Spark programs to ingest historical and incremental data.

Create SQOOP scripts to ingest historical data from EDW Module Vendors databases to Hadoop IOP, created HIVE tables and Impala views creation scripts for Dimension tables.

Participate in meetings to continuously upgrade the Functional and technical expertise.

Required Skills:                                                                       

8 Years of experience in Big Data, Hadoop on Data Warehousing or Data Integration projects.

Analysis, Design, development, support and Enhancements of ETL/ELT in data warehouse environment with Cloudera Big data Technologies.

8 Years of experience in Development.

8 Years of experience in Writing Hadoop/Hive/Impala scripts for gathering stats on table post data loads.

Writing complex SQL queries and performed tuning based on the Hadoop/Hive/Impala explain plan results.

Experience building data sets and familiarity with PHI and PII data.

Basic knowledge of UNIX/LINUX shell scripting.

Thanks & Regards

Priyabrata pradhan

Email: [email protected]

Keywords: green card
[email protected]
View all
Thu Dec 14 06:32:00 UTC 2023

To remove this job post send "job_kill 937078" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 9

Location: Columbus, Ohio