Home

Veeranjaneyulu Tokala - ETL Talend Developer
[email protected]
Location: Columbus, Ohio, USA
Relocation: Remote
Visa: H1B
Veeranjaneyulu Tokala
ETL Talend Developer
+1630-580-6406.
[email protected]
Columbus, OH
Remote
H1B


LinkedIn: linkedin.com/in/veeranjaneyulu-tokala-685673b7


Synopsis
A competent professional with more than 15+ years of professional experience
ETL, Big Data, DW Development, EDA, Data cleaning, Data Wrangling.
Involved in handling Talend and Informatica software and in-depth knowledge in managing all the projects under DWH and MDM projects.
Develop Low level and High level technical design documents for solution implementation.
Handle client communication regarding requirements, design, clarifications, etc.
Experience on Data Cleansing and Standardizations of Data.
Developing Talend Jobs based on the Mapping Specification Provided
Hands on Experience to use Talend Administration Center (TAC) and Code migration.
Breaking down the task and Putting down the efforts against it to analyze the business value that you provide to the client.
Owning the Ownership of the Jobs Developed and resolving the defects on the fly if any, once when moved to different environments.
Performed Unit Testing based on multiple test cases (Positive and Negative Scenarios) and prepared release notes to the framework Joblets.
Having a strong knowledge on generating Profiling Reports for creating of Business Rules.
Involved in Data Extraction, Transformation and Loading using ETL tool.
Extracted data from various cloud data sources like AWS-S3, Cassandra, Snowflake, Datalake including Parquet, Avro, CSV and Flat files.
Have designed metadata frame work job and migrated the individual source jobs to Single job and migrated the data into AWS and processed files using Glue into redshift tables.
Expertise in the domain of Insurance and Expert in designing ETL /DQ/MDM solutions and Data Migration Project Platforms.
Experience on AWS cloud services (EC2, S3, RDS, Redshift, IAM)
Good working Knowledge on Data warehousing concepts like Schemas, Dimensions and Fact tables.
Implemented end to end solution for Talend DI/ MDM Framework for generating of Golden Customer record across organizational level.
Proficient in arranging business requirements (requirement analysis), preparing the detailed process flow for functional requirements
An effective communicator with good problem-solving abilities, multitasking ability, willingness to take up ownership, facilitate teamwork and lead activities
Worked on various domains like BFSI, Health-care, Education and Logistics.
Achievements
Provided solution to one of the leading Bank to provide financial health check to improve processing time 5 Hrs to 2 mins.
Converted Legacy process interest calculation reports processing time 10 hr to 3 mins by leveraging Hadoop and spark frameworks.
Architected and implemented open source solution for an existing product which ended up in saving huge dollars and significant performance numbers.
IT Skills
Well versed with:
o Big Data Technologies: MapReduce, Hadoop, Spark, Hive, Pig, Zookeeper, Sqoop.
o Hadoop Distribution: CDH (5.7),HDP
o Programming: Scala, Python, Core java, SQL, PL/SQL,R
o Web Scrapping
o Data Stream: Kafka
o MDM Tool: Semarchy(Intelligent Data Hub)
o ETL Tools: Talend DI & Bigdata Studio, Talend cloud,Informatica.Abinitio
o DB: Microsoft SQL Server, Oracle, MySQL, PostgreSQL, RedShift
o OS: Windows, Unix/Linux
o IDE: SQL developer, Toad, Jupiter, Eclipse
o Dimensional Modelling(Star/Snowflake/Hybrid)
o Cloud: Services: AWS, Azure
o JIRA, MS-Office, MS-Project




Work Experience
MAK technologies(Texas,USA)
Client:VIRTUSA
Jul2022-tilldate
Project: AWS data lake
Responsibilities:
Creating ETL pipeline to load the data from XML files to redshift tables.
Extracting the data from Live person conversation messages and process them to load into ods layer.
Developing ETL jobs to load from ods tables to dimension and fact tables.
Developed framework jobs for generating parameter files and batch start and batch end mechanisms.
Consumed Bigdata feeds which will be generated out of logs from server to capture server metrics.
Efficiently designed error handling mechanisms and parameterized all the source and target information at configuration table.
Troubleshooting the talend jobs to identify the root cause of any Production jobs via problem ticket and releasing then thorough biweekly releases.

Virtusa Consulting Services Pvt Limited (Hyderabad,India)
Client: CITI BANK
MAY 2021-June2022
Project: Data stage to Talend Migration
Responsibilities:
Working in Data Stage exit project converting existing environment to Talend cloud jobs.
Doing reverse engineering and analyzing the existing workflows and stored procedures and Data stage jobs and documenting each stage.
Implementing data ingestion mechanisms based on the different file arrivals and drop box designs.
Worked on customized talend code to build SCD1 frame work jobs for Ciras module.
Consumed Bigdata feeds for ICAPS module which will be generated out of logs from server.
Efficiently designed error handling mechanisms and parameterized all the source and target information at configuration table.
Troubleshooting the talend jobs to identify the root cause of any Production jobs via problem ticket and releasing then thorough MTR process.

SS&C, Hyderabad
Client: DST
July 2019-Apr 2021
Project: Sales Connect 2.0
Responsibilities
Worked as MDM DEVELOPER for one of leading Mutual Funds Company using Semarchy.

Design ETL solution for loading various heterogeneous sources data like Database, XML, CSV, Spreadsheets
and web scrapping data into MDM layer using Talend Bigdata.
Good in writing validation rules and enrichment rules using semql.
Good in creating sql hooks for custom enrichment.
Good in building match and merge rules, surviourship rules and populating golden data.
Managed to delivery of ETL solution (design approach) to process any source file in generic way.
Work closely with Business Users to understand Current application flow using Third party Sales connect
product.
Provide delivery timelines/estimates for deliverables by agile methodology.
Automated the ETL jobs and Exposing API calls to POST data to semarchy layer and GET API calls.
Technologies used: Talend 7.2, Semarchy (MDM Tool), Postman, Jira, Postgres, PG Admin, UNIX, Java






Virtusa, Hyderabad
Client: CITI BANK
April 18 July 19
Project: Reginsight NAM(Abinitio to Talend Migration)
Responsibilities
Working as Data Architect/ ETL Architect(IC) for one of leading bank Citi Corporation.
Understood the Existing Abinitio Graphs and Framework and Redesigned the ETL soluction in Talend
Design ETL solution for evaluating regulatory rules before submitting Federal Corporation in USA.
Good in understanding XFR,Input and output XML schema files.
Manage delivery of ETL solution (design approach) to process any source file in generic way.
Work closely with Business analysts, ETL and BI developers, system testers to help them understand data model design and source to target mappings and data transformation rules
Provide delivery timelines/estimates for various country specific ETL streams.
Automated the ETL process by integrating Talend job as jar file in the user screen to run the job.


Capgemini, Hyderabad
Senior Consultant
Jun 13 Apr 18
Responsibilities:
1. GE Data lake
Created ingestion framework for creating Data Lake from heterogeneous sources like Flat files, Oracle Db, mainframe, SQL server Databases.
Design and Develop ETL Processes in AWS Glue to load data from external sources like S3, glue catalogue and AWS Redshift.
Used dynamodb to log the errors of the ETL process while validating input file with target table structure data type mismatches and all.
Developed complex ETL mappings for Stage, Dimensions, Facts and Data marts load. Involved in Data Extraction for various Databases & Files using Talend
Created Talend jobs using the dynamic schema feature. Have used Big Data components (Hive components) for extracting data from hive sources.
Ingested large size files around 600 GB files to S3 in an efficient way.
Performance tuning - Using the tmap cache properties, Multi-threading and tParallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in ETL process.
Using Glue job read the data from S3 and loaded into redshift tables by reading metadata from data catalogue in json format.
Extensively used S3 bucket, Lambda functions and dynamo Db services from AWS.
Part of data loading into data warehouse using big data Hadoop Talend ETL components, AWS S3 Buckets and AWS Services for redshift database.
Used more components in Talend and Few to be mentioned: tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback components etc. in many of my Jobs Design
Worked on Joblets (reusable code) & Java routines in Talend
Implemented Error Logging, Error Recovery, and Performance Enhancement's & created Audit Process (generic) for various Application teams.
Experience in using Repository Manager for Migration of Source code from Lower to higher environments.
Created Projects in TAC and Assign appropriate roles to Developers and integrated GIT repository.
Worked on Custom Component Design and used to have embedded in Talend Studio
Used to be On call Support if the Project is deployed to further Phases
Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis (Cron Trigger)

2. GES OneMD Analytics

Created Architecture design for GES project for leveraging AWS with Talend.
Created technical design document as per functional specification.
Developed Talend jobs to process files from S3 bucket and generated Dimensions and Facts.
Used various automatic validation techniques which will validate the data between various
layers.
Extensively worked on AWS ecosystem components like EC2, Redshift, S3, copy commands.
Configuring repositories and execution servers to TAC and deployments of developed jobs.
Involved in the upgradation of the project from 6.1 to 6.4 version of Talend.
Created web service call in Talend to pull the data from cvent application in XML files.
Generated CSV files from XML files using Talend job and created staging Tables.
Created semantic views from base tables as per the report requirements.
3.GE Aero Hyperion
Worked for GE as Informatica developer in Production support and enhancements
Monitoring project progress as per scheduled deadlines for various tasks and taking necessary steps to ensure completion within time, cost, and effort parameters
Identify source systems and data required to transfer them into ETL mappings.
Analyse and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements.


United health group (Optum), Hyderabad
Senior software Engineer
Sep 2011- Jun 2013
Responsibilities:
Worked on a Code Migration Project (US based) as designer & developer role in healthcare domain
Meet directly and interact with clients to extract a comprehensive list of needs and analyze what sorts of solutions will provide them with what they need
Identify source systems and data required to transfer them into ETL mappings.
Analyze and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements.


Tech Mahindra (Satyam Computers), Hyderabad
Software Engineer
Jun2007 -Jul 2011
Responsibilities:

Analyse and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements.
Worked for Logistic client as reports developer using oracle report builder 10g.
Identify source systems and data required to transfer them into ETL mappings.
Client/Management reporting for daily testing status. Involvement in status meeting with client/Business Analyst.
Worked as DB tester to validate data into corresponding dimensions and facts.

Education
Bachelor of Technology Electronics and Communications (May 2007)
JNTU- Hyderabad
Additional Information

Completed Data exploration and exploratory analysis on Data Science project
Expert on Semarchy MDM Tool
Talend Certified
Project management process
Keywords: cprogramm business intelligence sthree database rlang information technology microsoft procedural language Ohio

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];119
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: