Veeranjaneyulu Tokala - ETL Talend Developer |
[email protected] |
Location: Columbus, Ohio, USA |
Relocation: Remote |
Visa: H1B |
Veeranjaneyulu Tokala
ETL Talend Developer +1630-580-6406. [email protected] Columbus, OH Remote H1B LinkedIn: linkedin.com/in/veeranjaneyulu-tokala-685673b7 Synopsis A competent professional with more than 15+ years of professional experience ETL, Big Data, DW Development, EDA, Data cleaning, Data Wrangling. Involved in handling Talend and Informatica software and in-depth knowledge in managing all the projects under DWH and MDM projects. Develop Low level and High level technical design documents for solution implementation. Handle client communication regarding requirements, design, clarifications, etc. Experience on Data Cleansing and Standardizations of Data. Developing Talend Jobs based on the Mapping Specification Provided Hands on Experience to use Talend Administration Center (TAC) and Code migration. Breaking down the task and Putting down the efforts against it to analyze the business value that you provide to the client. Owning the Ownership of the Jobs Developed and resolving the defects on the fly if any, once when moved to different environments. Performed Unit Testing based on multiple test cases (Positive and Negative Scenarios) and prepared release notes to the framework Joblets. Having a strong knowledge on generating Profiling Reports for creating of Business Rules. Involved in Data Extraction, Transformation and Loading using ETL tool. Extracted data from various cloud data sources like AWS-S3, Cassandra, Snowflake, Datalake including Parquet, Avro, CSV and Flat files. Have designed metadata frame work job and migrated the individual source jobs to Single job and migrated the data into AWS and processed files using Glue into redshift tables. Expertise in the domain of Insurance and Expert in designing ETL /DQ/MDM solutions and Data Migration Project Platforms. Experience on AWS cloud services (EC2, S3, RDS, Redshift, IAM) Good working Knowledge on Data warehousing concepts like Schemas, Dimensions and Fact tables. Implemented end to end solution for Talend DI/ MDM Framework for generating of Golden Customer record across organizational level. Proficient in arranging business requirements (requirement analysis), preparing the detailed process flow for functional requirements An effective communicator with good problem-solving abilities, multitasking ability, willingness to take up ownership, facilitate teamwork and lead activities Worked on various domains like BFSI, Health-care, Education and Logistics. Achievements Provided solution to one of the leading Bank to provide financial health check to improve processing time 5 Hrs to 2 mins. Converted Legacy process interest calculation reports processing time 10 hr to 3 mins by leveraging Hadoop and spark frameworks. Architected and implemented open source solution for an existing product which ended up in saving huge dollars and significant performance numbers. IT Skills Well versed with: o Big Data Technologies: MapReduce, Hadoop, Spark, Hive, Pig, Zookeeper, Sqoop. o Hadoop Distribution: CDH (5.7),HDP o Programming: Scala, Python, Core java, SQL, PL/SQL,R o Web Scrapping o Data Stream: Kafka o MDM Tool: Semarchy(Intelligent Data Hub) o ETL Tools: Talend DI & Bigdata Studio, Talend cloud,Informatica.Abinitio o DB: Microsoft SQL Server, Oracle, MySQL, PostgreSQL, RedShift o OS: Windows, Unix/Linux o IDE: SQL developer, Toad, Jupiter, Eclipse o Dimensional Modelling(Star/Snowflake/Hybrid) o Cloud: Services: AWS, Azure o JIRA, MS-Office, MS-Project Work Experience MAK technologies(Texas,USA) Client:VIRTUSA Jul2022-tilldate Project: AWS data lake Responsibilities: Creating ETL pipeline to load the data from XML files to redshift tables. Extracting the data from Live person conversation messages and process them to load into ods layer. Developing ETL jobs to load from ods tables to dimension and fact tables. Developed framework jobs for generating parameter files and batch start and batch end mechanisms. Consumed Bigdata feeds which will be generated out of logs from server to capture server metrics. Efficiently designed error handling mechanisms and parameterized all the source and target information at configuration table. Troubleshooting the talend jobs to identify the root cause of any Production jobs via problem ticket and releasing then thorough biweekly releases. Virtusa Consulting Services Pvt Limited (Hyderabad,India) Client: CITI BANK MAY 2021-June2022 Project: Data stage to Talend Migration Responsibilities: Working in Data Stage exit project converting existing environment to Talend cloud jobs. Doing reverse engineering and analyzing the existing workflows and stored procedures and Data stage jobs and documenting each stage. Implementing data ingestion mechanisms based on the different file arrivals and drop box designs. Worked on customized talend code to build SCD1 frame work jobs for Ciras module. Consumed Bigdata feeds for ICAPS module which will be generated out of logs from server. Efficiently designed error handling mechanisms and parameterized all the source and target information at configuration table. Troubleshooting the talend jobs to identify the root cause of any Production jobs via problem ticket and releasing then thorough MTR process. SS&C, Hyderabad Client: DST July 2019-Apr 2021 Project: Sales Connect 2.0 Responsibilities Worked as MDM DEVELOPER for one of leading Mutual Funds Company using Semarchy. Design ETL solution for loading various heterogeneous sources data like Database, XML, CSV, Spreadsheets and web scrapping data into MDM layer using Talend Bigdata. Good in writing validation rules and enrichment rules using semql. Good in creating sql hooks for custom enrichment. Good in building match and merge rules, surviourship rules and populating golden data. Managed to delivery of ETL solution (design approach) to process any source file in generic way. Work closely with Business Users to understand Current application flow using Third party Sales connect product. Provide delivery timelines/estimates for deliverables by agile methodology. Automated the ETL jobs and Exposing API calls to POST data to semarchy layer and GET API calls. Technologies used: Talend 7.2, Semarchy (MDM Tool), Postman, Jira, Postgres, PG Admin, UNIX, Java Virtusa, Hyderabad Client: CITI BANK April 18 July 19 Project: Reginsight NAM(Abinitio to Talend Migration) Responsibilities Working as Data Architect/ ETL Architect(IC) for one of leading bank Citi Corporation. Understood the Existing Abinitio Graphs and Framework and Redesigned the ETL soluction in Talend Design ETL solution for evaluating regulatory rules before submitting Federal Corporation in USA. Good in understanding XFR,Input and output XML schema files. Manage delivery of ETL solution (design approach) to process any source file in generic way. Work closely with Business analysts, ETL and BI developers, system testers to help them understand data model design and source to target mappings and data transformation rules Provide delivery timelines/estimates for various country specific ETL streams. Automated the ETL process by integrating Talend job as jar file in the user screen to run the job. Capgemini, Hyderabad Senior Consultant Jun 13 Apr 18 Responsibilities: 1. GE Data lake Created ingestion framework for creating Data Lake from heterogeneous sources like Flat files, Oracle Db, mainframe, SQL server Databases. Design and Develop ETL Processes in AWS Glue to load data from external sources like S3, glue catalogue and AWS Redshift. Used dynamodb to log the errors of the ETL process while validating input file with target table structure data type mismatches and all. Developed complex ETL mappings for Stage, Dimensions, Facts and Data marts load. Involved in Data Extraction for various Databases & Files using Talend Created Talend jobs using the dynamic schema feature. Have used Big Data components (Hive components) for extracting data from hive sources. Ingested large size files around 600 GB files to S3 in an efficient way. Performance tuning - Using the tmap cache properties, Multi-threading and tParallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in ETL process. Using Glue job read the data from S3 and loaded into redshift tables by reading metadata from data catalogue in json format. Extensively used S3 bucket, Lambda functions and dynamo Db services from AWS. Part of data loading into data warehouse using big data Hadoop Talend ETL components, AWS S3 Buckets and AWS Services for redshift database. Used more components in Talend and Few to be mentioned: tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback components etc. in many of my Jobs Design Worked on Joblets (reusable code) & Java routines in Talend Implemented Error Logging, Error Recovery, and Performance Enhancement's & created Audit Process (generic) for various Application teams. Experience in using Repository Manager for Migration of Source code from Lower to higher environments. Created Projects in TAC and Assign appropriate roles to Developers and integrated GIT repository. Worked on Custom Component Design and used to have embedded in Talend Studio Used to be On call Support if the Project is deployed to further Phases Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis (Cron Trigger) 2. GES OneMD Analytics Created Architecture design for GES project for leveraging AWS with Talend. Created technical design document as per functional specification. Developed Talend jobs to process files from S3 bucket and generated Dimensions and Facts. Used various automatic validation techniques which will validate the data between various layers. Extensively worked on AWS ecosystem components like EC2, Redshift, S3, copy commands. Configuring repositories and execution servers to TAC and deployments of developed jobs. Involved in the upgradation of the project from 6.1 to 6.4 version of Talend. Created web service call in Talend to pull the data from cvent application in XML files. Generated CSV files from XML files using Talend job and created staging Tables. Created semantic views from base tables as per the report requirements. 3.GE Aero Hyperion Worked for GE as Informatica developer in Production support and enhancements Monitoring project progress as per scheduled deadlines for various tasks and taking necessary steps to ensure completion within time, cost, and effort parameters Identify source systems and data required to transfer them into ETL mappings. Analyse and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements. United health group (Optum), Hyderabad Senior software Engineer Sep 2011- Jun 2013 Responsibilities: Worked on a Code Migration Project (US based) as designer & developer role in healthcare domain Meet directly and interact with clients to extract a comprehensive list of needs and analyze what sorts of solutions will provide them with what they need Identify source systems and data required to transfer them into ETL mappings. Analyze and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements. Tech Mahindra (Satyam Computers), Hyderabad Software Engineer Jun2007 -Jul 2011 Responsibilities: Analyse and Identify existing database structure and prepare ETL specifications to convert them into new Data Warehouse requirements. Worked for Logistic client as reports developer using oracle report builder 10g. Identify source systems and data required to transfer them into ETL mappings. Client/Management reporting for daily testing status. Involvement in status meeting with client/Business Analyst. Worked as DB tester to validate data into corresponding dimensions and facts. Education Bachelor of Technology Electronics and Communications (May 2007) JNTU- Hyderabad Additional Information Completed Data exploration and exploratory analysis on Data Science project Expert on Semarchy MDM Tool Talend Certified Project management process Keywords: cprogramm business intelligence sthree database rlang information technology microsoft procedural language Ohio |