Home

Hari Reddy - Big Data Engineer
[email protected]
Location: Charlotte, North Carolina, USA
Relocation: YES
Visa: GC
HARICHANDAN REDDY

Sr. Data Engineer
Email: [email protected] 720-372-3616


Professional Summary:

Above 9+ years of experience as Sr. Data Engineer/Modeler with Data Analytics Professional in System Analysis, Data Architecture and Development, Testing and Deployment of business applications.
Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop and Flume.
Hands-on experience in architecting and data modeling for AWS Redshift, AWS Oracle RDS, AWS PostgreSQL and AWS Aurora.
Proficiency with A/B testing platforms and tools (e.g., Google Optimize, Optimizely, Adobe Target)
Good understanding and hands on experience with AWS S3 and EC2.
Experience in designing stunning visualizations using Tableau software and publishing and presenting Dashboards, Storyline on Web and Desktop platforms.
Experienced in generating and documenting Metadata while designing OLTP and OLAP systems environment.
Experience in writing SQL queries and optimizing the queries in Oracle, SQL Server, Netezza, Teradata and Big Data.
Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
Hands on experience in Normalization and De-Normalization techniques upto 3NF for optimum performance in relational and dimensional database environments.
Strong background in various Data Modeling tools using ERWIN, ER/Studio and Power Designer.
Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
Experience in metadata design, real time BI Architecture including Data Governance for greater ROI.
Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
Strong Database experience using Oracle, XML, DB2, Teradata, SQL server, Big data and NoSQL.
Proficient in using Python, SQL, Hadoop ecosystem for extracting data and building predictive models.
Extensive ETL testing experience using Informatica 9x/8x, Talend, Pentaho.
Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export
Experience in conducting Joint Application Development (JAD) sessions with SMEs, Stakeholders and other project team members for requirement gathering and analysis.
Strong hands on experience using Teradata utilities like BTEQ, Fast-Load, Multi-Load, Fast-Export, Tpump, Teradata Manager and Visual Explain.
Strong validation experience of data models by different measures such as AUC, ROC, and confusion matrix.
Proficient in project implementations using various Software Development Life Cycle (SDLC) methodologies like Waterfall, Agile (SCRUM) and RUP.
Experienced in Client-Server application development using Oracle, PL/SQL, SQL PLUS, SQL Developer, TOAD, SQL LOADER.
Expertise in SQL Server Analysis Services (SSAS) to deliver Online Analytical Processing (OLAP) and data mining functionality for business intelligence applications.
Extensive experience in SSIS Packages, SSRS reports and SSAS cubes on production server.
Experience in designing Enterprise Data Warehouses, Data Marts, Reporting data stores (RDS) and Operational data stores (ODS).
Experienced on creation of reports, dashboards, and ad-hoc analysis using the reporting tool Tableau

Technical Skills:

Data Modeling Tools: Erwin r9.6/9.5, ER/Studio 9.7, Sybase Power Designer
Languages: SQL, PL/SQL, ASP, Visual Basic, XML, Python, SQL, T-SQL, SQL Server, C, C++, JAVA, HTML , UNIX shell scripting, PERL.
Big Data Tools: Hadoop, Hive, Spark, Pig, HBase, Sqoop, Flume.
Database: Oracle 11g/12c, MS Access, SQL Server 2016/2014, Sybase and DB2, Teradata14/15, Hive
BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1,SAP Business Objects, Crystal Reports
Operating Systems: Microsoft Windows 8/7 and UNIX.
Applications: Toad for Oracle, Oracle SQL Developer, MS Word, MS Excel, MS Power Point, Teradata, Designer 6i
Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Waterfall Model
Project Execution Methodologies: Ralph Kimball and BillInmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)


Lowes, Morrisville, NC august 2021 Till date
Sr. Data Engineer

Responsibilities:

Massively involved in Data Engineer role to review business requirement and compose source to target data mapping documents.
Design and setting up A/B tests to evaluate changes in elements like copy, design, or functionality.
With IICS, performed various data integration tasks, such as data replication, data synchronization, data migration, and data warehousing. It supports both batch and real-time data integration workflows.
Installed and Configured Open Source Software like Pig, Hive, HBase, Flume and Sqoop.
Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas
Worked with Netezza and Oracle databases and implemented various logical and physical data models for them.
Involved in Normalization/De-normalization techniques for optimum performance in relational and dimensional database environments.
Worked on Amazon Redshift, AWS & Azure and architecting a solution to load data, create data models and run BI on it.
Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL access on Hadoop data
Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture.
Applied Data Governance rules for primary qualifier, Class words and valid abbreviation in table name and Column names.
Worked on Tableau for insight reporting and data visualization
Performed POC for Big data solution using Cloudera Hadoop for data loading and data querying
Created OLAP data architecture, analytical data marts, and cubes optimized for reporting.
Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
Cleaned and maintained the database by removing and deleting old data.
Involved in Logical modeling using the Dimensional Modeling techniques such as Star Schema and Snow Flake Schema.
Worked on Hadoop ecosystem, hive queries, MongoDB, Cassandra, Pig, and Apache Strom.
Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects.
Created SSIS packages for different data loading operations for many applications.
Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from legacy Oracle and SQL Server database systems.
Developed LINUX Shell scripts by using NZSQL/NZLOAD utilities to load data from flat files to Netezza database.
Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
Used Ab-Initio DQE for data quality solution for enterprise-level data processing and data management systems

Environment: Erwin9.6, Informatica, Big Data, LINUX, Teradata, SQL, Oracle, AWS, Netezza, Tableau, Hadoop, Hive, OLAP, OLTP, Flume, NZSQL/NZLOAD, MongoDB, Cassandra.

Verizon, NY November 2019 July 2021
Sr. Data Engineer

Responsibilities:

As a Data Engineer implemented MDM hub to provide clean, consistent data for a SOA implementation.
Gathered and translated business requirements into detailed, production-level technical specifications, new features, and enhancements to existing technical business functionality.
Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
Created and maintained Database Objects (Tables, Views, Indexes, Partitions, Synonyms, Database triggers, Stored Procedures) in the data model.
Worked with other Data Architects to design New Data Mart to design the Google Analytics data reports.
Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop.
Connected to Amazon Redshift through Tableau to extract live data for real time analysis.
Created Rich dashboards using Tableau Dashboard and prepared user stories to create compelling dashboards to deliver actionable insights.
Created Hive architecture used for real time monitoring and HBase used for reporting
Worked for map reduce and query optimization for Hadoop hive and HBase architecture
Worked with DBA group to create Best-Fit Physical Data Model from the Logical Data Model using Forward engineering using Erwin.
Generate DDL scripts for database modification, Teradata, Macros, Views and set tables.
Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase.
Performed data management and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like Perl, TOAD, MS Access, Excel and SQL.
Worked on Naming standards for Table/Column/Index/Constraints names thru Erwin Macros and Master Abbreviations file.
Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
Worked on Physical design for both SMP and MPP RDBMS, with understanding of RDMBS scaling features.
Performed Researching and deploying new tools, frameworks and patterns to build a sustainable Big data platform.
Extensively used agile methodology as the Organization Standard to implement the data Models.
Involved in performing extensive Back-End testing by writing SQL queries and PL/SQL stored procedures to extract the data from SQL Database.
Involved in the validation of the OLAP Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports.
Applies architectural and technology concepts to address scalability, security, reliability, maintainability and sharing of enterprise data
Designed Metadata Repository to store data definitions for entities, attributes & mappings between data warehouse and source system data elements.
Environment: Erwin9.6, Big Data, OLTP, OLAP, SMP, Teradata R13, Teradata SQL Assistant, Hadoop , Hive, Pig, HBASE, DB2, Big Data, Agile, MS-Office, Oracle , SQL Server.

Huntington Columbus, Ohio Sep 2016 0ct 2019
Sr. Data Engineer

Responsibilities:

Worked as a Data Engineer/Architect to generate Data Models using Erwin and developed relational database system.
Lead Architectural Design in Big Data, Hadoop projects and provide for a designer that is an idea-driven.
Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration
Designed Physical Data Model (PDM) using Erwin and Oracle PL/SQL.
Designed Normalization up to 3NF and performed Forward & Reverse Engineering using the Erwin.
Built relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
Involved in writing T-SQL, working on SSIS,IICS, SSRS, SSAS, Data Cleansing, Data Scrubbing and Data Migration.
Developing full life cycle software including defining requirements, prototyping, designing, coding, testing and maintaining software.
Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
Developed Business Intelligence architecture using Microsoft and Tableau products.
Provided guidance and solution concepts for multiple projects focused on data governance and master data management.
Performed extensive data profiling and data analysis for detecting and correcting inaccurate data from the databases and track the data quality.
Developed Source to Target Matrix with ETL transformation logic for ETL team
Participated with key management resources in the strategic analysis and planning requirements for Data Warehouse/Data Mart reporting and data mining solutions.
Managed the meta-data for the Subject Area models for the Data Warehouse environment.
Created data masking mappings to mask the sensitive data between production and test environment.
Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning.
Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data
Worked with Hadoop eco system covering HDFS, HBase, YARN and Map Reduce.
Generated and DDL (Data Definition Language) scripts using ERWIN and assisted DBA in Physical Implementation of Data Models.
Responsible for technical data governance, enterprise wide data modeling and database design.
Developed Data mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.

Environment: Erwin 9.5, Tableau, MDM, QlikView, PL/SQL, HDFS, Teradata 13, JSON, HADOOP (HDFS), MapReduce, PIG, Spark, AWS.
Cube IT Innovations, HydData analyst/Modeler

Responsibilities:

Worked as a Data Modeler/Analyst to generate Data Models using Erwin and developed relational database system.
Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin
Interacted with users for verifying User Requirements, managing Change Control Process, updating existing Documentation.
Developed dimensional model for Data Warehouse/OLAP applications by identifying required facts and dimensions.
Developed Data Migration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW)
Worked very close with Data Architectures and DBA team to implement data model changes in database in all environments.
Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
Performed data cleaning and data manipulation activities using NZSQL utility.
Generated DDL (Data Definition Language) scripts using Erwin 8 and supported the DBA in Physical Implementation of data Models.
Maintained warehouse metadata, naming standards for future application development.
Produced PL/SQL statement and stored procedures in SQl for extracting as well as writing data.
Extensively made use of Triggers, Table Spaces, Pre/Post SQL, Sequences, Materialized Views, Procedures and Packages in Data Models.
Performed data analysis and data profiling using SQL queries on various sources systems including Oracle and SQL Server 2008
Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information.
Facilitated meetings with the business and technical team to gather necessary analytical data requirements.
Development of database objects like Tables, views and materialized views etc using SQL.
Assisted in designing test plans, test scenarios and test cases for integration, regression and user acceptance testing.
Environment: ERWIN 9.0, OLAP, OLTP, SSIS, ODS, PL/SQL, Metadata, SQL Server 2008, Oracle9i
Keywords: cprogramm cplusplus business intelligence sthree active directory information technology microsoft procedural language Delaware New York North Carolina

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];2408
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: