Amarendra Muppalaneni - Data Architect/Solution Architect/Data Engineer |
[email protected] |
Location: Plano, Texas, USA |
Relocation: Remote |
Visa: H1B |
Amarendra Muppalaneni
Data Architect/Solution Architect/Data Engineer +1 832-400-9673 EXT 113 [email protected] Plano, TX Remote H1B PROFILE SUMMARY An eloquent, focused and dedicated professional with rich knowledge & experience; targeting opportunities as Solution Architect / Data Architect / Data Engineer with an established and growth-focused organization in IT Industry Results-driven professional leveraging over 13 years of extensive experience in in designing and implementing enterprise-level data solutions, leading data management teams, and driving data-driven decision-making processes Expert in architecture design and software development, with a strong focus on cloud-based solutions, microservices, and event- driven architectures Expertise in Data Warehouse for Extracting, Transforming and Loading of data from multiple sources and loading data into a Data Warehouse. Demonstrated expertise in data modeling, database design, ETL/ELT development, data warehousing, data governance, and data security, utilizing a wide range of technologies and platforms, including Snowflake, Teradata, Greenplum, SQL Server, Oracle, PostgreSQL, Hadoop, Spark, and Object storage Hands - on development in Spark SQL (Azure Databricks), AWS S3, Python, Hive, Kafka, Hadoop. File format like structured, JSON, PARQUET etc. Proven track record of successfully leading complex data architecture projects from conception to delivery, ensuring alignment with business objectives, technical requirements, and industry best practices Adept at conducting technical assessments, analyzing business requirements, and translating them into technical specifications, system designs, and architecture blueprints Experienced in Microsoft Azure ecosystem/ Informatica Power Center/ IICS/ Streamsets/ ADF (Azure Data Factory) / Azure Databricks and cloud data warehouses like Snowflake and Teradata Vantage. Proficient in agile methodologies, including Scrum with a strong focus on delivering value incrementally and continuously Strong analytical and problem-solving skills, with the ability to identify and resolve data-related issues, optimize data workflows, and improve data quality, accuracy, and completeness Displayed leadership abilities, including managing cross-functional teams, mentoring junior engineers, and collaborating with stakeholders to drive project success Strong communication skills, with the ability to translate complex technical concepts into clear and actionable recommendations for business stakeholders TECHNICAL SKILLS Data Integration / ETL Tools: Informatica Power Center 10.x/9.x/8.x, IICS, Streamsets, Qlik Replicate, Equalum Data Warehousing: Snowflake, Teradata On-Prem/ Vantage Cloud, VMware Greenplum Big Data Platforms: Cloudera Hadoop (HDFS / Hive / Pyspark / Impala / Kafka), Azure Databricks Languages: Python, Oracle PL/SQL Data Virtualization: Denodo Query Engine: Starburst Presto Data Storage Dell EMC ECS / Azure Blob / ADSL Gen2 / AWS S3 Cloud Data Warehouse: Snowflake & Teradata Vantage Databases / RDBMS: Oracle 19c/18c/12c/11g/10g/9i, Microsoft SQL Server, PostgreSQL Data Governance: Collibra, Alation Data Privacy and Classification: BigID Visualization tool: Tableau, Power BI Desktop Cloud Technologies: AWS S3, Azure Data Factory, Azure Blob, ADSL Gen2, Azure Databricks Data Modelling Tool: SAP Power Designer CORE COMPETENCIES Solution Design & Enterprise Architecture 8 Data Warehousing and ETL Processes 9 Data Modeling & Database Design 9 Data Integration and Migration 8.5 Cloud Computing 7.5 Proof of Concept (POC) 9 Product Management 7.5 Agile Methodologies 9 Business Analysis 9 Stakeholder Management 7 EDUCATION B.Tech. (Computer Science Engineering) from Bapatla Engineering College, Acharya Nagarjuna University, Andhra Pradesh, India WORK EXPERIENCE Title: Consultant Client: Dell Technologies Jan 2011 May 2023 Responsibilities: Key Roles: Commenced employment as a Senior Software Engineer and subsequently advanced to the positions of Advisor, Senior Advisor, and Consultant. Demonstrated versatility by undertaking a range of roles including ETL Developer, L3 Support, Teradata DBA, ETL Architect, Technical PM/PO, Data Architect, and COE Lead / Architect Key Result Areas: COE Technical Lead / Architect Lead and driven the implementation of innovative technology solutions by collaborating with cross-functional teams, overseeing the development of reference patterns and proof-of-concepts, and staying abreast of emerging trends and technologies to optimize business processes and achieve operational efficiency Spearheaded multiple data acquisition and integrations with various tools like IICS, Qlik replicate, Equalum, Azure Data Factory, Azure Databricks (ETL), Denodo (Data Virtualization), Dremio (Data Lake Query Engine), Starburst Presto (Query Engine), Snowflake (cloud), Teradata Vantage (on-prem/cloud), VMware Greenplum (Data Warehouses) etc. Strong SQL & PL/SQL coding skills on Snowflake, Teradata, Greenplum etc. Used AWS glue catalog with crawler to get the data from S3 and perform sql query operations Hands-on with Tasks, Copy, Stages, File patterns, Zero copy cloning, Snow pipe, streams, time travel etc. on Snowflake Used AWS glue catalog with crawler to get the data from S3 and perform sql query operations Create AWS Glue data ingestion modules for importing data in multiple levels in S3 and reporting using Athena and Quick sight. Hands-on with Teradata utilities like Fastload, Mload, Fastexport, Tpump, TPT, BTEQ, NOS etc. and Greenplum utilities like gpload, gpss, PXF etc. Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift Used JSON schema to define table and column mapping from S3 data to Redshift Spearheaded and guided team members through mentoring and coaching to enhance their skills and expertise, while overseeing project planning and execution to ensure timely and budget- compliant delivery Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift Worked on Big data on AWS cloud services i.e. EC2, S3, EMR Implemented the machine learning algorithms using python to predict the quantity a user might want to order for a specific item so we can automatically suggest using kinesis firehose and S3 data lake As a lead designer of product frameworks, enabled Teradata/Greenplum features like GPSS / TDAM that would effectively reduce ETL costs for the organization Developed server-side software modules and client-side user interface components and deployed entirely in Compute Cloud of Amazon Web Services (AWS) Used python Boto 3 to configure the services AWS glue, EC2, S3 Well versed with Big data on AWS cloud services i.e. EC2, S3, Glue. Demonstrating success in implementing projects from inception, executing migrations, and constructing streamlined pipelines Was responsible for creating on-demand tables on S3 files using Lambda Functions and AWS Glue using Python and PySpark Data Architect Built and implemented data architectures to meet business goals by designing data models, flows, and processing systems, and evaluating and selecting relevant technologies for data management Manage build phase and quality assure code to ensure fulfilling requirements and adhering to ETL, Azure cloud services. Established and maintained data governance policies and processes, including documentation such as data dictionaries and metadata repositories, to ensure accurate, consistent, and accessible data ETL Architect Developed and implemented effective ETL strategy, including tool selection and designing solutions to meet business data requirements Optimized ETL architecture for superior performance, scalability and accuracy, while collaborating with data experts for consistency across systems Performed tuning of ETL jobs and SQL s Development of pipelines using Python, AWS S3, Azure Blob / ADLS Gen2, Azure data factory, Azure Databricks and Teradata Vantage Databases. Teradata DBA Designed, implemented, and maintained Teradata database and associated components, including Teradata servers, utilities, storage, code reviews, performance tuning and security Implemented alerts using scripting and Teradata viewpoint. Implemented procedure to grant and revoke roles to users Used data mover extensively to copy sample data from Prod to lower environments for testing Good understanding of metadata tables and helped Dev teams with impact analysis L3 Support Single point of contact for code reviews, impact analysis, defect fixes, tuning queries pertaining to a large Services domain containing over 2K control-m jobs and over 10K informatica sessions Accomplishments: Designed data solution for Teradata, Greenplum and Snowflake, enabling data transfer to and from ECS (S3)/ ADLS Gen2 using native utilities, eliminating the need for ETL acquisition and ingestion Coordinated Oracle Golden Gate implementation for real-time ingestion from Oracle/SQL Server to Teradata / Kafka / ECS/ S3/ADLS Gen2 / Azure Blob Redesigned DELTA-based table loads, reducing processing time from 4.5 hours to 18 minutes Identified and dropped unused tables / addressed skew, reclaiming 36.9 TB of space Tuned high-impact CPU queries resulting in significant reductions in millions of CPU cycles Determined ETL & Control-M job standards for the entire organization Identified and resolved delays in Quarter end load process by purging necessary driver tables on Omega database, resulting in 2 hours gain in quarter end SLA's Created an ORACLE procedure to address a restage issue, saving 550 hours effort Won the EBI Teradata Innofest contest & Hackathons for innovative solutions Received various awards, including Game Changer, Champion, LTI, and Inspire Awards for outstanding performance Environment: Informatica 8.x/9.x/10.x, Informatica Cloud (IICS), Azure Data Factory, Azure Databricks, Streamsets, Qlik replicate, Oracle Golden Gate, Oracle, MS SQL Server, Snowflake, Teradata 13/14/15/16/17, Greenplum databases, PySpark, AWS S3, DELL EMC ECS, Azure Blob, ADLS Gen2, Python, Oracle PL/SQL, Kafka, Denodo 7 Title: Senior Software Engineer Client: UST Global Jan 2010 Jan 2011 Key Result Areas: ETL Informatica Power Center with emphasis on acquisition of data from RDBMS and file systems and integrate into Teradata data warehouse Worked closely with Project management, ETL Architects, data modelers/architects to understand project requirements and develop Informatica workflows/ sessions/mappings. Collaborated with Business Intelligence and business operational teams to understand data source and implement new requirements. Extensively used transformations like Joiner, Lookup, Filter, Aggregator, Sorter, Source Qualifier, Sequence Generator, Normalizer etc. and debugger for de-bugging purposes. Formulated and optimized end-to-end ETL processes to ensure accurate data extraction, transformation, and loading into the Data Warehouse Cleansed, transformed, optimized queries, and achieved flawless project execution with zero defects in every release Environment: Informatica Power Center 8.x/9.x, Oracle 10g/11g, Teradata 13.x, MS SQL Server, Linux, Control-M Keywords: business intelligence sthree information technology purchase order microsoft procedural language Delaware Texas |