Home

Amarendra Muppalaneni - Data Architect/Solution Architect/Data Engineer
[email protected]
Location: Plano, Texas, USA
Relocation: Remote
Visa: H1B
Amarendra Muppalaneni
Data Architect/Solution Architect/Data Engineer
+1 832-400-9673 EXT 113
[email protected]
Plano, TX
Remote
H1B

PROFILE SUMMARY

An eloquent, focused and dedicated professional with rich knowledge & experience; targeting opportunities as Solution Architect / Data Architect / Data Engineer with an established and growth-focused organization in IT Industry
Results-driven professional leveraging over 13 years of extensive experience in in designing and implementing enterprise-level data solutions, leading data management teams, and driving data-driven decision-making processes
Expert in architecture design and software development, with a strong focus on cloud-based solutions, microservices, and event- driven architectures
Expertise in Data Warehouse for Extracting, Transforming and Loading of data from multiple sources and loading data into a Data Warehouse.
Demonstrated expertise in data modeling, database design, ETL/ELT development, data warehousing, data governance, and data security, utilizing a wide range of technologies and platforms, including Snowflake, Teradata, Greenplum, SQL Server, Oracle, PostgreSQL, Hadoop, Spark, and Object storage
Hands - on development in Spark SQL (Azure Databricks), AWS S3, Python, Hive, Kafka, Hadoop. File format like structured, JSON, PARQUET etc.
Proven track record of successfully leading complex data architecture projects from conception to delivery, ensuring alignment with business objectives, technical requirements, and industry best practices
Adept at conducting technical assessments, analyzing business requirements, and translating them into technical specifications, system designs, and architecture blueprints
Experienced in Microsoft Azure ecosystem/ Informatica Power Center/ IICS/ Streamsets/ ADF (Azure Data Factory) / Azure Databricks and cloud data warehouses like Snowflake and Teradata Vantage.
Proficient in agile methodologies, including Scrum with a strong focus on delivering value incrementally and continuously
Strong analytical and problem-solving skills, with the ability to identify and resolve data-related issues, optimize data workflows, and improve data quality, accuracy, and completeness
Displayed leadership abilities, including managing cross-functional teams, mentoring junior engineers, and collaborating with stakeholders to drive project success
Strong communication skills, with the ability to translate complex technical concepts into clear and actionable recommendations for business stakeholders



TECHNICAL SKILLS

Data Integration / ETL Tools: Informatica Power Center 10.x/9.x/8.x, IICS, Streamsets, Qlik Replicate, Equalum
Data Warehousing: Snowflake, Teradata On-Prem/ Vantage Cloud, VMware Greenplum
Big Data Platforms: Cloudera Hadoop (HDFS / Hive / Pyspark / Impala / Kafka), Azure Databricks
Languages: Python, Oracle PL/SQL
Data Virtualization: Denodo
Query Engine: Starburst Presto
Data Storage Dell EMC ECS / Azure Blob / ADSL Gen2 / AWS S3
Cloud Data Warehouse: Snowflake & Teradata Vantage
Databases / RDBMS: Oracle 19c/18c/12c/11g/10g/9i, Microsoft SQL Server, PostgreSQL
Data Governance: Collibra, Alation
Data Privacy and Classification: BigID
Visualization tool: Tableau, Power BI Desktop
Cloud Technologies: AWS S3, Azure Data Factory, Azure Blob, ADSL Gen2, Azure Databricks
Data Modelling Tool: SAP Power Designer
CORE COMPETENCIES

Solution Design & Enterprise Architecture 8
Data Warehousing and ETL Processes 9
Data Modeling & Database Design 9
Data Integration and Migration 8.5
Cloud Computing 7.5
Proof of Concept (POC) 9
Product Management 7.5
Agile Methodologies 9
Business Analysis 9
Stakeholder Management 7


EDUCATION

B.Tech. (Computer Science Engineering) from Bapatla Engineering College, Acharya Nagarjuna University, Andhra Pradesh, India


WORK EXPERIENCE


Title: Consultant
Client: Dell Technologies Jan 2011 May 2023

Responsibilities:
Key Roles: Commenced employment as a Senior Software Engineer and subsequently advanced to the positions of Advisor, Senior Advisor, and Consultant. Demonstrated versatility by undertaking a range of roles including ETL Developer, L3 Support, Teradata DBA, ETL Architect, Technical PM/PO, Data Architect, and COE Lead / Architect

Key Result Areas:

COE Technical Lead / Architect
Lead and driven the implementation of innovative technology solutions by collaborating with cross-functional teams, overseeing the development of reference patterns and proof-of-concepts, and staying abreast of emerging trends and technologies to optimize business processes and achieve operational efficiency
Spearheaded multiple data acquisition and integrations with various tools like IICS, Qlik replicate, Equalum, Azure Data Factory, Azure Databricks (ETL), Denodo (Data Virtualization), Dremio (Data Lake Query Engine), Starburst Presto (Query Engine), Snowflake (cloud), Teradata Vantage (on-prem/cloud), VMware Greenplum (Data Warehouses) etc.
Strong SQL & PL/SQL coding skills on Snowflake, Teradata, Greenplum etc.
Used AWS glue catalog with crawler to get the data from S3 and perform sql query operations
Hands-on with Tasks, Copy, Stages, File patterns, Zero copy cloning, Snow pipe, streams, time travel etc. on Snowflake Used AWS glue catalog with crawler to get the data from S3 and perform sql query operations
Create AWS Glue data ingestion modules for importing data in multiple levels in S3 and reporting using Athena and Quick sight.
Hands-on with Teradata utilities like Fastload, Mload, Fastexport, Tpump, TPT, BTEQ, NOS etc. and Greenplum utilities like gpload, gpss, PXF etc. Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift
Used JSON schema to define table and column mapping from S3 data to Redshift
Spearheaded and guided team members through mentoring and coaching to enhance their skills and expertise, while overseeing project planning and execution to ensure timely and budget- compliant delivery Worked on AWS Data Pipeline to configure data loads from S3 to into Redshift Worked on Big data on AWS cloud services i.e. EC2, S3, EMR
Implemented the machine learning algorithms using python to predict the quantity a user might want to order for a specific item so we can automatically suggest using kinesis firehose and S3 data lake
As a lead designer of product frameworks, enabled Teradata/Greenplum features like GPSS / TDAM that would effectively reduce ETL costs for the organization
Developed server-side software modules and client-side user interface components and deployed entirely in Compute Cloud of Amazon Web Services (AWS)
Used python Boto 3 to configure the services AWS glue, EC2, S3
Well versed with Big data on AWS cloud services i.e. EC2, S3, Glue.
Demonstrating success in implementing projects from inception, executing migrations, and constructing streamlined pipelines Was responsible for creating on-demand tables on S3 files using Lambda Functions and AWS Glue using Python and PySpark
Data Architect
Built and implemented data architectures to meet business goals by designing data models, flows, and processing systems, and evaluating and selecting relevant technologies for data management
Manage build phase and quality assure code to ensure fulfilling requirements and adhering to ETL, Azure cloud services.
Established and maintained data governance policies and processes, including documentation such as data dictionaries and metadata repositories, to ensure accurate, consistent, and accessible data

ETL Architect
Developed and implemented effective ETL strategy, including tool selection and designing solutions to meet business data requirements
Optimized ETL architecture for superior performance, scalability and accuracy, while collaborating with data experts for consistency across systems
Performed tuning of ETL jobs and SQL s
Development of pipelines using Python, AWS S3, Azure Blob / ADLS Gen2, Azure data factory, Azure Databricks and Teradata Vantage Databases.

Teradata DBA
Designed, implemented, and maintained Teradata database and associated components, including Teradata servers, utilities, storage, code reviews, performance tuning and security
Implemented alerts using scripting and Teradata viewpoint.
Implemented procedure to grant and revoke roles to users
Used data mover extensively to copy sample data from Prod to lower environments for testing
Good understanding of metadata tables and helped Dev teams with impact analysis

L3 Support
Single point of contact for code reviews, impact analysis, defect fixes, tuning queries pertaining to a large Services domain containing over 2K control-m jobs and over 10K informatica sessions

Accomplishments:
Designed data solution for Teradata, Greenplum and Snowflake, enabling data transfer to and from ECS (S3)/ ADLS Gen2 using native utilities, eliminating the need for ETL acquisition and ingestion
Coordinated Oracle Golden Gate implementation for real-time ingestion from Oracle/SQL Server to Teradata / Kafka / ECS/ S3/ADLS Gen2 / Azure Blob
Redesigned DELTA-based table loads, reducing processing time from 4.5 hours to 18 minutes
Identified and dropped unused tables / addressed skew, reclaiming 36.9 TB of space
Tuned high-impact CPU queries resulting in significant reductions in millions of CPU cycles
Determined ETL & Control-M job standards for the entire organization
Identified and resolved delays in Quarter end load process by purging necessary driver tables on Omega database, resulting in 2 hours gain in quarter end SLA's
Created an ORACLE procedure to address a restage issue, saving 550 hours effort
Won the EBI Teradata Innofest contest & Hackathons for innovative solutions
Received various awards, including Game Changer, Champion, LTI, and Inspire Awards for outstanding performance

Environment: Informatica 8.x/9.x/10.x, Informatica Cloud (IICS), Azure Data Factory, Azure Databricks, Streamsets, Qlik replicate, Oracle Golden Gate, Oracle, MS SQL Server, Snowflake, Teradata 13/14/15/16/17, Greenplum databases, PySpark, AWS S3, DELL EMC ECS, Azure Blob, ADLS Gen2, Python, Oracle PL/SQL, Kafka, Denodo 7



Title: Senior Software Engineer
Client: UST Global Jan 2010 Jan 2011


Key Result Areas:
ETL Informatica Power Center with emphasis on acquisition of data from RDBMS and file systems and integrate into Teradata data warehouse
Worked closely with Project management, ETL Architects, data modelers/architects to understand project requirements and develop Informatica workflows/ sessions/mappings.
Collaborated with Business Intelligence and business operational teams to understand data source and implement new requirements.
Extensively used transformations like Joiner, Lookup, Filter, Aggregator, Sorter, Source Qualifier, Sequence Generator, Normalizer etc. and debugger for de-bugging purposes.
Formulated and optimized end-to-end ETL processes to ensure accurate data extraction, transformation, and loading into the Data Warehouse
Cleansed, transformed, optimized queries, and achieved flawless project execution with zero defects in every release

Environment: Informatica Power Center 8.x/9.x, Oracle 10g/11g, Teradata 13.x, MS SQL Server, Linux, Control-M
Keywords: business intelligence sthree information technology purchase order microsoft procedural language Delaware Texas

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];122
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: