Home

Bhaskara - Data Analyst
[email protected]
Location: Fairfax, Virginia, USA
Relocation: Yes
Visa: USC
Baskara S Muthusamy
[email protected]
410 929 9023




Summary:
A service oriented, customer focused Senior Data Architect, Data Analyst ,ETL Developer and IT professional with extensive hands on experience in support and delivery of Data Warehouse, Operational Data Store, BI solution, Data Quality, Master Data Management, Meta Data Management, Data Customer Relationship Management projects ,Big Data technologies and Cloud SaaS solutions
Advanced data modelling skills including of Conceptual Data Model (CDM), Logical Data Model (LDM) and Physical Data Models (PDM) for OLTP s and Dimensional Data Modelling (Star/Snowflake), normalization, de-normalization and implementation of Kimball and Inmon Methodology
Over 15 + Years of experience in Banking, Financial, Trading and Healthcare Vertical with HIPPA compliances.
Led DW teams and implemented multiple ETL, Business Intelligence and Data Analytics projects with resources from cross functional and technical teams. Scoped, planned and prioritized multiple projects simultaneously.
Expertise in Microsoft Purview, Informatica AXON for data governance
Hands on experience in designing, architecting, developing and implementing solutions on AWS and Azure.
Experience in core AWS services S3, Lambda, EC2, ELB, EBS, VPC,Auto Scaling and Redshift
Expertise in Informatica PowerCenter ,Informatica Data Quality (IDQ), Informatica Analyst, Informatica Master Data Management (MDM), Informatica Meta Data Manager,WebServices
Hands on experience in BI Reporting using Tableau, Tableau Prep Builder, SQL and R
Experience in Big Data subjects and technologies HDFS, Pig, Hive, Hadoop, HBase, NoSQL, Map R, Apache Spark, Kafka, AWS, Salesforce , Rest API, Java, SOA, XML, JSON, Web Services, Python,Azure ,
Rest API using Azure API Management and Azure Functions, Databricks,Azure Fabric
Experience in CI/CD pipe line tool like Jenkins,GitHub,Jira,Kubernetes Operators for deployments
Hands on experience in BI Reporting Tools using Power BI, Tableau, R, SAS, RStudio, Tidyverse Packages, ggplot2
Experience in AI/ML PyTorch,NumPy, SciPy, Matplotlib,Pandas,Scikit-Learn,NLP,RNN, Evaluated Decision Forests, Na ve Bayes Classification &k-Nearest Neighbors models
Led and participate in the development of short and long-term strategies to drive innovation, decrease speed to market, improve quality, and lower cost.
Experience in Big Data subjects and technologies HDFS, Pig, Hive, Hadoop, HBase, NoSQL,Web Services,Snowflake and Databricks
Extensive experience in Database Development in Oracle,DB2 ,Netezza, MySQL,SQL Server, PostgreSQL,MongoDB,IMS DB
Hands on experience in developing Stored Procedures, Functions, Views and Triggers, Complex SQL queries using Oracle PL/SQL and T-SQL.
Perform Business Analysis ,User Requirement Gathering, User Requirement Analysis, complex data analysis in support of ad-hoc and standing customer requests, develop programs, methodologies for analyzing and presenting data
Experience in creating various Models for AI/ML using Python,R and TensorFlow
Strong background in Autosys, AIX, UNIX, Linux, Shell Scripting (Korn Shell/Ksh) and Perl Scripting.
Involved in full life cycle development and support of Data Warehouse projects. Involved in application support with SEI-CMM LEVEL 5 Software Quality, Process development and Test Methodologies.
Experienced in establishing QA Procedures, QA Standards, proficient in writing and implementing test plans, documenting the testing procedures and writing in-depth test cases to meet the testing cycle.
Extensive experience in conducting requirements gathering sessions to define system requirements via interviews, questionnaires, brainstorming sessions, workshops and prototyping.
Hands on experience in Project planning, execution, monitoring and resource balancing skills with ability to support simultaneous projects in an organizational structure in both Agile/Scrum/Kanban and Waterfall Methods


Education:
Post Graduate Diploma in Computer Application, Loyola College, India.
BSc (Maths & Computer Science), Madurai Kamaraj University, India

Certifications/ Courses:
Certified Informatica Designer.
Informatica Data Quality and Informatica MDM Hub Certification.
IBM Information Framework (IFW/M1) Data Model
Salesforce, Visualforce, Apex
Amazon Web Services (AWS)/Redshift
Talend Studio
AI/ML using TensorFlow

Skills:
DATA
WAREHOUSING: Informatica PowerCenter 10.5.4/10.2, Informatica Developer (IDQ) 10.5.4,
Informatica Analyst, Informatica Metadata Repository, Informatica Metadata
Manager, AXON,Informatica MDM HUB 10.2.x,/10.x/Siperian MDM Hub XU, SIF,
Informatica Power Exchange, Data Stage 7.x, Microsoft SSIS, DTS
BI TOOLS: Power BI,Tableau 2023.2,/2021.4 /2019.2/10.2/9.1, R, ,Business ObjectsX1i /R4 4.2,Crystal Reports 2011,
MicroStrategy 9. x/10.x Microsoft SSRS/SSAS, Hyperion, SAS
DATABASES: Oracle 19C/12C, Sybase ASE/IQ, DB2 UDB 9.5,Microsoft SQL Server 2014, MySQL5.x, Netezza,PostgreSQL, Redshift, IMS DB
LANGUAGES: VBA,Python, R, PL/SQL,UNIX/LINUX Shell Scripting, C++, Java, Perl
WEB TECHNOLOGIES: J2EE, Java Script, ASP, ASP.NET, Web Services, SharePoint/MOSS 2007, AWS
DATA MODEL: ER Studio 9.5.1/8.0, Erwin r9.6/R7, Visio 2007.IBM M1 (IFW),
TOOLS: Toad, Clear Case, Clear Quest, ALM, HP Quality Canter 10, QTP, DOORS 9.6, JIRA, GitHub
Eclipse, MS Project 2007, SVN, Jenkins, Rally, Git, XML Spy, RStudio
ERP: SAP R/3.SAP BW, SAP S/4 HANA
CRM: Salesforce, VisualForce, Apex
SCHEDULER: AUTOSYS, UC4, CONTROL M.
MAINFRAME: IBM 3090, MVS, JCL, CICS, DB2, COBOL.
BIG DATA: HDFS, Pig, Hive, Hadoop, HBase, NoSQL, Map R, Apache Spark, Kafka, Cassandra, MongoDB
AWS/Azure: Amazon RDS, EC2, S3, Lambda, Redshift, DynomoDB, Athena, Snowflake,
Databricks,Azure Rest API ,Azure API Management, Azure Functions,Microsoft Purview
AI/ML : TensorFlow 2.x,NLP,RNN,KNN, Na ve Bayes Classification, SciPy, Scikit-Learn

Core Competencies:
Data Warehouse Application Design
Data Quality Management
Client management
Data Analysis
AI/ML
Master Data Management
Onsite/Offshore Team Management
Data Architecture
Middleware Technologies
Cross Functional Team Leadership
Database Performance and Tuning
Big Data
Technical Project Management



Professional Experience:

USPS, NC 11/20 Till Date
Technical Lead / Data Architect
Application : ETL/Data Quality /MDM

The Project is to Implement ETL , Data Governance ,Data Quality and Master Data Management(MDM) for USPS Enterprise Data Governance team .

Responsibilities:
Responsible for Data Architect, Data Modeling, scoping, mapping, logical and physical model efforts, planning and supporting the migration plan
Involved in Data Analyst ,Data Architect, Data Model for ETL ,Data Quality and MDM Staging Schema
Responsible for identifying Key Data Elements for Informatica Analyst
Created Data Catlog (EDC), data lineage using Informatica AXON for data governance
Worked with Data stewards to identify key data elements for Profiling,Data Masking, Standards,Data Cataloging,Data Quality rules
Responsible for identifying Key Data Elements and data sets for Tableau Reports
Developed standard, ad-hoc and key performance indicator (KPI) reports and dashboards
Create groups, hierarchies, datasets to develop summary ,detail reports and Analytical Reports. Publishing the Reports and Dashboards into Tableau Server.
Create advance analysis actions, calaculations,parameters,trendlines,statistics,maps in the reports.Facilitating Power users in Tableau Dashboard Development.
Created Workflow Tasks in AXON Workflow Designer for Change request,,Perfomed Bulk Updates,Buld Load,Created Dashboard and Widget in AXON
Performed data quality checks, identify the issues, Identify areas for data quality improvements improvements, implement fixes and improvements.
Resolve data quality problems through the error detection and correction, process control and the improvement.
Design, developed and deliver data quality and data profile best practices.
Involved in Development of ETL, Rest API, REST Web Service,REST API,Data Service, complex Informatica Mappings,Mapplets,Workflows and Informatica Application ,Workflow Job Scheduler
Involved in Performance tuning for Informatica Applications , Rest API,Database and SQL queries and Optimizations.
Developed Rest API for MDM Golden Records and Vehicle Tracking Reports
Developed Dashboard and Analytical Reports using Power BI, Azure Synapse
Identify Key attributes and Generate Analytical Reports for Mail Fraud detections.
Developed SQL Scripts for ETL, Unix Shell Scripts (ksh/sh) for AFT process and DQ Application for Informatica Scheduler
Involved in Informatica B2B Applications development and testing for JSON,Excel and Relational databases
Involved in migration of NO SQL scripts into Relational SQL
Create Complex PL/SQL,SQL scripts for ETL Load and QA Testing.
Developed Python scripts for standardize and cleanse the JSON,XML,Excel files for data anomaly
Developed DQ Rules ,Profiles and Scorecards in Informatica Analyst and Tableau
Define and Describe Business terms ,Concepts in glossary for Informatica Analyst
Responsible for Production Support for the Informatica ETL and DQ Applications.
Generate Informatica Applications XML for Export and Import into CAT and Production environment
Involved in creating Change Request ticket in Service Now for CAT and Production environment
Created data pipeline for new Machine Learning process Models using Python and TensorFlow for NLP

Environment: Oracle 19.x, Oracle PL/SQL,Erwin,Informatica Developer 10.5.4,Informatica PowerCenter 10.5.4, Informatica Analyst 10.5.4, Informatica AXON,Python 3.9, Linux,Unix, Hive,Informatica MDM Hub, AI/ML,NLP,RNN,EDC and EDG,ServiceNow,GitHub,Anoconda 2.4.2,ALM, MongoDB, Microsoft Purview, Power BI, SAP S/4 HANA,Neo4j,Azure Synapse,R,RStudio, ggplot2,forcats, Tableau Desktop 2023.2,Tableau Prep

DHS,MD 06/19 11/20
Technical Lead / Data Architect / Sr. ETL Developer
Application : ETL/ Data Warehouse/Big Data
The Project is to Implement ETL ,Big Data solutions on AWS for MD Think and CMS Projects (Centers for Medicare and Medicaid Service) for Midas Market Place Model

Responsibilities:
Responsible for Data Architect, Data Modeling, scoping, mapping, logical and physical model efforts, database planning, Data Lake and supporting the migration plan
Responsible for Design, Architect, Implement and support big data solutions, Design and development of cloud-based solutions, technical design and implementation of the architecture
Responsible for Data Architect, Data Modeling, logical and physical model for Midas Market Place Model
Responsible for Design, Development, Implementation of ETL and Data Science Applications
Research new technologies and products for their applicability to business processes, compare various solutions and determine the one which best fits the need
Responsible for the development efforts ,meeting the agreement with standards
Responsible for Oracle to PostgreSQL migration and conversion .
Design, development of ETL process in Informatica PowerCenter using Informatica Mappings, Mapplets, Workflows and Stored Procedures, Involved in Performance tuning of Informatica mappings
Responsible for data analysis, identifying the attributes ,Dimensions and Facts for the Reports
Implemented Data Lake solutions on AWS for Data Analysis
Design and Developed dashboards, create and maintain data visualizations reports in Tableau.
Develop parsing process to create DML, DDL for tables using Python.
Identify data points for Machine Learning process using Python ,TensorFlow
Design and Architect Kafka Streams using StreamSets for AWS Elastic Search.
Developed Spark SQL in Databricks for Data insights and Analytical Reports.

Environment: PostgreSQL, Oracle 12.2, Toad, Oracle PL/SQL, Informatica PowerCenter BDM 10.2, Power Exchange, Python 3.7, Linux, AWS, Lambda, EC2, Apache Spark, Scala, Cloudera Hadoop HDFS, Hive, Impala, Mongo DB, Kafka, GitHub, Rally, TensorFlow, Tableau 2019.x, R, Erwin, PGAdmin, Databricks,JIRA,Jenkins, MongoDB, Tableau Prep,Java, IMS DB

Common Securitization Solutions 05/17 05/19
Technical Lead / Data Architect / Sr. ETL Developer
Application : ETL/ Data Warehouse
The Project is to Implement ETL and Data Quality solutions on AWS for various modules like Disclosure, Bonds, Tax and Issuance, to automate the data validation process and to perform Reporting and Monitoring for Reconciliation Process.
Responsibilities:
Responsible for Data Architect, Data Modeling, scoping, mapping, logical and physical model efforts, database planning, Schemas ,Data Lake and supporting the migration plan
Design, development of Informatica Mappings, Mapplets, Workflows, Worklets using Informatica PowerCenter and Implemented performance tuning logic on Informatica Targets, Sources, Mappings and Sessions to provide maximum efficiency and performance.
Responsible for data analysis, identifying critical data elements, KPI, Measures and Metrics
Develop Unix Shell Scripts to parse flat files, Python Scripts to parse JSON files and to download files from AWS S3 buckets.
Data Extracted from AWS S3 files, JSON are loaded into Database
Loaded the aggregate data into a relational database for reporting, dash boarding and ad-hoc analysis.
Integrated new tools and developed technology frameworks/prototypes to accelerate the data integration process and empower the deployment of predictive analytics.
Implemented Data Lake solutions using structured and unstructured data on AWS for Data Analysis
Design and Developed dashboards, create and maintain data visualizations reports in Tableau.
Developed Oracle PL/SQL Scripts ,SQL Scripts and UNIX Shell Scripts.
Environment: Tableau 2019.1, Tableau Prep ,Oracle 12.2, Toad, Oracle PL/SQL, Informatica PowerCenter 9.6.1, Python 2.6, Linux, AWS, Redshift, S3, Athena, Lambda, EC2, Apache Spark, Tableau 9.1, Tableau Prep,Bash Scripting, R, Snowflake, Cassandra, GitHub, Rally

Fannie Mae 04/11 04/17
Technical Lead / Data Analyst / Sr.ETL Developer
Application : Data Quality, Data Governance and EDI Applications
The Project is to Implement ETL, MDM for Security Master and Multifamily Business-focused Data Quality Rules to automate the data validation process and to perform Data Analysis, Profiling, Cleansing, Standardization, Matching, Consolidation, Reporting and Monitoring. Maintaining centralized control and management of data quality standards. Provide management level reports that measures the Key Performance Indicators (KPI), Subject Area, key data quality dimensions such as Reasonableness, Completeness, Consistency, etc. Reconciliation between various sources and target system.
Responsibilities:
Conducted JAD sessions with Management, Business Analysts, SME, Users in capturing and understanding/analyzing business requirements, defining project scope and communicate with other stakeholders for open and pending issues.
Developed Tableau Dashboards, Reports for multidimensional analysis and Enhancements and Maintenance of reports
Generated interactive dashboards in Tableau with Quick filters, Parameters. Created adhoc reports to users by connecting various data sources. Published Workbooks in Tableau server
Design the ETL processes to load data / real time data into the target Oracle database from multiple sources such as XML, flat files and databases.
Involved in the development of Informatica Mappings, Mapplets and Workflows using Informatica PowerCenter and Informatica Developer for complex data quality rules. Created various tasks like Session, Command, Timer and Event wait.
Implemented performance tuning logic on Informatica Targets, Sources, Mappings and Sessions to provide maximum efficiency and performance.
Modified several of the existing mappings based on the user requirements and maintained existing mappings, sessions and workflows.
Participate in establishing and evolving architecture standards, reference models, guidelines and principles
Lead Role in Design, Architect and Implementation of Data Quality Projects.
Responsible for implementing MDM standards, MDM best practices, procedures, enhancements and support for Security Master ETL/MDM application for EDI.
Developed and implemented SOA using XML/XSD and Web Services for data integration.
Developed Oracle PL/SQL Scripts ,SQL Scripts and UNIX Shell Scripts.
Involved in development, implementation, and management of data strategies that optimize data quality and improve standardization and business information value derived from enterprise data.
Perform detailed analysis to support the planning and impact assessment for Data Quality initiatives.
Identify Key data elements for Mortgage Fraud Prevention,Identiy Theft,Income Misrepresentation ,loans originated by third-party originators and generate Analytical reports using AI/ML with Python,R and Databrics
Involved in Data Quality Management, Data Resource Management, Data Stewardship and Data Administration, Maintaining Data Standards, Data Governance.
Responsible for Data Architect, Data Modeling, scoping, mapping, logical and physical model efforts, database planning, Schemas and supporting the migration plan
Developed BOXI Crystal Reports, Tableau Dashboards, Reports for multidimensional analysis using R Scripts.
Generated interactive dashboards in Tableau with Quick filters, Parameters. Created adhoc reports to users by connecting various data sources. Published Workbooks
Responsible for creating data profiling methodology, identifying known data quality issues and communicate with team for handling the data.
Perform validation for Golden Records, Business Glossary, Catalog and Lineage analysis for IMR (Informatica Meta Data repository).
Involved in data quality development and support initiatives by designing new data models, helping to track down sources of data anomalies, etc.
Implemented DQ rules in Informatica Analyst (IDQ) for Adhoc Profiling and analysis for rules exceptions
Work with the Agile team on complex reporting / dashboard solutions, developed self-reporting BI platform to the end user community
Involved in Master Data Management, production support, QA/Testing, developing test scripts for Data Quality projects and reports, executing test cases in Quality Center and ALM.
Environment: Tableau 9.1, Tableau Desktop 9.1,Oracle 11g/12C,Toad, Informatica PowerCenter 9.6.1,Informatica MDM Hub 9.7.x, IDQ Developer 9.6.1, IDQ Analyst, 9.6.1, Informatica Metadata Manager, PowerExchange, Salesforce, TOAD, Linux, Shell Script(Ksh), Autosys, Tableau 9.1,R, RStudio, ggplot2, forcats ,Business Objects R4 (BOXIR4) /Crystal Reports 2011,ER Studio 9.5.1,TIBCO, Eclipse IDE, Java, SVN, Jenkins, AWS, Netezza 6.1.x, SQL Server 2014

Pfizer Inc 09/10 04/11
Technical Lead / Data Architect
Application: Heath Care Provider Information Systems (HCP)
Project Description
Data feeds from Pfizer applications from different countries extracted, standardized, cleansed before loading into Siperian MDM for Customer and Address Intelligent Matches for Heath Care Provider Information Systems (HCP).
Responsibilities:
Developed ETL procedures using Informatica PowerCenter and custom PL/SQL.
Created Informatica mappings for populating the data into the dimension, facts, and history tables from different sources systems.
Developed Workflows using task developer, worklet designer, and workflow designer in Workflow manager.
Tuned the performance of mappings by following Informatica best practices and also applied several methods to get best performance by decreasing the run time of workflows.
Developed procedures for monitoring of daily/weekly/monthly ETL processes and backup procedures for possible ETL load failures ,developed scheduler to run Informatica workflows
Responsible for development of Informatica Mappings, Mapplets, and Workflows using Informatica PowerCenter and Informatica Developer for complex data quality rules
Understand requirements to create, validate and implement Logical and Physical Data Models, DDL and Scripts.
Created target state data model, identified attributes for Landing, Staging Area to be maintained in the Siperian MDM Hub.
Develop process for data profiling, standardization, cleansing the data before loading into Siperian MDM Hub for match and merge for Customer and Address with HCP systems.
Define, advice and write rules to enhance MDM efficiencies in support of client data program
Analyze, validate, and interpret the content and quality of data within the Medical applications, Transactions, data stores and warehouses using Informatica Data Quality.
Perform detailed data analysis and present findings in a cohesive, high impact format to drive and inform business and technical decisions
Develop and maintain ORS Model and MDM rules and guidance etc.
Responsible to perform functional and performance testing.
Environment: Oracle 11g,TOAD,Unix, Siperian MDM Hub XU,SIF, Business Data Director (BDD),Informatica PowerCenter 9.1,Informatica Data Quality 9.1,Erwin7.3.x, Windows 2003 Server, Informatica Metadata Repository, Netezza 6.1.x,Business Objects X1i,JBoss Application Server, Visio

Walt Disney 03/10 08/10
Technical Lead / Data Architect
Application : Enterprise Data Warehouse
Data Extract from SAP BW and Loading to ODS/FDM. Various Segments OPA, DCP (Disney Consumer Products) and Studios loads these data into their respective segment Data Marts, Cognos Cubes, Planning and Forecasting is used to generate reports using this data
Responsibilities:
Developed Informatica mappings for extracting, cleansing, transforming and loading into Data Warehouse
Involved in performance tuning for Informatica Mappings and databases to identify and remove processing bottlenecks, implementing the best practice for development
Maintained source and target mappings, transformation logic and processes to reflect the changing business environment over time.
Used various transformations like Filter, Router, Expression, Lookup Aggregator, Sequence Generator,
Update Strategy, Joiner, Normalizer, Sorter and Union to develop mappings in Informatica Designer. Extensively used workflow variables, mapping parameters and mapping variables.
Created sessions, batches for incremental load into staging tables and scheduled them to run daily.
Used shortcuts to reuse objects without creating multiple objects in the repository and inherit changes made to the source automatically.
Involved in Creating Stored Procedures, Functions, Views, Index for loading the staging tables.
Performed Data Mapping and Data Migration from heterogeneous sources into the Data Warehouse
Involved in development of Perl and UNIX (Ksh) Shell Scripts to generate various file formats and reports.
Responsible for the design and delivery of a new Data Marts
Involved in Data Architect, Data Model, ETL Design and Development
Preparing release notes, deployment of Workflows and Jobs to test and production environment
Created attributes for the dimensions/Fact tables based on the traceability matrix & source systems.
Serve as a liaison between business and technical teams, database administrators and develop detailed functional, data and technical requirements
Defined Target Load Order Plan and Constraint based loading
Developed various Reports, Interactive Dashboards with drill-down capabilities, with various charts and views using filters. Created parameterized reports, Drill down and Drill through reports.
Automated and scheduled UNIX shell scripts for Informatica sessions and batches
Coordination with offshore in development activities and QA testing.
Environment: Oracle 11G,AIX, DB2 UDB 9.5,SQL Server 2005, Shell Script, Windows 2003 Server, Informatica PowerCenter 8.6, Erwin7.3, SAP R/3,SAP BW, MicroStrategy 9.x

Previous Clients:
Freddie Mac, Great American Insurance, ITG(Investment Technology Group),Standard Chartered Bank, Patheon, DBS Bank, Deutsche Bank, URA, Lucent Technologies, Housing Development Board, OKI, Harvard Medical School
Keywords: cplusplus continuous integration continuous deployment quality analyst artificial intelligence machine learning business intelligence sthree database sfour active directory rlang information technology business works hewlett packard microsoft procedural language Delaware Maryland North Carolina

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];1258
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: