Bala Anoop Reddy - Sr Data Analyst |
[email protected] |
Location: Morrisville, North Carolina, USA |
Relocation: Yes |
Visa: H1B |
Sr. Data Analyst
Email: [email protected] Ph.707-893-0022 Professional Summary: Data Analyst with 9+ years of experience in Data Analysis, Machine Learning, Data mining with large data sets of Structured and Unstructured data, Data Acquisition, Data Validation, Predictive Modeling, and Data Visualization. Good knowledge and experience in Software Development Life Cycle (SDLC) and its phase s requirement gathering, Analysis, Design, Implementing, Deployment, and Maintenance. Experience in Data Analysis, Data Validation, Data Modeling, Data mapping, Data Verification, Data loading, and Data mining through understanding various Requirements, analyses, and Designs. Highly skilled in using visualization tools like Tableau, ggplot2, dash, and flask for creating dashboards. Adept in statistical programming languages like R and Python. experience in the Alteryx platform, and involved in data preparation, data blending, and the creation of data models, and data sets using Alteryx. Offloading historical data to Redshift spectrum using Matillion, Python and AWS glue. Working in distributed computing. Defined stakeholders, established workflow processes, defined Roles and Responsibilities, provided training and support, maintained and managed data reporting, and created a user toolkit. Experiencein Data Extraction, Data Management, Data Cleansing, Data Profiling, Data Consolidation, and, Data Quality for various business data feeds. Expert in creating Complex Crystal Reports Ad Hoc Reports, Frequency Reports, Summary Reports, Drill-down, Dynamic grouping, graphical, aging reports Deployed reports on the web using Crystal publishing wizard and scheduled reports using e-Portfolio and Crystal management console Worked on the data architecture framework implementation for Informatica ETL environment. Worked with AWS Cloud platform and its features which include EC2, VPC, RDS, EBS, S3, CloudWatch, Cloud Trail, and Quick Sight. CloudFormation and Autoscaling etc. Sound experience in programming in SQL Server 2000/2005, SSIS 2012/2008r2 Used AWS command line client and management console to interact with AWS resources and APIs, JSON. Experience on Cloud Databases and Data warehouses (SQL Azure and Confidential Redshift/RDS). Experienced in creating Data Flow Diagrams, Use Cases, Use Case Diagrams, Activity diagrams, Entity Relationship Diagrams, Data Mapping, and Data Integration. Created roles and access level privileges and taken care of Snowflake Admin Activity end to end. Converted 230 views queries from SQL server snowflake compatibility. Experienced in creating Data Flow Understanding the client s business problems and analyzing the data by using appropriate Statistical models to generate insights. Excellent technical skills consistently outperformed schedules and acquired interpersonal and communication skills. Hands-on experience with Snowflake utilities, Snow SQL, Snow Pipe, Big Data model techniques using Python. Technical Skills: Programming Languages SQL, Python (NumPy, scikit-learn, Pandas, Matplotlib, etc.), R, HTML, CSS, JSON Machine Learning Linear Regression, Logistic Regression, Decision Tree, KNN, K-Means, Principal Component Analysis (PCA) etc. Databases Oracle, MS SQL Server 2012/2014/2016, MySQL, PostgreSQL Cloud Platform AWS (Redshift, S3, EC2, etc.) Operating systems Windows 7 or more, Windows Server 2008/2012, UNIX, LINUX, IOS Analytical Tools Jupiter Notebook, Excel. Visualization Tools Tableau, Qlik Sense, Qlik View, Power BI, SAS Viya RPA Automation Anywhere Project Management Tools Smartsheet, MS Office (Word, Excel, PowerPoint, Access, Outlook), MS Project PROFESSIONAL EXPERIENCE: Capital one, Plano, TX June 2023 to Present Sr Data analyst/Tableau Responsibilities: Involved in gathering the requirements and analyzing the requirements and interact with the customer to find the proper information. Developed effective and scalable Business Intelligence solutions including complex Analytical, Executive and operational Tableau Reports, Reporting Data Marts, and operational Reports. Prepared dashboards using the Level of Detail concept (LODs), calculated fields, parameters, calculations, groups, sets and hierarchies in Tableau. Created various actions in the dashboard like filter actions to send information to other tableau worksheets and to connect the same in analytical flows, URL actions to connect views of data to external resources and Highlight action to provide rich visual interaction between the views on the sheet. Working with team of developers to design, develop, and implement a BI solution for Sales, Product and Customer KPIs. Strong experience in migrating other databases to Snowflake. Export reports in various formats like MS Excel, PDF, MS Word, Text, and CSV Experience with Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source systems which include loading nested JSON formatted data into snowflake table. Installation and maintenance of PostgreSQL Databases migration to AWS Aurora Developed a 360-business dashboard with multiple panels and parameters that became a favorite sales tool for most of the sales force at CE. Created different Power BI reports utilizing the desktop and the online service and schedule refresh. Implemented Tableau mobile dashboards via Tableau mobile application. Create various complex reports using Crystal Reports XI R2 Developed dashboard prototypes using Cloud Dashboard Tools Looker and AWS Quick sight managing all aspects of the technical development. Compared the performance of each benchmark query among different solutions like Spark, Teradata, MySQL, Presto, Hive (using Tez engine) by creating a bar graph in Numbers. Responsible for Designing Logical and Physical data modeling for various data sources on Confidential Redshift Performed admin tasks related to users, groups, folders, and security using Crystal Management Console CMC. Generated Dashboards with Quick filters, Parameters, and sets to handle views more efficiently. Generated context filters and data source filters while handling huge volumes of data. Built dashboards for measures with the forecast, trend line and reference lines. Working with Project Management Team Foundation Server, SharePoint, Microsoft Project, and JIRA and source code version control SVN environment. Environment: Tableau Desktop 2018, 2019, SQL, Aurora, Tableau Server, data bricks, Ms Access, Hadoop, Snowflake, HDFS, Hive, Redshift, YARN, UNIX shell scripting, Hue, Alteryx, Service now Performance Analytics, SAP, JIRA, SharePoint Honeywell, Tempe, AZ January 2022 to May 2023 Role: Sr. Data Analyst Responsibilities: Extracted, interpreted, and analyzed data to identify key metrics and transform raw data into meaningful, actionable information. Used the Agile Scrum methodology to build the different phases of the Software development life cycle (SDLC). Analyzed User Stories, and concepts, translating data requirements into Conceptual, Logical and Physical Data Assist end users with problems installing the Power BI desktop, installing and configuring the Personal and On-Premises gateway, connecting to data sources and adding the different users. Created 7 PS (all Modules including UI) Training Classes based on info from SME interviews as well as previous SAP Audience size ranged from 8 to 300. Courses were developed with the ability to use for new hires going forward after go-live. Experienced in developing Power BI reports and dashboards from multiple data sources using data blending. Model, updating existing Data Flow, ER diagrams. Designed the Fact and Dimension table for Data Marts using ER/Studio. Collected, validated, and organized core data for use in metrics and process improvement. Developed various solution driven views and dashboards by developing different chart types including Pie Charts, Bar Charts, Tree Maps, Circle Views, Line Charts, Area Charts, Scatter Plots in Power BI. Took part in extraction, aggregation, and quality assurance of data from multiple sources in support of performing operational reporting, client reporting, and quantitative analyses of utilization and healthcare costs. Implemented loading and transforming of large data sets of different types of data formats like structured, semi-structured, and unstructured data. Utilized SQL queries to perform data cleansing through imposing data constraints. Designed the scalable infrastructure required for optimal ETL of data using Airflow, to move data from a variety of data sources to the data warehouse. Performed benchmarking of federated queries in Spark and compared their performance by running the same queries on Presto. Specialized in transforming data into user-friendly visualization to give business users a complete view of their business using Power BI. Worked on data modeling and produced data mapping and data definition documentation. Used SQL queries for organizing and abstracting data from MS Access databases. Performed the Data Accuracy, Data Analysis, and Data Quality checks before and after loading the data. Develop / Auto deploy content using AWS (Amazon Web Services), GIT/Bitbucket, Maven, Jenkins Develop custom tools/scripts/packaging solutions for AEM using Java/Unix Develop integration solutions between AEM, AWS (Lambda, S3, API Gateway, JSON, and Cloud Formation) and Superfast (Social) Platforms Effectively used data blending feature in Tableau to connect different databases like Oracle, and SQL Server. Provided guidance in the writing and review of SAP course materials. This will include working with the functional team to obtain SME input and to facilitate the overall complete development of materials Worked on predictive analytics use cases using Python language. Designed complex dashboards and reports by linking data from multiple data providers, using freehand SQL and functionalities like Combined Queries. Used Alteryx for Data Preparation and then Tableau for Visualization and Reporting. Processed data in Alteryx to create TDE for tableau reporting. Involved in Exploring, Designing, Building, and Deployment high-level and rich dashboards for decision-making of critical metrics using Tableau. Performed data management projects and fulfilled ad-hoc requests according to user specifications by utilizing data management software programs and tools like MS Access, Excel, and SQL Designed & developed various departmental reports by using SAS, SQL, PL/SQL, and MS Excel Performed admin tasks related to users, groups, folders, and security using Crystal Management Console CMC. Provided guidance in the writing and review of SAP course materials. This will include working with the functional team to obtain SME input and to facilitate the overall complete development of materials. Utilized ODBC for connectivity to Teradata via MS Excel to retrieve automatically from Teradata Database. Performed performance tuning in Oracle and SQL Optimizing. During the tuning process developed various indexes and partitions as needed. Created data extracts in Tableau by connecting to the view using the Tableau MSSQL connector. Used data joining and blending and other advanced features in Tableau on various data sources like MySQL Tables and Flat files. Environment: SQL, Python (NumPy/pandas), Microsoft Excel, Qlik Sense, SDLC, Agile, ETL, R, JSON Cigna Corp., Bloomfield, CT June 2020 to December 2021 Role: Sr. Data Analyst Responsibilities: Worked with datasets of varying degrees of size and complexity including both structured and unstructured data. Responsible for the analysis and reporting of consumer data generated for compliance across all touch points from multiple databases. Created SAP End User Training Plan to train SAP end-users in SAP basic navigation to in-depth analysis of Asset Accounting/Financial Accounting, Consolidations, GL financials, PS, and logistics processes. Assessed training community, locations, and preliminary risks/issues Reviewed and analyzed complex, integrated data, model and information from various ERP systems Load data using ETL process dealing with different data sources MS Excel, MS Access and SQL Server, CSV, Flat Files etc. into target database by performing different kinds of transformations using SQL Server Agent to automate the SSIS 2012/2008r2 package execution. Tested and managed the SSIS 2012 packages. Involved in extracting the data from OLTP to OLAP using SSIS. Day to day processing external files from Clients, using ETL some using Legacy DTS Packages for some transfer of data between servers in a SQL Server environment. Generated performance reports cluster/scatter charts along with various other basic charts Involved in Data Extraction, Transforming and Loading (ETL) using various tools such as data cleansing and Data profiling. Utilize SSIS in ETL data extraction, transformation, and loading. Conducted Data Mining and Data Modelling and performed collecting, cleansing, and providing data modelling and analyzing the structured and unstructured data. Experience in creating various views like Line chart, Bar Charts, Pie Chart, Table, Matrix table, Area Charts, tree map, Line and Stacked column chart and Creating Custom Designed and developed ETL workflows and datasets in Alteryx. Processed data in Alteryx to create TDE for tableau reporting. Involved in client meetings with other associates to discuss the schedules, progress of work, requirement changes and other deliverables issues. Creating the Data Modeling and writing the stored procedures for the Target tables used for the report development. Worked with Business Analyst and the Business users to understand the user requirements, layout, and look and feel of the application to be developed. Gave guidance and improved end-user process from current legacy system (PMIS) to fit how the processes would fit in SAP ECC 6.0 Created and maintained data for Training and Events Management Prepared Scripts in Python and Shell for Automation of administration tasks. Created SQL tables with referential integrity and developed advanced queries using stored procedures and functions using SQL server management studio. Created custom datasets using SQL by performing multiple joins on several tables in a database and developed BI solutions tailored to meet the requirements gathered from multiple cross functional teams Defined relationships, created actions, filters, Level-of-Detail expressions parameters, data blending, hierarchies, calculated fields, sorting, groupings, live connections, in-memory in both tableau and excel. Created Tables, Views, Indexes, and other SQL joins. Environment: SQL, matplotlib, NumPy, pandas, tableau, MS Word, MS Excel, Smartsheet, python, UNIX/Linux, Oracle, csv, SQL Server, MS Access, JIRA Verizon, Dallas, TX October 2019 to May 2020 Role: Data Analyst Responsibilities: Extensively involved in requirements gathering and data gathering to support developers in handling the design specification. Created reports utilizing SSRS, Excel services, Power BI and deployed them on SharePoint Server as per business requirements. Provided technical support, analyzing, and evaluating systems, equipment, and process improvements. Retrieved data from massive and complex data sets using statistical techniques and conducted data integration. Performed initial analysis to assess the quality of the data and conducted further analysis to determine the meaning of the data. Created Data Flow Diagrams, Use Cases, Use Case Diagrams, Activity diagrams, Entity Relationship Diagrams, Data Mapping, and Data Integration. Used Amazon IAM to grant fine-grained access to AWS resources to users. Also, managed roles including permissions of users to AWS account through IAM. Used packages like ggplot2 in R Studio for data visualization and generated scatter plots and high low graph to identify the relationship between different variables. Used Python to pre-process data and attempt to find insights. Deployed Reports for Internal, Departmental Executive Reporting based on the requirements using SQL Server Reporting Services 2012/2008r2. Maintained Report Manager and Reporting Server for SSRS to manage routine report production for clients. Used SQL Server Agent to automate the SSIS package execution. Tested and managed the SSIS packages. Designed, developed, and tested data mart prototype (SQL), ETL process (SSIS), and OLAP cube (SSAS). Worked on SQL Server concepts SSIS (SQL Server Integration Services), SSAS (Analysis Services) and SSRS (Reporting Services). Use of data transformation tools such as DTS, SSIS, Informatica or Data Stage. Will be pulling the data from multiple databases (Hadoop, Teradata, Vertica, Oracle, etc.) compiling data into one readable format, and moving it into an internal client tool. Performed Data Extraction, Data Management, Data Cleansing, Data Profiling, Data Consolidation, and Data Quality for various business data feeds. Experience and knowledge of Data Warehouse best practices, Teradata, Hadoop, and Informatica ETL. Will be Developing queries, reports, dashboards, and related reporting and analytical objects for use by end-users, per the defined methodology in Hive and running Unix/Linux commands. I work on the Hadoop Ecosystem tools (e.g., Spark, Map Reduce/Tez) or other big data technologies preferred. Involved in Data Analysis, understanding trends and patterns in data. Performed data migration, Extracting, Transforming and Loading (ETL) data from Excel and various other data sources to Tableau (Upload, Transform and Visualize). Involved in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, Data Export and writing SQL statements Stored Procedures, Functions, Triggers, and packages. Performed in-depth analysis of the data pulled for ad-hoc requests and prepared graphs Using MS Excel and MS PowerPoint. Used SQL to analyze, query, sort and manipulate data according to defined business rules and procedures. Performed data analysis and data validation by writing complex SQL queries using Teradata SQL Assistant. Developed advanced SQL queries to extract, manipulate, and/or calculate information to fulfill data and reporting requirements including identifying the tables and columns from which data is extracted. Involved in writing complex SQL queries using correlated subqueries, joins, and recursive queries. Worked on SQL to extract, transform and load data to the database. Worked on Data mapping, and logical data modeling used SQL queries to filter data within the Oracle database tables. Environment: SQL, matplotlib, NumPy, pandas, tableau, MS Word, MS Excel, Smartsheet, python, UNIX/Linux, Oracle, CSV, SQL Server, MS Access. Amazon, Hyderabad, Telangana, India June2017to September 2019 Role: Data Analyst Responsibilities: Extensively involved in maintaining requirements traceability and communicating any changes to these business and system requirements and release planning to different business stakeholder groups, third party vendors/supply chain partners and technical teams with JAD sessions, solution demos and conduct solution reviews. Interact with internal business users and external vendors to support and initiate Vendor setups, communication, and order releases in a production environment to optimize order and contract management system as well as integrate invoicing modules. Assist as a business process analyst in testing the tickets and SCRs related to IDOC errors and managing a backlog and prioritization service improvement opportunities. Acting as part of a de-centralized Service Management team, upholding the ITSM quality while partnering with the delivery team to maximize velocity. Verify and validate data mappings and specifications data content; met testing guidelines and deadlines, manage VAN activity, and provide application support and enhancements to existing applications. Monitor, investigate, and resolve production incidents; reviewed daily incident error/exception reports analyzed process and recommend improvements and solutions to vendor s contacts, trading partners, Preparing Requirements Traceability Matrix (RTM) and Gap Analysis for additional details on requirements traceability. Creating the Business Process Master List (BPML) and level-specific test plans in collaboration with the Test Architect/Lead. Configuration of P2P & MM-IM requirements, including Plants, Storage Locations, MRP and Spend data. As well, as the Purchasing requirements for internal sources (Stock Transfer Orders), including Purchasing Groups and Special Procurement Types. Gathering MM and WM business process procedures documentation for preparing Test plans. Working with Integration Testing leads to developing Test Scenarios Unit testing, String testing, Integration Cycle Testing and Interface Testing for Make to Deploy (QM/PP), Procure to Pay (MM/SRM), Plan to Schedule, Order to Cash (SD), Hire to Retire (HR), Record to Report (FI) and Lead to Customer (CRM) business processes (BPs). Building and tracking transactional codes, BI, RICEFs and URS (User Requirements Specifications) requirements coverage Reviewing the Test Cases, End-to-End Test Scenarios, completeness, and audit of Test cases. Actively involved in Reporting, Forecasting, Analysis, and Decision Support related to security role management. Actively involved in reading, extracting data from varied file types creating standard reports using MS Access, ECC backend production data and, generating delimited files and SQL and VBA scripts to load data into SQL tables and executed for real-time reporting and analytical needs of executive and organizational reporting. Understand the client s data needs and translate them into meaningful, interactive dashboards and visualizations using Tableau and convert complex reporting needs into technical specifications, including calculations, custom groups, parameters, filtering criteria, aggregations, etc. Created data mappings that helped assess and cleanse the data that will migrate from the data marts of Enterprise Data warehouse (EDW) Teradata to the integration system and ensure that data is effectively transformed by the application of business logic and the business rules are not violated at any point of data integration. Drafted functional requirement specifications (FRSs), data field mapping and data dictionaries for customer master data, quotation processing, sales order processing, consignment processing, billing, returns processing, contract processing, rebates processing, and backorder processing. Ensured the reporting requirements of the analyst team were met in the system and authored progress and completion reports, which were submitted to the program manager and business owner on a weekly basis. Prepare and review UAT test document and developed and uploaded test scenarios for validating the business requirements, mapped business requirements to test scenarios through Requirement Traceability Matrix Environment: Scrum, HTML5, CSS3, AngularJS, Bootstrap, SAP, jQuery, MySQL, Access, Excel, VBA, SAP, SAP ECC. SAP-BW, SAP-BOBJ, API, SOAP, Informatica, XML, REST, JSON, MS Office, JIRA Confluence, Tableau 10.3 Server, Selenium, Moqups, SAP Success factor CSC, Hyderabad, Telangana, India June2015to May2017 Role: Data Analyst/Power BI Responsibilities: Coordinated with business users and business analysts on requirements. Gathered requirements and documented technical and business metadata about the data. Created Excel Pivot and VBA macros for various business scenarios. Extracted and mined data for analysis to aid in solving business problems. Developed reports to visually explore data and create an interactive report Excel and Power BI. Responsible for building data analysis infrastructure to collect, analysis and visualize data. Formulated SQL queries, Aggregate Functions, and database schema to automate information retrieval. Demonstrated ability to move data between production systems and across multiple platforms. Manipulating data to fulfill analytical and segmentation requests. Using Data Visualization tools and techniques to best share data with business partners. Review code and system interfaces and extracts to handle the migration of data between systems/databases. By using Power BI desktop connected to various data sources, working with different visualizations. Created Data Quality Scripts using SQL and Hive to validate successful data load and quality of the data. Developed and published reports and dashboards using Power BI and written effective DAX formulas and expressions. Created several types of data visualizations using Python and Tableau. Utilized Power Query in Power BI to Pivot and Un-pivot the data model for data cleansing and data massaging Designed & developed various departmental reports by using SAS, SQL, PL/SQL, and MS Excel Developed and executed business reports using the Teradata SQL advanced techniques like rank, row number etc. Involved in writing stored procedures and packages using SQL. Environment: SQL, matplotlib, NumPy, tableau, MSWord, MS Excel, Python, UNIX/Linux, Oracle, SQL Server, MS Access. EDUCATION: Bachelor s Degree in Computer Science 2015. Keywords: user interface materials management business intelligence sthree active directory rlang information technology golang container edition business works microsoft procedural language Arizona Connecticut Delaware South Dakota Texas |