Ganesh - data analyst |
[email protected] |
Location: Aledo, Texas, USA |
Relocation: yes |
Visa: h1b |
GANESH KURRA
Data Analyst (404)-939-3457 [email protected] PROFESSIONAL SUMMARY: Having 7+ years of dedicated experience as a Data Analyst within the product domain industry, my expertise extends across ETL Development and Data Modeling realms. Performed Data Cleaning, Data Profiling, Data Analysis and Data Mapping operations to bring more insights onto the raw finance data and creating Data Flow Diagrams, Use Cases, Use Case Diagrams, Activity diagrams, Entity Relationship Diagrams and Data Integration. Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, Data Export and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers, and packages. Proficiently managed and documented code repositories on GitHub, enabling efficient code sharing and collaboration among team members. Created Python-based data models to support predictive analytics and statistical analysis, driving data-driven decision-making for key business processes. Created compelling and data-driven PowerPoint presentations to communicate complex analytical findings and business insights to stakeholders and executive teams. Developed visually engaging slides that transformed raw data into clear, actionable insights, using charts, graphs, and infographics to illustrate key trends and metrics. Built and managed ETL data pipelines using Informatica Cloud Data Integration (CDI) to streamline the flow of data across multiple systems, ensuring seamless integration and high data quality. Leveraged Informatica Suite (CDI, CAI, CDQ) to manage complex data transformations and automate data workflows, improving operational efficiency. Ensured reference data quality for both Salesforce and non-Salesforce sources, performing validation and data cleansing to maintain accurate and reliable datasets. Utilized Snowflake as a cloud-based data warehousing solution, optimizing query performance and enabling large-scale data analytics. Created data frames and performed high-level analytics using Python, leveraging libraries like Pandas and NumPy to derive actionable insights from datasets. Developed basic data quality metrics and dashboards using Tableau Desktop to monitor data accuracy and integrity, providing visual insights to key stakeholders. Worked with project management tools like Jira to track issues, manage tasks, and collaborate effectively within agile teams. Managed version control of Python scripts using Git, ensuring code integrity and facilitating collaboration within the analytics team. Developed and maintained Python-based tools for data validation and quality assurance, ensuring the accuracy and reliability of data used in analysis. Utilized Python to perform data profiling and exploratory data analysis (EDA), identifying trends, anomalies, and correlations in large datasets. Automated data import and export processes with Python, enabling seamless integration between various data sources and analytics platforms. Experienced in working with Spark SQL for data processing and querying, contributing to effective data processing and extraction. Designed and automated ETL workflows using Alteryx to streamline data integration processes. Utilized Alteryx for data cleansing, transformation, and preparation, enhancing data quality for analytical reporting. Developed and maintained ETL pipelines using Databricks to integrate data from various sources, improving data accuracy and timeliness. Involved in identifying the defects and developed a Defect Tracking report using Mercury Quality Centre this involved the Bug life cycle. Collaborated with cross-functional teams to design and implement data models in Databricks, enhancing data accessibility and reporting efficiency. TECHNICAL SKILLS: Big records Technology: Hadoop, MapReduce, Hive, Presto, Apache Spark (PySpark), Sqoop, Apache Airflow, Autosys, Snowflake, Teradata, Oracle, RDMS,Glue, Python, Scala. Database/Datawarehouse: Oracle, SQL Server, Snowflake, Teradata, Redshift Programming languages: Python, Scala , Java and UNIX shell scripting CI/CD Tools: Jenkins, GitHub, Jira, Confluence, Tableau IDE &Tools: Eclipse, PyCharm, MS access, MS Excel Cloud Platform: GCP,AWS, Cloudera, Microsoft Azure. Certifications: Google Data Analytics, AWS Cloud Solution Architect, Microsoft Power BI Data Analyst, Tableau Intelligence Analyst Professional Certificate PROFESSIONAL EXPERIENCE: Client:Elevance Health , Texas Oct 2023 to till date Role: Data Analyst Developed interactive dashboards and reports in Power BI to visualize key performance indicators (KPIs) and metrics related to patient care, cost efficiency, and operational effectiveness. Integrated external health data sources, such as public health databases and research repositories, into Databricks for comprehensive analysis and enhanced data accuracy. Developed custom data models and algorithms to identify potential healthcare fraud and abuse, improving compliance and reducing financial losses. Designed and optimized ETL workflows in Informatica CDI, improving data throughput and minimizing latency in high-volume data processing environments. Conducted in-depth data audits using Informatica CAI to ensure compliance with industry standards and internal data governance policies. Developed and maintained real-time data pipelines, leveraging Informatica CDI and Snowflake, ensuring seamless integration and up-to-date information for analytics and reporting purposes. Defined and monitored key performance indicators (KPIs) for data quality using Tableau Desktop, providing a proactive approach to identifying and resolving data discrepancies. Implemented automated data reconciliation processes between Salesforce and non-Salesforce systems, minimizing data inconsistencies and ensuring unified data views across platforms. Customized data transformation logic using Informatica CDI, creating reusable components that streamlined the data migration process and reduced project turnaround times. Developed Python-based data processing scripts to clean and prepare datasets for machine learning models, enhancing the organization's ability to perform predictive analytics. Collaborated with business stakeholders to gather requirements for data quality improvement initiatives, leveraging Informatica CDQ to implement rule-based validations. Developed and maintained attribution models to track and measure the impact of different touchpoints in the customer journey, ensuring accurate performance evaluation and optimization of marketing spend. Designed and executed multi-channel attribution studies, integrating data from different sources to provide a holistic view of marketing effectiveness and customer behavior. Conducted risk stratification analysis to identify high-risk patients and implement targeted interventions for improved care management and outcomes. Worked closely with clinicians and healthcare providers to translate data insights into actionable recommendations for clinical practice and care delivery. Designed and developed data pipelines on AWS using services such as S3, Lambda, and EMR, enabling seamless data movement and processing at scale. Utilized AWS Athena for ad-hoc querying of data stored in S3, facilitating quick and efficient data analysis for business stakeholders. Implemented data lake architecture on AWS, consolidating data from various sources into a centralized repository for analytics and reporting purposes. Environment: AWS, EC2, Redshift, Lambda, Power BI, SQL,Jira , MS Excel ,Python , Teradata, GIT, Athena, Snowflake, SQL server, Tableau, Airflow, MongoDB. Client: Apple, Texas Jan 2023 to Sep 2023 Role: Data Analyst Responsibilities: Utilized Guide wire Insurance Now and Guide wire Policy Center for both Personal Lines and Commercial Lines in the Property and Casualty Insurance domain. Applied Databricks for efficient data processing and analysis within the Property and Casualty Insurance domain. Facilitated Agile project management by creating and managing Jira boards, sprints, and backlogs, ensuring alignment with project goals and timelines. Generated detailed reports and dashboards in Jira to provide stakeholders with insights into project status, team performance, and deliverable progress. Prepared and presented ad-hoc analysis and project updates in PowerPoint, providing stakeholders with timely and relevant information on key metrics and performance indicators. Ensured data accuracy and consistency in PowerPoint presentations by cross-referencing with source data and analytical outputs, maintaining high standards of quality in all reports. Utilized Databricks SQL and Apache Spark to handle large-scale data sets, performing complex queries and transformations to support data-driven projects. Engineered end-to-end ETL pipelines using Informatica CDI, integrating disparate data sources into a centralized repository while ensuring minimal data loss and high data accuracy. Customized and deployed Informatica Cloud Application Integration (CAI) workflows to automate cross-system data synchronization, improving operational efficiency and reducing manual intervention. Designed complex data transformation logic in Informatica CDI, enabling the efficient handling of unstructured and semi-structured data within cloud and on-premise environments. Led initiatives to improve reference data quality by implementing robust validation rules and cleansing processes within Informatica CDQ, significantly enhancing data reliability and trustworthiness. Performed detailed data quality assessments using Informatica Data Quality (IDQ) tools, applying industry-standard metrics to measure and improve the quality of data flowing through ETL pipelines. Integrated Informatica CDI with cloud platforms such as Snowflake, enabling scalable data storage and processing solutions that optimized performance in large datasets. Developed and executed data transformation scripts using Azure Data Factory and Azure Databricks, improving data quality and streamlining ETL processes. Conducted data analysis and generated actionable insights using Python for data manipulation and statistical analysis, integrating Python scripts with Azure services for enhanced analytical capabilities. Integrated data from multiple sources, including on-premises databases and cloud-based systems, using Azure Data Factory and Azure Synapse Analytics to provide a unified view of business data. Utilized Azure Machine Learning to build and deploy predictive models, incorporating machine learning algorithms to enhance data analysis and forecasting capabilities. Implemented data security measures and compliance protocols in Azure environments, ensuring data privacy and adherence to regulatory requirements. Managed incidents and service requests using ServiceNow, enhancing IT service management through efficient tracking, resolution, and reporting. Employed JSON for RESTful API development and data serialization, improving data exchange and system interoperability. Involved in translating business concepts into XML vocabularies by designing XML Schemas with UML. Worked on following applications Business Objects, Enterprise Architect, Toad, Plainview (project management), Microsoft Suite (Word, Excel, PowerPoint, Visio, Access, Project). Environment: Azure, Azure Virtual Machines, Azure Blob Storage, Azure Data Factory, Power BI, Airflow, SQL, MS Access, MS Excel, Python, Teradata, GIT, Apache Spark, Azure Synapse Analytics, Snowflake, SQL Server, Tableau, Microsoft Purview, Oracle Client: Ryder System , Georgia Jan 2022 to Jan 2023 Role : Data Quality Analyst Responsibilities: Worked with the fundraising teams to gather data migration and ETL requirements and preparing requirement documents. Performed Data Cleaning, Data Profiling, and Data Analysis and Data Mapping operations to bring more insights onto the raw finance data. Conducted exploratory data analysis (EDA) and feature engineering in Databricks to support machine learning models and enhance predictive analytics capabilities. Created and automated complex dashboards and reports in Databricks, providing actionable insights and performance metrics to drive business strategies and initiatives. Utilized Google Ads and Microsoft Ads platforms to analyze advertising campaign performance, identifying trends and generating actionable insights that significantly improved ROI. Created detailed reports and dashboards to track key performance metrics such as CPC, CPA, and conversion rates, using tools like Google Analytics and Microsoft Power BI. Developed and implemented optimization strategies based on data analysis, leading to enhanced cost-efficiency and improved click-through rates. Conducted A/B testing on ad creatives and landing pages, providing data-driven recommendations that increased user engagement and conversion rates. Integrated data from Google Ads and Microsoft Ads with other marketing data sources to produce comprehensive performance reports, aiding in strategic decision-making. Automated routine data extraction and reporting processes using Python and SQL, significantly reducing the time required for report generation. Monitored and analyzed advertising spend across various platforms, ensuring alignment with budgetary constraints and optimizing resource allocation. Performed trend analysis and market research to identify emerging opportunities and potential risks in digital advertising strategies. Designed and developed interactive reports and dashboards in Power BI to visualize data from various sources, including SQL databases, Excel files, and web services. Implemented data modeling techniques in Power BI to create relationships between different data tables and optimize query performance. Used DAX (Data Analysis Expressions) in Power BI to create calculated columns and measures for advanced data analysis and calculations. Collaborated with business users to gather requirements and customize Power BI reports to meet their specific needs and preferences. Proficient in utilizing Tableau for data visualization and analysis, demonstrating the ability to create interactive and insightful dashboards to support informed decision-making. Experience in leveraging Tableau's advanced features for complex data modeling, including handling large datasets, creating calculated fields, and implementing effective data blending techniques. Strong understanding of Tableau's integration capabilities with various data sources, ensuring seamless connectivity and extraction of valuable insights from diverse datasets. Environment : SQL, MS access, MS excel,Python , Airflow, Microsoft Purview, Oracle, Teradata, Power BI, GIT , AWS, Redshift, Mysql, Snowflake, SQL server, Tableau. Client: Cloud Big Data Technologies Group, India March 2017 to Dec 2020 Role: Data Analyst Responsibilities: Created Source to Target Mapping documents, documented business and transformation rules and participated in working sessions, and ensured full business participation throughout the process. Demonstrated expertise in various RDBMS platforms, including SQL Server and DB2. Generated Data Definition Language (DDL) statements to define and manage database structures. Tested the FSDs and ETL mapping which were developed to load from different source systems in the Teradata Staging /Target areas. Conducted comprehensive attribution analysis to determine the effectiveness of various marketing channels and campaigns, providing insights that optimized resource allocation and improved overall marketing strategy. Implemented and managed A/B testing for various marketing initiatives, including email campaigns, landing pages, and ad creatives, leading to data-driven decisions that enhanced user engagement and conversion rates. Developed automated scripts for data backup and recovery processes in both SQL and NoSQL environments, ensuring data availability and disaster recovery. Conducted internal and final DFD reviews with the ETL team EIS, Data Quality and Metadata Management team, data architects, and business and application data stewards. Involved in identifying the defects and developed a Defect Tracking report using Mercury Quality Centre this involved the Bug life cycle. Utilized advanced functions and formulas in MS Excel, including VLOOKUP, INDEX-MATCH, and PivotTables, to manipulate and analyze large datasets. Developed complex Excel models and macros to automate repetitive tasks and streamline data analysis processes. Created dynamic charts and graphs in MS Excel to visualize trends, patterns, and outliers in data effectively. Conducted data validation and cleansing in MS Excel to ensure data accuracy and integrity for reporting purposes. Prepared executive-level reports and presentations in MS Excel, summarizing findings and recommendations for key stakeholders. Environment: ER Studio, Teradata, Oracle, Informatica Power Center 8.6, Business Objects XI R3.1, Teradata SQL Assistant, Windows XP, Share Point, Altova XML SPY, XML, MS Access, MS Excel, SharePoint Portal. Education: Master s in Computer science from University of central missouri 2021 Keywords: continuous integration continuous deployment business intelligence sthree active directory information technology microsoft procedural language |