Utthman Khan - Data Analyst |
[email protected] |
Location: Dallas, Texas, USA |
Relocation: Hybrid or Remote or Onsite in Dallas, TX |
Visa: USC |
Uthmaan Khan
[email protected] Senior Data Analyst Phone: (817) 482-3535 Dallas, TX US Citizen Summary: Business Data Analyst with over 7 years of diversified experience including strong Business Data Analysis, Data Management and Data Governance in requirements gathering and documentation with solid oral and written communication skills seeking a business data analyst position which will utilize and incorporate my current skills and experience, as well as expand them to assist the organization in being successful. Skilled in analyzing business processes, eliciting requirements, defining business needs and translating it to Business Requirement Document (BRD), Functional Requirement Document(FRD), Software Requirement Specification (SRS)/ Technical Requirement Documents (TRD), User Stories, Use-case Specification and Non-functional Requirements. Great experience in interacting with Stakeholders/Customers, gathering requirements through interviews, workshops, and existing system documentation or procedures, defining business processes, identifying and analyzing risks using appropriate templates and analysis tools. Experienced in various phases of Software Development life cycle (Analysis, Requirements gathering, Designing) with expertise in documenting various requirement specifications, functional specifications, Test Plans, Source to Target mappings. Experienced in using Redshift as a data source for business intelligence tools like Tableau, Power BI, or Looker for reporting and visualization. Capable of designing data warehouses and data marts using dimensional modeling techniques, such as star and snowflake schemas. Integrated multiple data sources into Looker, including cloud databases (BigQuery, Redshift), SQL databases, and flat files. Proficient in using Databricks for collaborative, interactive data analysis, and processing using SQL, Python, R, or Scala. Experienced in using CRM systems to track tasks, appointments, follow-ups, and other activities related to customer interactions. Proficient in understanding and analyzing different data schemas, structures, and formats, including relational databases, XML, JSON, CSV, etc. Proficient in Python, with a strong grasp of the language's syntax, data structures, and libraries. Worked on Data Management using Collibra for Enterprise Data Governance projects on areas of Data Quality, Reference data management, Data Dictionary, Business Glossary, Data Discovery, Metadata management. Worked in projects involving Data Mapping, Data Profiling, Data Migration, Data Conversion, Data Quality, Data Governance, Data Lineage, Data Integration. Worked in accordance with the Agile Safe /Scrum methodology for various phases of the business and created business vision, business architecture documents and mock reports. Experienced in extracting data from various sources, such as databases, APIs, and files, using ETL processes, data integration tools, and scripting languages like SQL Proficient in using SAS for statistical analysis, including descriptive statistics, regression analysis, hypothesis testing, and time series analysis. Designed and implemented data marts to support specific business functions, ensuring efficient data storage and retrieval. Led the implementation of MDM solutions to ensure data consistency, accuracy, and reliability across the organization. Worked excel Pivot tables and V-Lookup tables for Data Analysis. Proficient in data analysis using SQL, Tableau, and other data analysis techniques. Used SQL to extract data from Oracle ERP databases to analyze the data. Strong understanding of project life cycle and SDLC methodologies including RUP, RAD, Waterfall and Agile. Expertise in Master Data Management, Meta Data, Informatica Business Glossary & Data Quality Involved in analyzing the data using MS Excel, SQL and Access and generating different reports to gain more knowledge about the business status. Strong experience in using Excel and MS Access to dump the data and analyze based on business needs. Involved in defining the source to target data mappings, business rules and data definitions. Worked on the implementation of Metadata Repository, Maintaining Data Quality, Data Clean-up procedures, Transformations, Data Standards, Data governance program, Scripts, Stored Procedures, triggers and execution of test plans. Proficient in writing complex SQL queries using Snowflake SQL to extract, manipulate, and analyze data stored in Snowflake data warehouses. Utilized SSIS to design and implement data integration workflows, seamlessly moving and transforming data from diverse sources for analysis. Developed robust ETL processes using SSIS, cleansing and enriching data to ensure accuracy and consistency for analytical insights. Industrious, results oriented professional with technical and business experience dedicated to organizational growth and profitability through excellent project management, business analysis and leadership capabilities. Implemented and Administered Collibra Data Governance: Setup Business Glossary, Workflows, Data Governance Assets, Setup users and Data Catalogue. Ingested business and technical metadata into Collibra via excel import template based on Source, Region and Assets type. Ensure Data Governance function standards - Metadata, Data Lineage, Data Mappings, Data Ingestion into the EDL appropriately, etc to support business and technology needs to locate data, understand data and ensure appropriate use. Serve as an expert in Business Metadata in Collibra and promote understanding of the data definition at the application, data element and data domain level. Education: Bachelor s in Information System Tarleton State University, Stephenville, Texas. Skill Matrix: Data Analysis and Modeling: Power Designer, SQL, R, Informatica Project Management: MS Project, MS Excel, Confidential Rational Portfolio Manager Database Management: Oracle SQL, Python, MySQL, MS SQL Server, MS Access, Data Lineage, Hadoop (Hive, PIG), Teradata Data Visualization: Tableau, Power BI, Excel, Macros Cloud Platform AWS, Google Cloud, Azure Version Control and Documentation: MS Office Word, MS Project, Advanced Excel, Lotus Notes, GitHub repository Business Intelligence: Data Warehousing, RDBMS, Hypothesis Testing, A/B Testing, Data Mining Data Visualization and Reporting: Tableau, Power BI, Excel, Macros Collaboration Tools: JIRA, Confluence Testing and Quality Assurance: User Acceptance Testing (UAT), Test Case Design, Test Planning Professional Work Experience: GM Financial Remote Sept 2023 - Present Sr Data Analyst/Business Analyst Worked on Data Analysis, Business Analysis, Design, Development, Implementation and Testing of Data Warehousing using Data Conversions, Data Extraction, Data Transformation and Data Loading (ETL). Created and maintained Data Governance Policies, Business Glossary, Data Dictionary, Reference Data, Metadata, Data Lineage, and Data Quality Rules. Developed and maintained interactive data visualizations and dashboards using Tableau, enabling business stakeholders to easily understand and analyze complex data. Worked on interactive dashboard and Reports in Tableau for monitoring the operation performance on a day-to-day basis. Designed and developed interactive and visually appealing dashboards in Looker to support data-driven decision-making Defined the data model and schema for data marts, aligning with business requirements and ensuring data integrity. Experience in leveraging Databricks for data engineering tasks, such as Extract, Transform, Load (ETL) processes, data cleansing, and transformation. Conducted MDM maturity assessments and identified areas for improvement to enhance data management capabilities. Proficient in integrating CRM systems with other tools or databases and customizing the CRM to align with specific business needs. Proficient in Python, with a strong understanding of its core concepts, data structures, and syntax. Developed and implemented risk mitigation strategies, collaborating with cross-functional teams to ensure effective execution and risk reduction. Experienced in writing and optimizing SQL queries to analyze large-scale data sets in various formats like CSV, JSON, Parquet, or ORC stored in S3. Skilled in integrating with APIs using Python to retrieve and exchange data from various sources. Understanding of evaluation metrics for assessing the performance of data mining models, such as accuracy, precision, recall, F1 score, ROC curve, and AUC-ROC. Experienced in providing training to team members on CRM usage, features, and best practices. Experience in data manipulation, cleaning, and transformation using SAS tools like DATA step, PROC SQL, and SAS Macros for efficient data processing. Demonstrated ability to solve complex business problems by leveraging Python's capabilities. Proficient in implementing data governance and stewardship processes using Collibra to ensure data quality, compliance, and accountability. Proficient in using Alteryx to prepare and blend data from multiple sources, including databases, spreadsheets, and cloud-based platforms, to create clean, standardized datasets for analysis. Worked with excel in creating PivotTables and PivotCharts to summarize and visualize large datasets, enabling efficient data analysis and reporting. Applied dimensional modeling techniques (star schema, snowflake schema) to structure data marts for optimal query performance. Skilled in working within the collaborative workspace provided by Databricks, facilitating team collaboration and sharing of notebooks and data. collaborated with cross-functional teams, including developers, data scientists, and business stakeholders, to develop and deploy Python-based solutions. Collaborated with stakeholders to define data governance policies and standards for master data management. Worked on information security, including the implementation and management of security controls, policies, and procedures to protect sensitive data and ensure compliance with industry regulations. Worked with the data steward and data owners in creating metadata, Lineage, Data Quality rules and guidelines. Experienced in defining and implementing data governance policies, standards, and guidelines within Collibra. Experience in designing workflows in Alteryx by arranging and configuring tools to automate data processes, transformations, and analytics. Experience in optimizing SQL queries, indexing strategies, and query performance tuning techniques to enhance query execution speed in Teradata. Skilled in using Python libraries such as NumPy, pandas, and Matplotlib for data analysis and visualization. Capable of performing data cleansing, deduplication, validation, and enrichment using Alteryx's functionalities and tools. Experience using Python for web development, including frameworks like Django and Flask to create web applications. Proficient in using MDM tools such as Informatica MDM, IBM InfoSphere, SAP Master Data Governance, and Microsoft Master Data Services. Experience in evaluating and validating data mining models using techniques such as cross-validation, holdout validation, and performance metrics to ensure robustness and generalization. Skilled in transforming XML data using technologies like XSLT (Extensible Stylesheet Language Transformations) for presentation and data conversion. Proficient in Oracle Database installation, configuration, and administration tasks, including Oracle RDBMS and Oracle Exadata Established risk monitoring processes and prepared regular risk reports for senior management, highlighting key risks and mitigation status. Experience in using specific replication tools or platforms like Oracle GoldenGate, SAP LT Replication Server, AWS Database Migration Service (DMS), SQL Server Replication, etc. Skilled in SAS programming language, creating data sets, performing transformations, generating reports, and automating tasks using SAS procedures. Adept at writing Data Mapping Documents, Data Transformation Rules and maintaining Data Dictionary and Interface requirements documents. Experienced in integrating UiPath with different applications and systems via APIs or connectors for seamless automation. Capable of integrating XML data with various applications and systems, including databases and web services. Collaborated with business stakeholders to gather requirements and translate them into actionable Looker dashboards and reports. Provided ongoing maintenance and support for data marts, including data refreshes and schema updates. Skilled in performing data cleansing and transformation tasks using Alteryx's intuitive drag-and-drop interface, including removing duplicates, handling missing values, and formatting data for consistency. Established end-to-end data lineage processes to track the flow of data across various systems, applications, and processes. Familiarity with using Excel add-ins for specialized functionalities, such as statistical analysis, financial modeling, or industry-specific tools. Developed and implemented efficient record management policies and procedures to ensure compliance with legal and regulatory requirements. Proficient in using Excel's built-in data analysis tools, including regression analysis, data tables, and goal seek for more in-depth analysis. Conducted data lineage analysis to identify dependencies, validate data integrity, and ensure compliance with regulatory requirements. Designed and developed SSRS reports and dashboards to visualize data and provide insights to stakeholders. Hands-on experience in designing and implementing security frameworks and controls, such as access controls, authentication mechanisms, encryption protocols, and network security solutions. City National Bank Feb 2020 Sept 2023 Dallas, TX Data Analyst / Business Data Analyst Worked on projects involving Data Analysis, Data Management , SQL, problem-solving, and multi-tasking between multiple projects. Wrote complex SQL queries on daily basis to retrieve data and worked on Teradata, Oracle, MS Access, SQL Server etc. Involved in documenting the Data quality scores in Collibra using integrations with Informatica Data Quality. Performed data analysis and data profiling on a day-to-day basis using complex SQL on various source systems including Oracle and Teradata. Tested Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables. Wrote the business requirements in Jira Agile Epics and user stories format and uploaded and assigned the user stories in Jira for both Agile (sprints) methodology. Proficient in using data modeling tools like ERwin, ER/Studio, IBM Data Architect, or Lucidchart to design data models. Proficiency in Python for data science tasks, such as data cleaning, feature engineering, and building predictive models. Experience in creating visualizations and dashboards using Databricks notebooks to present and communicate insights derived from data analysis. Implemented and configured MDM software to support master data management initiatives. Proficient in using XPath and XQuery for querying XML data and selecting specific elements or attributes. Ensured data quality and consistency within data marts through validation and cleansing processes. Provided training and support to end-users, enabling them to leverage Looker for their analytical needs. Experienced in web development using Python, including web frameworks like Django and Flask to build web applications and APIs. Experience in integrating SSRS reports with other Microsoft technologies like Power BI, SharePoint, or third-party applications. Capable of creating Python scripts to automate repetitive tasks and enhance workflow efficiency. Experience in integrating Snowflake with other tools, databases, or ETL processes, ensuring seamless data transfer and synchronization. Experience in using SAS BI tools for creating dashboards, data visualization, and generating insights through reports and analytics. Experience in using Python for data science tasks such as statistical analysis, predictive modeling, data mining, and feature engineering, leveraging libraries like SciPy, Scikit-learn, StatsModels, and Pandas. Worked closely with business stakeholders to gather requirements and translate them into data mart design specifications. Proficient in optimizing Spark jobs, tuning configurations, and improving performance for efficient data processing within Databricks. Experience working with databases using Python, including database management systems like MySQL, PostgreSQL, and MongoDB. Implemented and maintained risk management frameworks, developing risk policies and procedures to standardize practices across the organization. Utilized Looker s API and SDK for advanced integrations and custom functionality. Prepared test Data sets and performed data analysis and used MS excel for data mining, data cleansing, data mapping, data dictionary and data analysis. Skilled at integrating with and consuming APIs using Python to gather and exchange data. Worked on Data Management using Collibra for Enterprise Data Governance projects on areas of Data Quality, Reference data management, Data Dictionary, Business Glossary, Data Discovery, Metadata management., Managed all data collection, extraction, transformation and load (ETL) activities using Microsoft SSIS, Talend and informatica including data profiling, data cleansing, data conversion and quality control Experienced in using data modeling languages like SQL Data Definition Language (DDL) to create and modify database objects. Implemented JSON schema validation using ajv and Joi to ensure data integrity and compliance with defined standards. Skilled in serializing structured data into XML format for storage or transmission. Experience in monitoring query execution, analyzing query execution plans, and using Snowflake's query history for performance analysis. Reviewed the data model and reporting requirements for Cognos Reports with the Data warehouse/ETL and Reporting team. Serialized and deserialized JSON data for network communication, optimizing data transfer performance. Experienced in working with XML-based messaging for SOAP (Simple Object Access Protocol) and REST (Representational State Transfer) web services. Proficient in creating and utilizing SAS Macros for automating repetitive tasks, improving code efficiency, and facilitating reusability of code segments. Proficient in visualizing manufacturing data using tools like Tableau, Power BI, or custom dashboards to communicate insights and KPIs effectively. Prepared and transformed data for analysis in Looker using ETL tools and Looker s data preparation features. Performed detailed analysis of complex business processes, data requirements and proposes solutions and worked directly with internal customers to understand the business environment and needs. Involved in several complex financial reports related to Non-performing Loans portfolio, which includes Credit Loss Reporting Package and Annual Reports Package. Created (Extract, Transform, Load) ETL design mapping sheets, data reconciliation strategy, data archival strategy- ETL framework, stored procedures and built SQL query objects to detect data loss. Involved in writing Data Mapping Documents for the system and involved in documenting the ETL process and writing SQL Queries for the retrieval of the data including MS SQL Server and Teradata. Created Interactive Tableau Dashboard after gathering and analyzing the data from the warehouse to illustrate the metrics of the business process. Responsible from the user perspective to analyze dashboards and analytical reporting requirements using Tableau. Worked with the business users to understand & document the design and functionality requirements. Involved in testing and documentation for applications developed in SDLC Environment. Performed BI and ETL Analyst role interfacing with subject matter experts to understand and define the functional requirements supporting reporting applications by outlining data strategy and report specifications. Migrated data from different data sources (SQL Server, Flat files, Excel source files, Oracle) and loaded the data into target tables, data warehouse by performing different kinds of transformations such as conditional split, derived columns, aggregation etc., to ensure data validation, before loading based on the business need using SSIS. Conducted financial risk analysis, utilizing models and simulations to evaluate credit risk, market risk, and liquidity risk. Used Data Integrator to load data from various sources such as MS Access, MS Excel, CSV files to MS SQL Server. Extracted data from SQL Server database tables and Oracle database tables into flat data files and Excel sheet Used SQL Loader and SQL Developer to load data from flat data files and excel sheet into Oracle database. Designed and implemented data profiling and data quality improvement solution to analyze, match, cleanse, and consolidate data before loading into data warehouse. Tenet Healthcare Nov 2017 Jan 2020 Dallas, TX Business Data Analyst Created data dictionary, Data mapping for ETL and application support, DFD, ERD, mapping documents, metadata, DDL and DML as required. Create and maintain product backlog items, bugs, features, etc. in JIRA and TFS / Azure DevOps for story tracking and documentation. Worked on Snowflake cloud database and used tools like Alteryx and Tableau to present study, analyze and present data Skilled in using XML parsing libraries and APIs, such as DOM (Document Object Model), SAX (Simple API for XML), and StAX (Streaming API for XML). Prepared user stories documentation in JIRA for Agile projects, documented requirements for the development and data migration epic to convert customer data to oracle instance with upgrades. Conducted and automated the ETL operations to Extract data from multiple data sources, transform inconsistent and missing data to consistent and reliable data and finally load it into the Multi-dimensional data warehouse. Lead the go-live implementation plan from the business side for digital banking initiatives and partner with other management leads in IT to align organizational readiness plans with delivery plans. Documenting the Data quality scores in Collibra using integrations with Informatica Data Quality. Performed data analysis and data profiling using complex SQL on various source systems including Oracle and Teradata. Proficient in extracting insights from CRM data to inform business strategies, customer retention efforts, and marketing campaigns. Developed cross-platform applications using Python, leveraging tools like PyQt and Kivy. Experience in setting up automated workflows and scheduling jobs within Databricks for regular data processing or ETL tasks. Provided training and support to data stewards and other stakeholders on MDM best practices and tools. Skilled in creating comprehensive documentation for data models, including data dictionaries, naming conventions, and data model metadata. Prepared business glossary and metadata with Confidential for all deposit related CDE's. Used advanced Excel to work on large datasets by developing formulas, creating Pivot tables, V-lookup and H-lookup formula development, regression analysis. Developed and maintained interactive data visualizations and dashboards using Tableau, enabling business stakeholders to easily understand and analyze complex data. Working experience on interactive dashboard and Reports in Tableau for monitoring the operation performance on a day-to-day basis. Experienced in using CRM systems to track tasks, appointments, follow-ups, and other activities related to customer interactions. Knowledgeable about XML databases and NoSQL databases designed to store and query XML data, such as MarkLogic or eXist-db. Debugged and resolved issues related to JSON data parsing and formatting, ensuring smooth application functionality. Developed and executed MDM strategies and roadmaps aligned with business objectives and data management goals. Proficient in writing complex SQL queries and utilizing reporting tools like Tableau, Power BI, or Looker to extract insights and generate reports from the data warehouse. Experienced in connecting to and manipulating databases using Python, including SQL databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB). Proficient in data modeling for NoSQL databases like MongoDB, Cassandra, or Neo4j, considering their unique structures and querying requirements. Experience in creating Data Governance Policies, Business Glossary, Data Dictionary, Reference Data, Metadata, Data Lineage, and Data Quality Rules. Use SQL to query into databases and analyze data. Worked on Snowflake cloud database for Data Analysis and used ETL tools like Alteryx and Visualization tools like Tableau to Clean, Prep , analyze and present data. Extract, transform, and load data from source systems to Azure Storage services using a combination of Azure Data Factory, T-SQL, Spark SQL, and U-SQL of Azure Data Lake Analytics Created SQL queries for Extracting, Transforming, and Loading (ETL) mapping using PostgreSQL, Oracle, and SQL Server. Proficient in testing Python code and debugging techniques to ensure high-quality software. Experienced in using Redshift as a data source for business intelligence tools like Tableau, Power BI, or Looker for reporting and visualization. Worked with Finance, Risk, and Investment Accounting teams to create Data Governance glossary, Data Governance framework and Process flow diagrams. Proficiency in multiple databases like MongoDB, Cassandra, MySQL, ORACLE, and MS SQL Server Designed logical model with all relationships for database; forward engineered it to SQL Server with Erwin. SQL Server, Visual Studios, PowerBi, Tableau, and SharePoint. Created data stories, reports and visualizations using Tableau and Power BI Managed full SDLC processes involving requirements management, workflow analysis, source data analysis, data mapping, metadata management, data quality, testing strategy and maintenance of the model. Assisted in the business process design or redesign and documentation as needed. Created business scenarios and insights that are meaningful and which impact on critical customer facing issues and help navigate process improvements/decision Extracting data through SQL query and business intelligence solutions, analyzing and interpreting Designed and implemented ETL processes based in SQL, T-SQL, stored procedures, triggers, views, tables, user defined functions and security using SQL SERVER 2012, SQL Server Integration Service (SSIS). Keywords: business intelligence sthree database rlang information technology golang microsoft Texas |