Seshuram - data engineer (python) |
[email protected] |
Location: Durham, North Carolina, USA |
Relocation: Onisite/remote |
Visa: H1B |
Seshuram Sundaramoorthy
PH : 919-655-5636 HIGHLIGHTS Experienced Data Analyst with a strong background in data warehousing, data modeling, and SQL. Proficient in Python, PySpark, Hadoop, Impala, and Hive. Seeking to leverage my expertise to contribute to a dynamic team, extend data warehouse capabilities, and enhance data-driven decision-making processes. Developed and maintained scalable and efficient Python applications, demonstrating strong analytical and problem-solving skills Skilled SQL Data Engineer with a strong background in SQL, data manipulation, and data reconciliation. Proficient in Excel, Python, and R, with the ability to work in a flexible data environment. Advanced expertise in statistical modelling for predictive and prescriptive analytics using SQL, NOSQL, Python, R and Scala in Spark framework. Expert level knowledge in packaging and deploying Scala Jars for SAAS applications. Experience in architecting and developing data pipelines in Microsoft Azure, AWS, Docker Jenkins, Databricks and Hadoop. Experienced SSRS for developing interactive and customized reports and dashboards. Proficient in developing, testing, and maintaining Hadoop ecosystem components. Advanced knowledge in designing and implementing Machine Learning and Deep Learning algorithms via Tensorflow, Keras, Sci-Kit Learn, SciPy and numPy. Strong coding and debugging skills in C/C++, Java, Python. EXPERIENCE Data Analyst (Engineer), PricewaterHouseCoopers LLP, Jul 2022 Present Designed and developed the backend database of multiple PowerApps applications used by partners and stakeholders of PWC to monitor their assets using Microsoft SQL server and stored procedures. Designed and monitored data pipelines using Azure Data Factory to port legacy data from older database to the SQL server used by the PowerApps applications. Gained hands-on experience in SQL query optimization. Data Analyst (Engineer), Cisco Systems, Inc. Oct 2019 Jun 2022 Performed ETL operations on multiple datasets related to reverse supply chain product data and analysed the consolidated dataset to perform feature extraction. Proficient in SSIS for designing and executing ETL processes to extract, transform, and load data. Visualized the consolidated dataset using python s Matplotlib to analyse trends in the products lifecycle. Developed Python scripts and applications to transform, clean, and process data for analysis. Optimized Hadoop cluster performance and ensured data reliability and availability. Developed and tuned SQL queries for populating data warehouse summary tables/views. Built a predictive model to classify customer issues into being critical or not for every return material authorization order placed. Assisted in analyzing reporting requirements and data model needs Data Analyst (Engineer), Zabble Inc. July 2019 - Sep 2019 Developed and maintained Zabble s IOS application by working on algorithms for real-time waste analytics, micro- services and their integration with back-end services in AWS coded in Python. Designed and developed data pipeline using AWS Athena, Lambda, DynamoDB, S3, Textract and Quicksight, tested for performance quality and responsiveness. Followed agile principles in developing the software. Data Scientist, Coca-Cola May 2018 - Dec 2018 Sales and Volume Prediction Developed an E2E data science pipeline using PySpark framework and packages like mllib in Azure Databricks environment to predict volume and revenue. Skills: SQL - Advanced Excel skills (Pivot tables, formulas, VLOOKUP) - Data manipulation and transformation - Data reconciliation - Python and R for data transformation - Data analysis -SQL Server -SSIS -SSRS -.NET - Data Analysis: Proficient in data analysis techniques and tools. - Data Warehousing: Over 5 years of experience in data warehousing. - Data Modeling: Strong background in data modeling. - Programming Languages: Python, PySpark, and PLSQL. - Big Data Technologies: Hadoop, Impala, Hive. - Databases: Knowledge of Snowflake. - Other Tools: Experience with Databricks. - SQL Skills: Excellent SQL skills. - Operating Systems: Familiar with Unix/Linux environments. - Communication: Strong written and verbal communication skills. - Problem Solving: Demonstrated ability to solve complex problems. - Teamwork: Capable of working effectively in a team. - Self-Starter: Proven track record as a self-motivated individual. Academic Projects Information Extraction Using NLP Techniques Oct 2018 Dec 2018 Developed an information extraction system for a given template using NLP techniques such as N-grams, Named Entity Recognition, tokenizing, semantic parsing etc. Image Classification using CNN Jan 2019 - May 2019 Developed a convolution neural network model for classifying images from Yelp s dataset challenge as to whether an image in the dataset is a food or a drink. Model was built using Scala and Spark, bundled into a Scala Jar and uploaded into Amazon EMR service as a Spark application. Final model yielded an accuracy of 85%. Keywords: cprogramm cplusplus sthree rlang |