Mohan Krishna - Senior GCP Data Engineer |
naresh@vorizoit.com |
Location: Jersey City, New Jersey, USA |
Relocation: Open |
Visa: H1 |
Mohan Krishna
________________________________________ Professional Summary 9+ years of experience in Data Engineering, Data Pipeline Design, Development, and Implementation. Expertise in migrating on-premise ETLs to Google Cloud Platform (GCP) using BigQuery, Cloud Data Proc, Cloud Storage, and Composer. Skilled in Kafka, PySpark, Spark SQL, Data Science Integration, and Machine Learning Model Deployment. Strong knowledge of AWS (Redshift, Athena, EC2, EMR, S3, Glue, Lambda) and Azure (Data Lake, Data Factory, Synapse, Data Bricks). Experienced in Hadoop ecosystem (HDFS, YARN, Hive, Sqoop, Oozie, HBase, Spark, Kafka, Flume, Zookeeper). Proficient in ETL, Data Visualization (Tableau, Google Data Studio, Power BI), and Kubernetes deployments. ________________________________________ Technical Skills Programming: Python, Java, Scala, SQL, PL/SQL, Shell Scripting Databases: MySQL, PostgreSQL, Oracle, Cassandra, MongoDB, Teradata Cloud Platforms: GCP, AWS, Azure Big Data Tools: Spark, Hive, Kafka, Sqoop, Impala, Flume ETL & Visualization: Informatica, Talend, Power BI, Tableau, Databricks Scheduling & Build Tools: Jenkins, Maven, Oozie, Airflow, Zookeeper Operating Systems: Linux, Windows, Unix ________________________________________ Professional Experience Senior GCP Data Engineer UPS (Jan 2023 Present) Developed ELT pipelines using Data Proc (PySpark), BigQuery, and Cloud Composer. Migrated Oracle SQL ETLs to GCP, leveraging Cloud Pub/Sub, Data Proc, and BigQuery. Implemented Airflow pipelines using GCP Composer for workflow orchestration. Worked on Kubernetes deployments and Stackdriver log monitoring. Designed data pipelines for BigQuery to Azure Data Warehouse migration. Assisted Data Science team in deploying machine learning models using Cloud Dataflow. Senior GCP Big Data Engineer Alpha Technologies, India, Client: JP Morgan Chase (Jan 2021 May 2022) Architected multiple data pipelines and developed ETL workflows in GCP. Used Google Cloud Functions to load GCS files into BigQuery. Migrated AWS EMR Spark workflows to GCP Cloud Dataflow. Deployed multi-node Kubernetes clusters and automated Datadog monitoring. Built Airflow DAGs for automation of ETL pipelines. Senior AWS Data Engineer Avon, India. Client: Anthem Health Care (Nov 2019 Dec 2020) Developed Spark applications using Python and Apache Spark for large-scale data processing. Implemented real-time streaming with Kafka and Spark Streaming. Migrated on-premises applications to AWS (S3, EMR, Redshift, Glue). Utilized AWS Glue & PySpark to load and transform data in S3 & Hive. Azure Data Engineer Ace Hardware Corporate, India, Client: Qualcomm (Nov 2018 Nov 2019) Led migrated legacy databases to Azure SQL, Data Lake & Synapse Analytics. Designed Azure Data Factory (ADF) pipelines for ETL workflows. Developed Azure Data bricks transformations using PySpark. Implemented performance tuning in Spark clusters. Data Analyst Couth Infotech (May 2016 Oct 2018) Designed Hive tables, developed Sqoop jobs for Oracle to HDFS migration. Created Oozie workflows for automating ETL processes. Developed Spark SQL transformations for structured and semi-structured datasets. Implemented partitioning and bucketing in Hive as per the specific project requirements. Engaged in the conversion of Hive SQL queries into Spark transformations using Spark SQL and Scala. ________________________________________ Education Master of Science in Information Studies Trine University (2023) Bachelor of Engineering in Computer Science JNTUH (2016) ________________________________________ Certifications: Google Cloud Certified - Professional Data Engineer AWS Certified Data Analytics - Specialty ________________________________________ Keywords: business intelligence sthree procedural language |