Roopesh - Data Engineer |
[email protected] |
Location: Miami, Florida, USA |
Relocation: |
Visa: |
Roopesh Reddy
Phno: 972-913-7067 Email: [email protected] Education: Masters in International Studies (MS) - Oklahoma State University 2009 Bachelor of Engineering (B.E) in Computer Science and Engineering - VTU India 2004 Journal Publications: Preprint: ABCMETAapp: R Shiny Application for Simulation-based Estimation of Mean and Standard Deviation for Meta-analysis via Approximate Bayesian Computation (ABC) https://bit.ly/2HfXbdr Abstract: Association between acute leukemias and levels of leptin in blood: A meta-analysis https://bit.ly/3cigs9z Qualification Summary: Over 16+ years of experience in information technology in various SDLC levels of Application development which include Analysis, Design, Coding, Testing, Debugging and Complete Support to the large scale applications. Involved in cloud computing platforms like Amazon Web Services (AWS), Azure and Google Cloud (GCP) and CUDA Extensive knowledge and experience in R programming and related technologies, Python and other Machine Learning tools. Worked on Relational and Non Relational databases in Business and Academic domain. Advanced working experience in GIS related applications(ArcGIS, PostGIS, Spatial Index) Strong programming skills in Multithreading and I/O management. Experience in Agile/Scrum Development Methodology and Test Driven Development Experience in Perl and Shell scripting Good knowledge and work experience on developing applications using Java Swing, AWT, RCP (Rich Client Platform), SWT, JFace and Applets. Experience in writing JUnit test cases for projects developed. Academic familiarity with Scripting Languages, DBMS, Project Management, Software Engineering, Web Development, Event Driven programming and API Experience in developing server side script using JSP, PHP, Perl, JavaScript, HTML Extensive knowledge of Java Servlets with database connectivity using JDBC, ODBC. Experience in HL7 data parsing with intra and inter connection data interface Effectiveness in coordinating with Business team and Quality Assurance team to fix bugs Involved in Integration testing, Regression testing, Unit testing, statistical testing and System testing for the Enhancements. Strong in Domain Understanding of Requirements Engineering and other phases of Software Engineering. Sound knowledge of Web Services, Flex and AJAX implementation in J2EE. Strong knowledge in the concepts of RDBMS and academic familiarity with the concepts of Data Mining and Bio-Informatics. Computer Expertise: Programming Languages : Java 9, J2EE, C++, C, SQL, Pascal, Basic, Java, Python, Bash AWS Services: S3, Redshift, EMR, EC2, RDS, SNS, KMS, Glue, Glue Catalog Technologies : Servlets, JSP, JDBC, Applets, XML Schema, DOM Open Source : BIRT 3.7 (Business Intelligence and Reporting Tool), Apache Scripting Languages : Python, JavaScript, UNIX Shell Script, VB Script, PHP, Perl, HTML Operating Systems : Windows 2003/XP/2000/NT/98/95, UNIX, Linux (Debian and Fedora) Database : PostgreSQL w/PostGIS, MS Access, MS SQL Server 2000, MySQL, Oracle 9i/8i, SQLite, Qlik Frameworks : Model View Controller Pattern (MVC), Yii 1.0 and Yii 2.0 Application & Web Servers : Apache Tomcat 9, IIS, ArcGIS Server Tools : Eclipse, MS Visio, NetBeans, Toad, Dreamweaver,JIRA, Bamboo, Confluence, log4j, Maven, FindBugs, SonarQube, QGIS, Ansible Source Code Repositories : Subversion, Github Design : UML. KML Data Processing: PySpark, Managed Airflow Version Control: Bitbucket Infrastructure as Code: Terraform Work Experience: Lead Data Engineer Transamerica Nov 2020-Present Designed and implemented a comprehensive data processing ecosystem on AWS utilizing various technologies. Duties: Developed and maintained Terraform infrastructure as code scripts for automated provisioning and management of AWS resources during my tenure at Transamerica. Applied GitOps practices for version control and automated deployment workflows while working for Transamerica. Designed and implemented ETL pipelines using PySpark at Transamerica, ensuring efficient data processing and transformation. Managed data pipelines in GitLab CI/CD for continuous integration and deployment of ETL workflows during my time at Transamerica. Optimized data workflows at Transamerica to ensure scalability, reliability, and performance. Demonstrated proficiency in debugging techniques to identify data patterns and meet business requirements while employed at Transamerica. Executed data migration projects at Transamerica, transferring on-premises data to AWS data lake buckets. Collaborated with stakeholders at Transamerica to enhance data accessibility, quality, and integrity across teams. Implemented security measures using AWS IAM roles and policies for access control and data protection at Transamerica. Handled importing of data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS. Designed and implemented automated data validation checks to ensure data accuracy and integrity across pipelines. Implemented logging and monitoring solutions using AWS CloudWatch for proactive issue identification and resolution. Enhanced data processing efficiency by implementing partitioning and clustering strategies in AWS Redshift. Automated data ingestion processes using AWS Lambda functions triggered by S3 events, improving data availability and timeliness. Developed custom data connectors and adapters to integrate external APIs with AWS services for seamless data flow. Implemented disaster recovery strategies for critical data systems, ensuring business continuity and data resilience. Conducted performance tuning and optimization of SQL queries and data processing workflows to meet SLAs and improve efficiency. Collaborated with data scientists to deploy machine learning models into production environments for real-time data processing. Sr. Data Engineer July 2018 - Oct 2020 University of Miami, SCCC(Sylvester cancer center) Worked across the entire Microsoft BI suite of products including SQL servers, SSIS, SSAS, SSRS and Power BI SDLC. Developing and maintaing custom Applications like MAS, PAS, SRS, CIS, CCSG and Publications. Setting-up of version control and creating AD-Hoc reports. Duties: Created and integrated custom visual based on business requirements using Power BI desktop. Developed custom calculated measures using Dax in Power BI to satisfy business needs. Utilized AWS Glue for Extract, Transform, and Load (ETL) processes, maintaining a centralized Glue Catalog for metadata management. Executed cost-saving analyses on AWS services (S3, Redshift, EMR) to optimize resource utilization. Integrated security measures using AWS KMS for encryption and managed sensitive information through secure storage practices. Integrating R programming language for statistical analysis in Power BI Involved in the development of various Ad-Hoc reports using SSRS. Maintaining various In-house projects that build on legacy systems. Involved in setting-up version control across all applications and created action plans for the entire application. Created interfaces to communicate between UM servers and PUBmed servers to import and process various publications of SCCC stakeholders. Environment: Power Bi, SSIS, SSAS, SSRS, Java, Javascript, Confusion, SQL, Node.js, Angular 2.0, R Programming Language, VB, SQL server studio, Apache Tomcat. Technical Lead University of Miami, Diabetes Research Institute(DRI) May 2012 - June 2017 Developing Full Life Cycle Projects Including Trial Based EMR, Billing System (CITP), Developed HUMALINK and GLUCOMETER servers, Diabetes educational software (Med EDUSOFT), Creating apps in SharePoint server, Small Animal Database System. Duties: Responsible for developing, customizing and maintaining clinical islet transplant protocol application for the DRI using FileMaker Building interfaces from HUMALINK and GLUCOMETER servers to the application, that access data using Open Database Connectivity Extensively involved in the process of re-engineering the billing module, which involved replicating data effectively from the vendor servers using Open-MRS Responsible for requirements gathering, design, construction, testing and deployment of Med EDUSOFT. Provided technical insight in the software development phase of the project. Leading the team in user acceptance testing. Extensively worked on Yii framework to fit the needs of the application. Reports developed for various applications (Using BIRT Reporting Tool). Involved in analyzing the user requirements and creating the Small animal database. Environment: Oracle Apps R12, Toad, Oracle DB Sr. Software Engineer/ SME 2010-2012 SHREE Consulting Ltd, India Developed Google Cloud based application for Telemed, Created Machine learning analysis on using Azure Machine learning. Duties: Provided Subject Matter expertise on health care applications including security and HIPPA. Created microservices on google cloud For authentication of the patient and doctor workflows. To analyze large volumes of pharmacy data sets using a machine learning model to optimize prescription services. Built and managed a reliable integration and deployment infrastructure for keeping the Google Cloud instances up to date with the latest codebase. Environment: Oracle DB, Node.js, MongoDB, Google Cloud, Docker Oracle Apps Functional Consultant 2008-2010 American Red Cross (Washington DC) Worked on implementing an enterprise financial and supply chain management system. This system, the Consolidated Financial System (CFS), is a web-enabled suite of applications that supports national and biomedical operations. Duties: Involved during business requirement gathering and scope analysis phase. Configured Set of books, Chart of accounts, Currency, Calendar, Periods, Period statuses and exchange rates. Involved in designing the reconciliation process in AP, AR and GL modules. Responsible for defined accounting setups in AR and AP modules Developed BR 100 for configurations in SysAdmin, GL, AP and PO modules. Defined Security Rules, Descriptive Flex fields, period types, Accounting Calendar, Currency. Worked on OuterBay/Outer Bay'' a third party tool for Purging Transaction in AP, AR, CM and GL. Environment: Java SDK 8, Apache Tomcat 9, Postgresql, ArcGIS Server, Ansible, MySQL, kml, Mirth, Oracle Trainee/Intern Group Leader Indian Space Research Organization(ISRO), Bangalore, India 2004 Worked as Group Leader for the project Quality Maintenance Support For System Duties: Was Responsible for developing a Software module that would track and notify any discrepancy in the aerospace rocket stage testing which is presently been used extensively in the servers of ISRO Maintained Web and Database servers and was involved in migrating Windows Applications to Linux Application Gained knowledge in future technology that can be implemented with Linux Environment: PHP, Mysql, Perl, Linux Servers Certifications: OCP(Oracle Certified Professional) CCNA(Cisco Certified network associate) MCSE(Microsoft Certified System Engineer) Amazon Web Services(Brainbench Certified) Microsoft Azure Architect Technologies(AZ-300) Currently enrolled in, obtaining certifications Google Associate Cloud Engineer Currently enrolled in, obtaining certifications Keywords: cprogramm cplusplus continuous integration continuous deployment javascript business intelligence sthree database active directory rlang purchase order microsoft Arizona Arkansas |