Home

Shivani - Python Developer
[email protected]
Location: Hicksville, New York, USA
Relocation: any where USA
Visa: GC
Title: Sr. Python Developer
Name: Shivani P |
Cont: 980-304-8258 |
Email Id:[email protected]
PROFESSIONAL SUMMARY:
Around 9 years of experience as a python developer, software development and design using Python, Django, XML, HTML, DHTML, C#,Oracle PL/SQL, Postgres.
Experienced in installing, configuring, modifying, testing, and deploying applications with AWS (Amazon Web services).
Experience in all phases of Software Development Life Cycle (SDLC)-Waterfall, agile Process across various workflows (Requirement study, Analysis, Design, Coding, Testing, Deployment and Maintenance) in Web & Client/Server application development.
Extensive experience in developing web applications using Python, Django, Flask frameworks.
Experience in working with several python libraries including Beautiful soup, NumPy, matplotlib, SciPy, PyQt, SQLAlchemy.
Hands-on experience with industry-standard IDEs like PyCharm, Sublime, NetBeans.
Extensive experience in developing applications by implementation of web based and client/server application using HTML, Angular JS, Node JS, CSS, JavaScript, and various Java/J2EE technologies.
Good experience in Shell Scripting, Oracle RDBMS, SQL Server, UNIX, and Linux.
Good knowledge of TCP/IP, UDP, HTTP, HTTPS, SSH and SSL protocols.
Expert at version control systems like Git, SVN and CVS, Migrated repos from svn to GitHub.
Proficient in writing SQL Queries, Stored procedures, functions, packages, tables, views, triggers using relational databases like Oracle, MYSQL Server.
Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node,
Experience with source code testing frameworks like JUnit, PyUnit.
Experience object-oriented programming (OOP) concepts using Python.
Experienced in business domains like Investment Banking (Equity Research), Retail Loan Financing, Public Pension Funds, and Insurance products .
Used Scala sbt to develop Scala coded spark projects and executed using spark-submit
Data Node, MapReduce concepts responsible for writing MapReduce programs and setting up standards and processes for Hadoop - based application design and implementation.
Excellent knowledge with Python engineering under Linux OS (Debian, Ubuntu, SUSE Linux, RedHat Linux, Fedora). Experienced in creating web-based applications using Python, DJANGO, GTK, QT, C++, XML, CSS, HTML, DHTML, JavaScript, and jQuery.
Implemented pre-defined operators in spark such as map, flat Map, filter, reduceByKey, groupByKey, aggregate ByKey and combine ByKey etc.
Experience processing Avro data files using Avro tools and MapReduce programs.
Worked mainly on Core Java, C++, STL, data structures, UNIX, multithreading
Involved in development using C++, bug fixing and unit testing of the Layout commands.
Well versed with design and development of presentation layer for web applications using technologies like HTML, CSS, jQuery and JavaScript.
Good experience in working with Amazon Web Services (AWS) like AWS EC2, S3, VPC, SES, ELB, EBS, RDS, Glacier, DynamoDB etc.
Expertise with different tools in Hadoop Environment including Pig, Hive, HDFS, MapReduce, Sqoop, Spark, Kafka, Yarn,
Good knowledge on C, C++ languages including threads, templates, Polymorphism and OO concepts.
Experience in using XML, SOAP, REST web Service for inter operable software applications.
Hands on experience with bug tracking tools JIRA and Bugzilla.
Experience in Agile development processes ensuring rapid and high-quality c delivery
Highly motivated, quality minded developer, with proven ability to deliver applications against tight deadlines.

EDUCATION: Bachelor s in IT Services, 2015



TECHNICAL SKILLS:
Programming Languages Python
Cloud Services AWS, Microsoft Azure, GCP
Defect Tracking JIRA, Bugzilla, and VersionOne
Web Server Apache Webserver, IIS
Integration Tools Jenkins, IBM integration and Web Builder
Tools Django, Spyder, Visual Studio, Tableau Analytics Wynsure
Databases Oracle 10/11g, MySQL, SQL Server, and PostgreSQL, MongoDB
Operating Systems Unix, Linux, Ubuntu, Solaris, Windows and MacOS

PROFESSIONAL EXPERIENCE:

Client: Crowdstrike, Sunnyvale, CA Jan 2022 Present
Sr. Python Developer
Responsibilities:
Worked with Terraform for automating VPCs, ELBs, security groups, SQS queues, S3 buckets, and continuing to replace the rest of our infrastructure.
Hands-on configuring applications on AWS EC2 instances and the stage on S3 buckets. Perform S3 buckets creation, policies and IAM role-based policies and configuring user access levels.
Utilized Python Libraries like Boto3, NumPy for AWS.
Hands-on experience using IAM for creating roles, users, groups, and MFA to provide additional security to AWS account and its resources.
Familiar with building multi-threaded applications, and understanding distributed systems like Spark and Storm as well as
Worked on CSV files while trying to get input from the MySQL database.
Using raw files loaded data files from URL s to Amazon S3 Bucket.
Developed AWS Lambda functions in Python using S3 and SQS triggers to automate workflows and developed Python scripts to store and retrieve objects in AWS S3 buckets. Developed scripts in Python using boto3 API s to retrieve messages from various events.
Implemented Spark using Scala and utilizing Data frames and Spark SQL API for faster processing of data.
For Paccar customer application we have created near real time Api s and integrated with salesforce.
Implemented single sign on for the Paccar customer apps. Enabled customers to seamlessly access all the customer app s with single identity.
Developed high-performance distributed queueing system. Scala, Redis, Akka, closure, MQ messaging, Json Developed new functionality for interactive healthcare system.
Worked on fixing theBuild issuesduring the deployment procedures to all lower lanes and higher environments. O Implemented PySpark using Python and utilizing data frames and temporary table SQL for faster processing of data.
Implemented Spark using Scala and Spark SQL for faster testing and processing of data.
Worked wif Amazon Web Services (AWS) using EC2 for hosting and Elastic map reduce (EMR) for data processing wif 53 as storage mechanism.
Used Kubernetes to deploy scale, load balance, and worked on Docker Engine, Docker HUB, Docker Images, Docker Compose for handling images for installations and domain configurations.
Worked on reading queues in Amazon SQS, which have paths to files in Amazon S3 Bucket.
Aggregate clean files in Amazon S3.
Build IBM DataStage and Mainframe jobs to extract data from various file systems, Databases like DB2, MS SQL & Oracle that contain Consumer Banking, Investment Banking, Mortgage and Credit card data, perform transform operations as per business needs and load into the data warehouse.
Used Amazon EMR for map reduction jobs and test locally using Jenkins.
Also worked on Amazon EC2 Clusters to deploy files into Buckets.
Developed and maintained the Verification and Validation group's project schedules and roadmap ensuring project on schedule.
Experienced in Agile Methodologies and SCRUM Process.
Implemented PySpark using Python and utilizing data frames and temporary table SQL for faster processing of data. O Experience working on yaml serialization.
Automated setting up server infrastructure for the DevOps services, using Ansible, shell and python scripts.
Exported/Imported data between various data sources.
Designing and implementing CI (Continuous Integration) system: configuring Jenkins Servers, Jenkins nodes, TFS creating required scripts (Perl & Python), and creating/configuring VMs (Windows/Linux).
Used Azure Container Services with Kubernetes for Orchestrating the servers.
Created database using MySQL, wrote several queries to extract/store data.

Environment: AWS, Glue, EMR, GIS, spark- SQL, Scala, Lambda, processing data , python, MySQL, Auth0, Step functions, Jenkins.

Client: Advance Autoparts, Raleigh, NC Mar 2021 Dec 2022
Python Developer
Responsibilities:
Imported the data from different sources like cloud services, Local file system into Spark RDD and worked on cloud Amazon Web Services (EMR, S3, EC2, Lambda).
Developed the application by following Agile methodology and Scrum method of project management
Imported the data from different sources like cloud services, Local file system into Spark RDD and worked on cloud Amazon Web Services (EMR, S3, EC2, Lambda).
development utilizing Python, Apache Spark, Amazon EMR, S3, AWS Lambda, and Jenkins.
Worked with Spark for improving performance and optimization of the existing algorithms in Hadoop using Spark Context, Spark YARN.
Worked with testing methodologies like unit testing, Pytesting.
Creating the lambda for copying the files automatically when the files are available.
Developed Preprocessing job using Spark Data frames to flatten Json documents to flat file.
Analyze the customer data and account information before transmission and create the post and preprocessing record length.
Detokenizing the sensitive information using Turing before the file transfer.
Creating and Updating EFG routes as per requirement and mentoring the file transmission.
Strong experience programming, pointer manipulation, link list handling, hash - table lookup implementation & data- structure programming, recursive routines and UNIX library interface. OO programming using classes with inheritance and polymorphism, data structure programming using templates, exception handling, Stream & FilelO, threads, sockets, distributed systems.
Coordinating with different teams with file transmissions and failures.
Updating the IAM roles as per requirements and providing the access for users.
Created a highly-scalable, distributed system for running Web applications and Web services by working with architects and
developers and ensuring teh implementation of Restful Web services and Fast API design.
Fallow up with Bureaus regarding files failures and finding the solution and developing the application as per solution.
Worked on Jenkins for continuous integration to build code and deploy into remote servers.
Involving the developing the ATDD functionality and processing the test files.
Start involving with Quantum to start working with bureaus file process.
Involved in production deployments and implementing the deployments on prod.numpy

Environment: Python 3.7, JIRA, Spark, Spark-Streaming, Hadoop 2.7, AWS, Scala , distributed system EMR,S3,Mongo DB, Git, GitHub, AWS, Amazon Lambda, Jenkins,EC2,Apachi spark, Json

Gilead Sciences, Foster city, CA Jul 2019 Feb 2021
Python Developer
Responsibilities:
Worked in Agile development following Scrum process, Sprint, and daily stand-up meetings
Developed Views and Templates with Django view controller and template language to create a user-friendly website interface.
Worked on frontend and backend modules using Python on Django Web Framework.
Created application structure based on Angular style guides and best practices
Built S3 buckets and managed policies for S3 buckets and used S3 bucket and Glacier for storage and backup on AWS
Design and Developed Web Services, RESTful API s for Mobile Apps using Python Django-REST and Flask Frameworks
Developed complex database objects like stored Procedures, Functions, Packages, and triggers using SQL and PL/SQL.
Created a Lambda Deployment function, and configured it to receive events from your S3 bucket
Designing and implementing Docker workflow reduced built and deployment times.
Implemented user interface guidelines and standards throughout the development and maintenance of the website using the HTML, CSS, JavaScript, and jQuery.
Extensively used the advanced features of PL/SQL like records, Tables, Object types and Dynamic Sql.
Created APIs, database Model and Views Utilization Python to build responsive web page application.
Used GitHub for Python source code version control, Jenkins for automating the build Docker containers, and deploying in Mesos
Developed multiple React Redux based pages that provided functionality for Login/Logout, display of vendor data like vendor opportunity, top vendors, a tile-based display of important overview information, summary metrics etc.
Designed and implemented a dedicated MySQL database server to drive the web apps and report on daily progress.
Used PyUnit, the Python unit test framework, for all Python applications. Build application and database servers using AWS EC2 and create AMIs as well as use RDS for Oracle DB .
Developed Complex database objects like Stored Procedures, Functions, Packages and Triggers using SQL and PL/SQL.
Worked on Jenkins continuous integration tool for deployment of project.
Design and implement availability, scalability, and performance plans for the cloud managed service environment.
Extensively used bulk collection in PL/SQL objects for improving the performing.
Used Git for the version control.
Created a Git repository and added the project to GitHub.
Collaborated with internal teams and external teams to solve the issues and giving feedbacks.
Worked on object-oriented programming (OOP) concepts using Python, Django and Linux.
Used JIRA for bug tracking and issue tracking.
Practical experience with working on multiple environments like development, testing, production.
Environment: Python, Django, HTML5/CSS, MySQL, JavaScript, Eclipse, Linux, Shell Scripting, Ansible, jQuery, AWS, GitHub, JIRA, MS Office, and UNIX.

LPL, San Diego, CA Apr 2017 to June 2019
Python Developer
Responsibilities:
Followed Agile Methodologies to manage full life-cycle development of the project.
Developed views and templates with Python and Django view controller and templating language to create a user-friendly website interface.
Worked on large-scale, high-traffic Python web applications to define and implement new features, enhance core functionality, and integrate with other platforms and services.
Installation, configuring, upgrading, migrating, and patching of DB2, Oracle, and SQLServer.
Expertise in establishing database connections for Python by configuring packages like MySQL- Python.
Handled potential points of failure through error handling and communication of failure.
Performed efficient delivery of code based on principles of Test Driven Development (TDD) and continuous integration to keep in line with Agile Software Methodology principles.
Worked on installing and setting up the channels by using Django. Used JavaScript and JSON to update a portion of a webpage.
Build SQL queries for performing various CRUD operations like create, update, read and delete.
Skilled in using Collections in Python for manipulating and looping through different user defined objects.
Deployed the project into Heroku using the GIT version control system and Microservices.
Environment: DB2, SQL, Oracle, SQL Server, MySQL, Python, TDD, JSON, Django, GitHub, JIRA, MS Office, and UNIX.

People Tech, Hyderabad, India Jun 2015 - Nov 2016
Role: Python Developer.
Responsibilities:
Worked on back-end modules of the project are developed using Python with Django Framework.
Designed and developed the data management systems using MySQL databases.
Written Python Scripts to parse the XML documents and load the data in the database.
Utilized the existing Python and Django modules and rewritten them to deliver data in required formats.
Written indexes, views, constraints, stored procedures, triggers, cursors, and user-defined functions or subroutines in MySQL.
Responsible for debugging and troubleshooting the application.
Utilized a Subversion control tool to coordinate teamwork.
Developed views and templates with Python and Django have to create a user-friendly website interface.
Utilized Python libraries like wxPython, NumPy, SciPy, Web2py.
Designed Restful Web services using FLASK, with emphasis on improved Security for the service using FLASK-HTTP Auth with HTTPS.
Used Selenium Libraries to write a fully functioning test automation process.
Environment: Python 2.6, Django, UNIX, MySQL and Bugzilla.
Keywords: cprogramm cplusplus csharp continuous integration message queue javascript sthree database information technology microsoft procedural language quasar toolkit California Idaho North Carolina
Keywords: cprogramm cplusplus csharp continuous integration message queue javascript sthree database information technology microsoft procedural language quasar toolkit California Idaho North Carolina

To remove this resume please click here or send an email from [email protected] to [email protected] with subject as "delete" (without inverted commas)
[email protected];1563
Enter the captcha code and we will send and email at [email protected]
with a link to edit / delete this resume
Captcha Image: