Urgent Need for Data scientist - TCS - Hopkins, MN at Hopkins, Minnesota, USA |
Email: [email protected] |
From: Nithiya Bharath, Teamware solutions [email protected] Reply to: [email protected] Hi This is Bharath from Teamware solutions. Hope you are doing well. This email is regarding a job opportunity, Hope I reached the perfect candidate. ______________________________________________________________________________________ Below is the Job Description for your reference Data Scientist Hopkins, MN Onsite Contract Keywords: Python, Bigdata, PySpark, Competencies: Data Build Tool, Azure Data Factory Required: Develop and Maintain Data Pipelines using python bigdata pyspark Design, build, and maintain efficient, reusable, and reliable data pipelines. Integrate data from various sources, including databases, APIs, and flat files. Implement ETL processes to support data transformation and loading. Ensure the robustness and reliability of data processing systems. Develop scripts to automate repetitive tasks and improve data processing efficiency. Ensure scripts are well-documented and maintainablemanage Data and migration: Work with relational and NoSQL databases to store and retrieve data. Optimize database performance and ensure data integrity. Also contribute to the migration of data from RDMS to cloud data source. Design and implement robust data models to support analytical use cases. Work closely with data analysts, data scientists, and other stakeholders to understand data requirements and deliver solutions. Participate in code reviews and provide constructive feedback. Implement Data Strategy:Contribute to the development and execution of the data strategy. Assist in the design and implementation of data governance and data quality frameworks. Tune Performance:Identify bottlenecks and bugs, and devise solutions to address these issues. Optimize the performance of data processing workflows.Document processes:Maintain comprehensive documentation for all data processes, pipelines, and systems. Ensure that documentation is up-to-date and accessible to relevant stakeholders. Formal training or certification on Python and relevant libraries concepts and proficient applied experience. Proficient at Python and relevant libraries (e.g., Pandas, NumPy, SQLAlchemy). Expert in Database, PL\\SQL, Performance tuning, DB modelling, Erwin, DB query review, database query optimization. Experienced development in a data lake area using Databricks, Redshift or Snowflake tools. Experience of working on streaming data applications such as Spark Streaming, Kafka, MSK, Kinesis.Knowledge of cloud platforms (e.g., AWS, Azure, GCP) and their data services. Knowledge of machine learning and data science principles. Understanding of data governance and data quality principles.. Preferred qualifications, capabilities, and skills Experience working with ETL development Experience with DevOps including Continuous Integration (CI) and Continuous Deployment (CD) tools e.g. Jenkins, Sonar. Exposure to scheduling tools like Autosys / Control-M. Experience with big data technologies (e.g. Spark, Hadoop). Familiarity with data integration tools (e.g., Apache Airflow) and data warehousing solutions (e.g.Databricks, Redshift). Work effectively within Agile development framework to ensure timely and efficient project delivery If you're okay with that please share you most updated resume to my below email address. email: [email protected] Keywords: continuous integration continuous deployment database procedural language Minnesota Urgent Need for Data scientist - TCS - Hopkins, MN [email protected] |
[email protected] View all |
Fri Nov 08 23:33:00 UTC 2024 |