Data EngineerS at Remote, Remote, USA |
Email: [email protected] |
From: Vicky, istaffx [email protected] Reply to: [email protected] Develop, maintain, and optimize data pipelines to extract, transform, and load large datasets from diverse sources into our data ecosystem. Design and implement efficient and scalable data models that align with business requirements, ensuring data integrity and performance. Collaborate with cross-functional teams to understand data needs and deliver solutions that meet those requirements. Work closely with data scientists, analysts, and software engineers to ensure seamless integration of data solutions into larger systems. Identify and resolve data quality issues, ensuring accuracy, reliability, and consistency of the data infrastructure. Continuously monitor and improve data pipelines and processes, identifying opportunities for automation and optimization. Stay updated with emerging trends, technologies, and best practices in data engineering, data modeling, and backend Java engineering. Provide technical guidance and mentorship to junior team members, fostering their growth and development. Requirements: Bachelor's or Master's degree in Computer Science, Engineering, or a related field. 5+years of hands-on experience as a Data Engineer, working on complex data projects and implementing data modeling solutions. Data Engineering Must have Solid understanding of SQL and expertise in working with relational databases (e.g., PostgreSQL, MySQL). In-depth knowledge of data modeling techniques and experience with data modeling tools. Proficiency in designing and optimizing data pipelines using ETL/ELT frameworks and tools (e.g., Informatica, Apache Spark, Airflow, AWS Glue). Working knowledge on Data warehousing Familiarity with cloud-based data platforms and services (e.g., Snowflake, AWS, Google Cloud, Azure). Experience with version control systems (e.g., Git) and agile software development methodologies. Strong communication skills to effectively convey technical concepts to both technical and non-technical stakeholders. Excellent problem-solving skills and the ability to work independently and collaboratively in a fast-paced environment. Good to Have JAVA 8, REST APIs, and microservices, Spring Boot framework Alteryx UNIX scripting Primary Skillset: Data Engineering Solid understanding of SQL and expertise in working with relational databases (e.g., DB2, MySQL). Data Modelling knowledge cloud-based data platforms like Snowflake Working knowledge on Data warehousing Cloud-based data platforms and services (e.g., AWS, Google Cloud, Azure) Alteryx (good to have) ETL/ELT tools like Informatica, Apache Spark UNIX scripting Good to have: Client specific tools like Alloy Registry, LEGEND tool for Alloy data modelling, PURE, Data Browser, Data Lake etc Keywords: |
[email protected] View all |
Fri Jan 26 22:23:00 UTC 2024 |