Home

SR. DATA ENGINEER - Hybrid -Laurel MD local candidate only or near by who can travel for 3 days onsite at Laurel, New York, USA
Email: [email protected]
Hi Employer,

Please share Sr, resumes
12+
and Local only

Position:
Sr Data Engineer

Location:
14501 Sweitzer Lane, Laurel, MD 20707

Hours 8-5pm

3 days in office, 2 days remote

Work Auth:
Open

Duration:
2-year contract

Work Schedule
Hybrid

Background check:
Yes

Vaccination required: No

Scope Of Work:
Build and manage scalable data pipelines and solutions in Azure, utilizing Synapse Analytics and Microsoft Fabric to support analytics needs. The Senior Data Engineer with expertise in Azure Synapse Analytics and Microsoft Fabric will design, develop, and implement scalable data solutions to support analytics and reporting needs. The role involves creating, optimizing, and managing data pipelines to efficiently move and transform data across the Azure ecosystem. The candidate will be responsible for setting up and managing data lakes to store large volumes of structured and unstructured data, ensuring high availability and security. They will collaborate with cross[1]functional teams to gather data requirements and create efficient, scalable architectures. Strong experience in ETL development, data modeling, and cloud technologies like Azure Data Factory, Azure Data Lake, and Synapse Analytics is essential. The candidate will also ensure data quality, security, and compliance with governance standards.

Create and optimize complex data pipelines using Azure Data Factory, Synapse Analytics, and other Azure tools to extract, transform, and load data efficiently.

Implement and maintain Azure Data Lake solutions to store large volumes of structured and unstructured data, ensuring scalability, performance, and security.

Integrate data from various sources, including relational databases, NoSQL databases, APIs, and flat files, into the Azure environment for analysis and reporting

Design and develop robust data architectures, optimizing for performance and scalability in Azure Synapse Analytics and Azure Data Lake environments.

Develop efficient ETL/ELT processes using Azure Data Factory or other Azure tools to ensure timely and accurate data loading and transformation.

Ensure data pipelines run smoothly by monitoring, troubleshooting, and resolving issues to minimize downtime and data inconsistencies.

Continuously optimize data pipelines and query performance, especially within Azure Synapse to handle large data sets and complex transformations efficiently.

Work closely with data scientists, analysts, and business teams to understand data requirements and deliver scalable data solutions that support analytics needs.

Implement and enforce security best practices, ensuring data lakes, pipelines, and analytics solutions comply with Azure security standards and data governance policies.

Design and implement logical and physical data models that support high[1]performance querying and reporting within Azure Synapse.

Implement data quality checks, data validation processes, and error handling within data pipelines to ensure accuracy and consistency of data.

Ensure adherence to data governance frameworks, managing data lineage, metadata, and ensuring compliance with organizational and regulatory requirements.

Implement data partitioning and indexing strategies to improve query performance within data lakes and Synapse.

Automate data ingestion, transformation, and processing tasks to ensure efficient and scalable data workflows within the Azure environment.

Create and maintain detailed documentation for data architectures, pipelines, processes, and data models, ensuring transparency and ease of maintenance.

Provide technical guidance and mentorship to junior data engineers, sharing best practices and ensuring adherence to high-quality engineering standards.

Monitor resource utilization in Azure environments, planning for future data growth and ensuring efficient use of cloud resources.

Strong knowledge of Medallion architecture.

Experience in setting up parquet and delta file structures.

Experience in working with non-structured data sources.

Strong knowledge of consuming and exposing data from various data sources like XML, JSON etc.

Continuously stay informed on the latest features and best practices in Azure Synapse Analytics, Microsoft Fabric, and the Azure ecosystem, implementing improvements as needed.

Strong knowledge of python for creating and scheduling data pipelines

Implement real-time data ingestion and processing pipelines using technologies like Azure Stream Analytics, Event Hubs.

Design and implement a data mesh architecture to support decentralized data ownership and self-service data infrastructure, ensuring scalable and flexible data management across the organization.

Architect and manage multi-cloud data solutions, integrating data across different cloud platforms (e.g., AWS, OCI) with Azure Synapse for a unified data and analytics ecosystem.

Design and manage hybrid data architectures that integrate on[1]premises data centers with Azure cloud environments, ensuring seamless data movement and synchronization between cloud and on[1]prem systems.

Utilize advanced data cataloging tools such as Azure Purview to create an enterprise-wide data catalog, enabling efficient data discovery and usage across various teams.

Create and automate end-to-end machine learning pipelines that integrate data ingestion, feature engineering, model training, and deployment using Azure ML, Python (scikit-learn, TensorFlow, PyTorch), and Azure Synapse Analytics.

Utilize Python-based data augmentation techniques or synthetic data generation (e.g., GANs or SMOTE) to enrich datasets for machine learning training, especially in cases where data is limited or imbalanced.

Preferred Experience/Qualification/Knowledge Skills

Education: Bachelors Degree in Information Systems, Computer Science, or related scientific or technical field and three (5) years minimum of relevant experience.

General Experience:

Work Experience: 5+ years of experience designing and implementing data solutions and creating data pipelines at enterprise-level applications.

Industry Knowledge: Preferred to have experience in water and wastewater industry understanding of oracle utility applications.

Project Experience: Demonstrated experience working on large-scale data projects in diverse team environments with a focus on analytics, business intelligence, and enterprise systems

Specialized Experience

Data Modeling: Extensive experience with data modeling and database design.

Enterprise Analytics: Proven expertise in implementing enterprise[1]wide analytics and business intelligence solutions, including data integration from multiple systems into a single data repository

Skillset Database & Data Structures: Deep understanding of database design principles, SQL, PL/SQL, and Oracle database management systems, including performance optimization and troubleshooting.

o

Data Governance & Quality: Familiarity with data governance frameworks, ensuring data integrity, quality, and security within an enterprise context.

o

Data Lakes: Strong experience in creating data lakes and data warehouses.

o

Python: Strong knowledge of writing python code to create and manage data pipelines

o

Communication & Collaboration: Excellent verbal and written communication skills, with the ability to work closely with stakeholders to translate business needs into technical solutions.

o

Problem-Solving: Strong analytical skills and problem-solving abilities, especially when working with large, complex datasets

Thanks & Regards

Mohd shuaib

[email protected]

www.signinsol.com

To follow and receive more updates please
Click Here

--

Keywords: machine learning information technology procedural language Maryland
SR. DATA ENGINEER - Hybrid -Laurel MD local candidate only or near by who can travel for 3 days onsite
[email protected]
[email protected]
View all
Wed Oct 23 00:17:00 UTC 2024

To remove this job post send "job_kill 1866507" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 18

Location: , Maryland