Home

senior data engineer at Boston, MA for Hybrid at Quincy, Massachusetts, USA
Email: [email protected]
From:

alka,

Allied Business Consulting

[email protected]

Reply to:   [email protected]

Sr Data Engineer

Location: Quincy, MA - hybrid

Open to any visa

Key Responsibilities:
Data Integration & API Development: Integrate cybersecurity data sources and build data APIs for real-time insights and streamlined access.
Pipeline Engineering: Design and optimize large-scale ETL/ELT pipelines on Databricks using Python and PySpark.
Data Quality & Governance: Implement automated data checks, manage data lineage, and ensure compliance with cybersecurity standards.
Analytics & Visualization: Create data models, visualizations, and dashboards for threat and vulnerability insights using Databricks and React.js.
Data Architecture: Develop secure, scalable environments on AWS (S3, ELB, Lambda) for data storage and processing.
CI/CD for Data & Security: Use CI/CD pipelines to automate testing and deployment in alignment with Agile practices.
ML Integration: Deploy ML models for threat detection, anomaly detection, and risk scoring.
Mentorship: Guide junior engineers, establish best practices, and foster a high-performance, innovative culture.

Qualifications:

The CASM  Platform Product & Engineering team is looking for a Sr. Data Engineer, Cyber AI App & Attack Surface Management. CASM Platform team is focused on building enterprise-level robust and scalable Cyber AI applications and Platform to manage vulnerability, risk, security configurations, and attack surfaces oversight across the bank. We are seeking an experienced, hands-on Senior Data Engineer to join our CASM Platform team. Ideal

candidates bring a deep understanding of data engineering within a cybersecurity context, extensive

experience with cloud-native tools (especially AWS), and a passion for automation, security, intelligent

orchestration, and resilience. Youll work in high-paced Agile Product Engineering cross-functional

teams, including Product, AI application, DevOps, and Data Quality Management (DQM), to transform

the CASM platform into a single pane of glass for secure and insightful network asset and risk visibility.

We have multiple openings for this role, and it is open to candidates with varying levels of experience.

Role Responsibilities

As a Senior Data Engineer, CASM Platform, you will:
 Data Integration, API Development: Integrate diverse cybersecurity data sources using varietyof API mechanisms and to standardize and streamline data across the data and user planes.
Build and maintain data APIs for seamless access to data pipelines, enabling real-time insights for applications, machine learning models, and analytical layers.
 Data Pipeline Engineering Optimization: Design, develop, and optimize large-scale ETL/ELTpipelines on Databricks to efficiently process and transform cybersecurity data. Utilize Python,PySpark, and Databricks to automate and standardize data workflows across stages (raw,cleaned, curated), ensuring scalability and high performance.
 Data Quality & Governance: Implement automated data quality checks, leveraging DatabricksDQM tools and CI/CD pipelines to uphold data integrity and governance standards. Ensure datlineage, metadata management, and compliance with cybersecurity and privacy regulations,applying rigorous quality standards across data ingestion and processing workflows.
 Data Analytics & Visualization: Design centralized data models and perform in-depth data analysis to support cybersecurity and risk management objectives. Develop visualizations anddashboards using tools like Databricks, encapsulate data to spin up to React.js application layerto provide stakeholders with actionable insights into threat landscapes, vulnerability trends, andperformance metrics across the platform.
Scalable & Secure Data Architecture: Architect and manage secure, high-performance dataenvironments on Databricks, utilizing AWS services such as S3, ELB, and Lambda. Ensure dataavailability, consistency, and security, aligning with AWS best practices and data encryptionstandards to safeguard sensitive cybersecurity data.
 Agile Product; Engineering Continuous Delivery: Collaborate with advanced Agile Product; Engineering cross-functional teams to deliver data-driven insights through analytics tools andcustom visualizations that inform strategy and decision-making. Empower stakeholders withtimely, actionable intelligence from complex data analyses, enhancing their ability to respond toevolving cybersecurity risks.
 Data Science & ML Integration: deploy machine learning models, including predictive analytics,anomaly detection, and risk scoring algorithms, into the CASM platform. Leverage Python andPySpark to enable real-time and batch processing of model outputs, enhancing CASM Platforms proactive threat detection and response capabilities.
 Mentorship; Best Practices Promotion: Mentor junior engineers, establishing best practices in data engineering, DevOps, data science, and analytics. Encourage high standards in modeldeployment, data security, performance optimization, and visualization practices, fostering aculture of innovation and excellence.

Education Qualifications:

Minimum Qualifications

Education: B.S., M.S., or Ph.D. in Computer Science, Data Science, Information Systems, or a

related field, or equivalent professional experience.

Technical Expertise: 8+ years in data engineering with strong skills in Python, PySpark, SQL,

and extensive, hands-on experience with Databricks and big data frameworks. Expertise in

integrating data science workflows and deploying ML models for real-time and batch processing

within a cybersecurity context.

Cloud Proficiency: Advanced proficiency in AWS, including EC2, S3, Lambda, ELB, and

container orchestration (Docker, Kubernetes). Experience in managing large-scale data

environments on AWS, optimizing for performance, security, and compliance.

Security Integration: Proven experience implementing SCAS, SAST, DAST/WAS, and secure

DevOps practices within an SDLC framework to ensure data security and compliance in a high-

stakes cybersecurity environment.

Data Architecture: Demonstrated ability to design and implement complex data architectures,

including data lakes, data warehouses, and lake house solutions. Emphasis on secure, scalable,

and highly available data structures that support ML-driven insights and real-time analytics.

Data Quality ; Governance: Hands-on experience with automated data quality checks, data

lineage, and governance standards. Proficiency in Databricks DQM or similar tools to enforce

data integrity and compliance across pipelines.

Data Analytics; Visualization: Proficiency with analytics and visualization tools such as

Databricks, Power BI, and Tableau to generate actionable insights for cybersecurity risks, threat

patterns, and vulnerability trends. Skilled in translating complex data into accessible visuals and

reports for cross-functional teams.

CI/CD and Automation: Experience building CI/CD pipelines that automate testing, security

scans, and deployment processes. Proficiency in deploying ML models and data processing

workflows using CI/CD, ensuring consistent quality and streamlined delivery.

Agile Experience: Deep experience in Agile/Scrum environments, with a thorough

understanding of Agile core values and principles, effectively delivering complex projects with

agility and cross-functional collaboration.

Preferred Experience:

Advanced Data Modeling; Governance: Expertise in designing data models for cybersecurity

data analytics, emphasizing data lineage, federation, governance, and compliance. Experience

ensuring security and privacy within data architectures.

Machine Learning; Predictive Analytics: Experience deploying ML algorithms, predictive

models, and anomaly detection frameworks to bolster CASM platforms cybersecurity

capabilities.

High-Performance Engineering Culture: Background in mentoring engineers in data

engineering best practices, promoting data science, ML, and analytics integration, and fostering

a culture of collaboration and continuous improvement.

Keywords: continuous integration continuous deployment artificial intelligence machine learning javascript business intelligence sthree information technology Massachusetts
senior data engineer at Boston, MA for Hybrid
[email protected]
[email protected]
View all
Thu Nov 07 19:16:00 UTC 2024

To remove this job post send "job_kill 1909361" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 6

Location: Quincy, Massachusetts