Remote Work --- GCP Engineer --- Long term contract at Remote, Remote, USA |
Email: [email protected] |
From: Mohit Jaiswal, Intelligenz IT [email protected] Reply to: [email protected] Hi, I hope you and your family are doing well. I have a good position for you with my client. This point of time I don't know whether you are looking for a new job or not. But just thought if I can share the details and then confirm from you about your interest level for the opportunity. If you are interested and an available consultant, send me your most updated resumes in word format and contact details. Job Title: GCP Engineer Location: Remote Work Duration: Long term contract Must Green card Holder & Us Citizen (NO H1B, GC EAD) Start Date ASAP... Steer Clear No one < 3 years of experience on GCP No one with only experience in AWS and Azure Data Engineering Requirement GCP BigQuery Dataflow (Apache Beam) Cloud Composer (Airflow) GCS GKE Dataform (Optional to dbt) Programming SQL Python Java (Optional) Tools dbt / Dataform(On GCP) Test Automation on Data Misc.(Good to Have): Docker Kubernetes Microservices Experience: Data Platform Building (Mandatory) Ingestion/Migration Transformation/ETL Analysis (Optional) Visualization (SAP BOBJ/Looker/PowerBI) Governance (Unitiy Catalog(tool in databricks)/Colibra(tool)/Data Catalog(Service in GCP) Security (GCP Services IAM, KMS, DLP. Techniques ACL's, Row/Column level in BigQuery) Deployment CI/CD Github Cloud Build (Service in GCP) + Terraform Requirements: 10+ (for senior) 15+ (for principal) of proven experience in modern cloud data engineering, broader data landscape experience and exposure and solid software engineering experience. Prior experience architecting and building successful enterprise scale data platforms in a green field environment is a must. Proficiency in building end-to-end data platforms and data services in GCP is a must. Proficiency in tools and technologies: BigQuery, Cloud Functions, Cloud Run, Dataform, Dataflow, Dataproc, SQL, Python, Airflow, PubSub. Experience with Microservices architectures - Kubernetes, Docker and Cloud Run Experience building Symantec layers. Proficiency in architecting and designing and development experience with batch and real time streaming infrastructure and workloads. Solid experience with architecting and implementing metadata management including data catalogues, data lineage, data quality and data observability for big data workflows. Hands-on experience with GCP ecosystem and data lakehouse architectures. Strong understanding of data modeling, data architecture, and data governance principles. Excellent experience with DataOps principles and test automation. Excellent experience with observability tooling: Grafana, Datadog. Regards, Mohit Jaiswal Intelligenz IT Work : 646-502-7441 Maillot: [email protected] Keywords: continuous integration continuous deployment information technology green card Remote Work --- GCP Engineer --- Long term contract [email protected] |
[email protected] View all |
Fri Jun 21 19:43:00 UTC 2024 |