Home

Big Data AI / Big data Admin : Houston, TX (Onsite) : Immediate Interviews at Houston, Texas, USA
Email: [email protected]
Hi,

Hope you are in fine spirits of health!

My name is 
Love Juneja
 and I am working with Alpha Silicon.

We are headquartered in Santa Clara, CA, serving customers nationwide in the United States.

We have the following job opening....

Job Title: Big Data AI Developer

Experience Required: 3-5 years

Location: Houston, TX (Onsite)

Job Summary:

We are seeking a skilled Big Data AI Developer to join our Big Data team. The role involves designing and deploying AI models using machine learning and deep learning techniques within big data environments. You will work with distributed computing technologies to create scalable and efficient AI solutions.

Key Responsibilities:

Develop AI models using supervised, unsupervised, and semi-supervised learning techniques.

Design and implement end-to-end machine learning pipelines from data ingestion and preprocessing to model training, evaluation, and deployment.

Utilize TensorFlow, PyTorch, and other frameworks for building deep learning models applicable to computer vision, NLP, and other AI domains.

Manage and optimize Spark-ML and Flink-ML jobs within distributed environments for large-scale machine learning tasks.

Implement NLP and computer vision algorithms to solve complex problems in data analytics.

Engage in graph-based machine learning using Neo4j to uncover insights from connected data.

Work with the Cloudera suite to maintain and manage big data workflows, ensuring compatibility with AI model requirements.

Optimize data storage and processing using technologies like Kafka, HDFS, HBASE, KUDU, and Cloudera Machine Learning.

Collaborate with infrastructure teams to deploy models using cloud-native technologies and Kubernetes orchestration.

Develop custom data models and algorithms, and apply advanced Python programming skills to solve challenging data science problems.

Stay current with AI research and implement novel algorithms that can contribute to business goals.

Collaborate with stakeholders to understand business challenges and translate them into technical solutions.

Ensure models comply with data privacy and security regulations, applying best practices in data governance.

Run AI Models on GPUs, tweak Models to utilize distributed GPUs.

Design and implement robust MLOps (Machine Learning Operations) workflows to automate the machine learning lifecycle, from data collection and model development to deployment and monitoring, ensuring continuous integration and delivery (CI/CD) for AI products.

 Technical Qualifications:

Strong expertise in machine learning and deep learning with frameworks like TensorFlow and PyTorch.

Experience with distributed data processing frameworks, particularly SPARK-ML and Flink-ML.

Proficient in Python for data science, with a solid understanding of libraries and toolkits for AI and machine learning.

Experience in RAPIDS & GPU-Aware Scheduling.

Familiarity with Cloudera Data Services and the broader Cloudera suite in a big data context.

Proven ability in leveraging Kubernetes for deploying scalable AI applications.

Knowledgeable in advanced areas of AI, such as large language models (LLM), and the ability to apply them to real-world scenarios.

Experience with computer vision, NLP, and graph-based machine learning techniques.

Understanding of data management practices and ETL processes within distributed environments.

 Education:

Bachelors or Masters degree in Computer Science, Data Science, AI, or related fields.

Certifications:

Relevant certifications in AI, big data technologies, or Kubernetes are advantageous.

Soft Skills:

Excellent problem-solving ability and critical thinking skills.

Strong communication skills for effectively conveying complex technical concepts.

Team-oriented mindset with experience working in collaborative environments.

Commitment to continued exploration and staying abreast of emerging AI technologies.

Job Title: Big Data Administrator

Experience Required: 3-5 years

Location: Houston, TX (Onsite)

Job Summary:

As a Big Data Administrator, you will be responsible for managing our Big Data infrastructure and tools within the Cloudera ecosystem. You will play a critical role in our data team, ensuring the high availability, performance, and security of our data solutions.

Key Responsibilities:

Install, configure, and maintain Cloudera Big Data clusters across multiple environments, ensuring optimal performance and resource utilization.

Perform cluster maintenance tasks such as patching, upgrades, and migrations, applying best practices for minimal downtime.

Ensure high availability and manage the performance of data services like Kafka, HDFS, HBASE, Hive, and Impala.

Architect, develop, and maintain scalable data pipelines, ensuring seamless data integration and flow across the organization.

Administer and manage cloud-native Big Data technologies and container orchestration systems, with a focus on Kubernetes (Openshift or equivalent).

Implement and manage security and compliance measures, applying Cloudera Ranger, Ranger KMS, KTS, and Atlas tools to protect data integrity.

Optimize Cloudera Data Services for data engineering, data warehousing, and machine learning in collaboration with the respective teams.

Monitor system performance actively, tune system settings, and carry out capacity planning to meet service level agreements (SLAs).

Write and maintain advanced scripts in Shell, Basic Java/Scala/Python for automation and routine task management.

Plan and execute system backups and recovery strategies to maintain data safety and high service availability.

Coordinate with IT infrastructure and development teams to align Big Data systems with organizational goals.

Keep up-to-date with the latest Big Data technologies and contribute to knowledge sharing and team training.

Document system configurations, maintain standard operating procedures, and manage version control for infrastructure changes.

Liaise with software vendors for technical support, ensuring service standards and levels are maintained.

Responsible for the implementation, management, and troubleshooting of Kerberos authentication protocol within the big data ecosystem to ensure secure access control across all nodes and services. This includes configuring Kerberos principals, keytab files, and managing Kerberos tickets.

You will also be expected to keep the Kerberos infrastructure updated and secure, and work closely with the security team to audit and improve authentication measures regularly.

Ensure the confidentiality, integrity, and availability of data by implementing industry-standard encryption methods. This includes configuring TLS (Transport Layer Security) for data in transit to protect data as it travels across the network, and managing encryption for data at rest to prevent unauthorized access. The role also involves regular audits, developing, and maintaining security policies, and procedures for data encryption, as well as performing vulnerability assessments and remediation.

Technical Qualifications:

Proficiency in the Cloudera suite, including Kafka, HDFS, HBASE, KUDU, Zookeeper, HIVE, Impala, NIFI, SPARK, FLINK, Oozie, Yarn, Atlas, Ranger, RangerKMS, and KTS.

Experience with Cloudera ECS and Cloudera Data Services such as Cloudera Data Engineering, Cloudera Data Warehouse, and Cloudera Machine Learning.

Familiarity with container orchestration systems like Kubernetes (Openshift or equivalent).

Strong background in Linux system administration and shell scripting.

Basic programming skills in Java, Scala, or Python.

Experience with graph databases like Neo4j is a plus.

Knowledge of cloud-native Big Data technologies and architectural principles.

In-depth knowledge and hands-on experience in configuring and managing Kerberos authentication for a secure big data environment. Proficiency in setting up Kerberos realms, troubleshooting Kerberos authentication issues, and performing regular security checks and configuration audits.

Strong understanding of encryption protocols and standards for securing data. Experience with configuring TLS/SSL for data in transit across distributed networks and implementing encryption solutions for data at rest, using technologies like Hadoops encryption zones or third-party encryption tools. Familiarity with key management practices and security compliance requirements for data protection.

Education:

Bachelors degree in Computer Science, Information Technology, or a related field.

Certifications:

Cloudera Certified Administrator is preferred.

Kubernetes certification is advantageous.

Soft Skills:

Excellent problem-solving and analytical skills.

Strong communication and collaboration abilities.

Proactive attitude with a willingness to take initiative.

Thanks and Regards

Love Juneja

US IT RECRUITER

Phone: 951-800-2167

Email Id: [email protected]

--

Keywords: continuous integration continuous deployment artificial intelligence machine learning access management information technology California Idaho Texas
[email protected]
View all
Thu Nov 30 19:49:00 UTC 2023

To remove this job post send "job_kill 896332" as subject from [email protected] to [email protected]. Do not write anything extra in the subject line as this is a automatic system which will not work otherwise.


Your reply to [email protected] -
To       

Subject   
Message -

Your email id:

Captcha Image:
Captcha Code:


Pages not loading, taking too much time to load, server timeout or unavailable, or any other issues please contact admin at [email protected]
Time Taken: 9

Location: Houston, Texas