




**Key Responsibilities** * Design, implement, and automate ML lifecycle workflows using tools like **MLflow** , **Kubeflow** , **Airflow** and **OCI Data Science Pipelines** . * Build and maintain **CI/CD pipelines** for model training, validation, and deployment using **GitHub Actions** , **Jenkins** , or **Argo Workflows** . * Collaborate with data engineers to deploy models within **modern data lakehouse architectures** (e.g., **Apache Iceberg** , **Delta Lake** , **Apache Hudi** ). * Integrate machine learning frameworks such as **TensorFlow** , **PyTorch** , and **Scikit\-learn** into distributed environments like **Apache Spark** , **Ray** , or **Dask** . * Operationalize model tracking, versioning, and drift detection using **DVC** , model registries, and ML metadata stores. * Manage **infrastructure as code (IaC)** using tools like **Terraform** , **Helm** , or **Ansible** to support dynamic GPU/CPU training clusters. * Configure real\-time and batch data ingestion and feature transformation pipelines using **Kafka** , **Goldengate** and **OCI Streaming** . * Collaborate with DevOps and platform teams to implement robust **monitoring, observability** , and **alerting** with tools like **Prometheus** , **Grafana** , and the **ELK Stack** . * Support **AI governance** by enabling model explainability, audit logging, and compliance mechanisms aligned with enterprise data and security policies. **Required Qualifications** * Bachelor’s or Master’s degree in **Computer Science** , **Data Science** , or a related technical discipline. * **5–8 years** of experience in **ML engineering** , **DevOps** , or **data platform engineering** , with at least **2 years in MLOps** or model operations. * Proficiency in **Python** , particularly for automation, data processing, and ML model development. * Solid experience with **SQL** and distributed query engines (e.g., **Trino** , **Spark SQL** ). * Deep expertise in **Docker** , **Kubernetes** , and cloud\-native container orchestration tools (e.g., **OCI Container Engine** , **EKS** , **GKE** ). * Working knowledge of **open\-source data lakehouse frameworks** and **data versioning** tools (e.g., **Delta Lake** , **Apache Iceberg** , **DVC** ). * Familiarity with model deployment strategies, including **batch** , **real\-time inference** , and **edge deployments** . * Experience with **CI/CD pipelines** (GitHub Actions, GitLab CI, Jenkins) and **MLOps frameworks** (Kubeflow, MLflow, Seldon Core). * Competence in implementing monitoring and logging systems (e.g., **Prometheus** , **ELK Stack** , **Datadog** ) for ML applications. * Strong understanding of **cloud platforms** (OCI, AWS, GCP) and **IaC tools** (Terraform, CloudFormation). **Preferred Qualifications** * Experience integrating AI workflows with **Oracle Data Lakehouse** , **Databricks** , or **Snowflake** . * Hands\-on experience with orchestration tools like **Apache Airflow** , **Prefect** , or **Dagster** . * Exposure to **real\-time ML systems** using **Kafka** or **Oracle Stream Analytics** . * Understanding of **vector databases** (e.g., **Oracle 23ai Vector Search** ). * Knowledge of **AI governance** , including model explainability, auditability, and reproducibility frameworks. **Soft Skills** * Strong **problem\-solving** skills and an automation\-first mindset. * Excellent **cross\-functional communication** , especially when collaborating with data scientists, DevOps, and platform engineering teams. * A collaborative and **knowledge\-sharing** attitude, with good documentation habits. * Passion for **continuous learning** , especially in the areas of AI/ML tooling, open\-source platforms, and data engineering innovation.


