




**Job Description** We are seeking a **Senior Big Data Engineer** with solid experience in distributed platforms and Hadoop ecosystems, specifically with **Cloudera 7**, to join a large-scale data analysis and processing project at a leading financial institution. The candidate will be responsible for developing, maintaining, and optimizing data processes in distributed environments, ensuring efficiency, quality, and reliability of information flows. **Responsibilities** * Develop and maintain data pipelines on Cloudera 7 platforms. * Create, optimize, and execute queries in **Hive**. * Manage and monitor resources using **YARN**. * Design and administer workflows using tools such as Oozie or equivalents. * Implement and optimize processes in **Hadoop, HDFS, and MapReduce**. * Develop data applications and transformations in **Spark** (PySpark or Scala). * Troubleshoot issues, analyze performance, and apply improvements to distributed clusters. * Collaborate with business, architecture, and data teams to understand technical requirements. * Document processes, best practices, and development standards. * **Requirements (Mandatory)** * Degree in Systems Engineering, Computer Science, Computing, or related field. * **3–5+ years of experience** in Big Data ecosystems. * Proven experience with **Cloudera 7**. * Proficiency in: * **Hive** * **Hadoop / HDFS** * **YARN** * **MapReduce** * **Spark** (PySpark or Scala) * Workflows (Oozie, Airflow, or similar) * Experience handling large volumes of data. * Intermediate-to-advanced Linux knowledge. * Experience optimizing distributed processes. Employment type: Full-time Salary: $40,000.00 – $45,000.00 per month Work location: Remote


