Description:
Key Responsibilities:Design and build scalable big data applications using open-source technologies like Spark, Hive, Kafka Develop data pipelines and orchestrate workflows using Apache Airflow Implement and optimize ETL/ELT pipelines in Google Cloud Platform (Dataproc, GCS, BigQuery) Model and design schemas for data lakes and RDBMS platforms Automate data workflows and manage multi-TB/PB scale datasets Provide ongoing support, maintenance, and participate in on-call rotations Collaborate with
Jul 24, 2025;
from:
dice.com