Description:
Proficiency in managing and manipulating huge datasets in the order of terabytes (TB) is essential. Expertise in big data technologies like Hadoop, Apache Spark (Scala preferred), Apache Hive, or similar frameworks on the cloud (Google Cloud Platform preferred, AWS, Azure etc.) to build batch data pipelines with strong focus on optimization, SLA adherence and fault tolerance. Expertise in building idempotent workflows using orchestrators like Automic, Airflow, Luigi etc. Expertise in writing SQL
Jun 6, 2025;
from:
dice.com