Description:
Key ResponsibilitiesDesign and implement end-to-end data architecture on Databricks (Spark, Delta Lake, MLflow). Develop and optimize large-scale ETL/ELT pipelines using PySpark/SQL. Architect data lakes and lakehouses integrating cloud storage systems (e.g., ADLS, S3, GCS). Define and enforce best practices around security, data governance, and cost optimization. Lead technical workshops and collaborate with data engineers, data scientists, and DevOps. Implement data orchestration workflows (e.
Jun 6, 2025;
from:
dice.com