Job Description: Delivery Manager – Data Engineering (Databricks & Snowflake)
Position: Delivery Manager – Data Engineering
Location: Bavdhan/Baner, Pune.
Experience: 7-10 years
Employment Type: Full-time
Job Summary:
We are seeking a Delivery Manager – Data Engineering to oversee multiple data engineering projects leveraging Databricks and Snowflake. This role requires strong leadership skills to manage teams, ensure timely delivery, and drive best practices in cloud-based data platforms. The ideal candidate will have deep expertise in data architecture, ETL processes, cloud data platforms, and stakeholder management.
Key Responsibilities:
Project & Delivery Management:
• Oversee the end-to-end delivery of multiple data engineering projects using Databricks and Snowflake.
• Define project scope, timelines, milestones, and resource allocation to ensure smooth execution.
• Identify and mitigate risks, ensuring that projects are delivered on time and within budget.
• Establish agile methodologies (Scrum, Kanban) to drive efficient project execution.
Data Engineering & Architecture Oversight:
• Provide technical direction on data pipeline architecture, data lakes, data warehousing, and ETL frameworks.
• Ensure optimal performance, scalability, and security of data platforms.
• Collaborate with data architects and engineers to design and implement best practices for data processing and analytics.
Stakeholder & Client Management:
• Act as the primary point of contact for clients, senior management, and cross-functional teams.
• Understand business requirements and translate them into technical solutions.
• Provide regular status updates and manage client expectations effectively.
Team Leadership & People Management:
• Lead, mentor, and develop data engineers, architects, and analysts working across projects.
• Drive a culture of collaboration, accountability, and continuous learning.
• Ensure proper resource planning and capacity management to balance workload effectively.
Technology & Process Improvement:
• Stay up-to-date with emerging trends in Databricks, Snowflake, and cloud data technologies.
• Continuously improve delivery frameworks, automation, and DevOps for data engineering.
• Implement cost-optimization strategies for cloud-based data solutions.
Required Skills & Experience:
Technical Expertise:
-10+ years of experience in data engineering and delivery management.
-Strong expertise in Databricks, Snowflake, and cloud platforms (AWS, Azure, GCP).
-Hands-on experience in ETL, data modeling, and big data processing frameworks (Spark, Delta Lake, Apache Airflow, DBT).
-Understanding of data governance, security, and compliance standards (GDPR, CCPA, HIPAA, etc.).
-Familiarity with SQL, Python, Scala, or Java for data transformation.
Project & Team Management:
-Proven experience in managing multiple projects simultaneously.
-Strong knowledge of Agile, Scrum, and DevOps practices.
-Experience in budgeting, forecasting, and resource management.
Soft Skills & Leadership:
-Excellent communication and stakeholder management skills.
-Strong problem-solving and decision-making abilities.
-Ability to motivate and lead cross-functional teams effectively.
Preferred Qualifications:
🔹 Experience with data streaming (Kafka, Kinesis, or Pub/Sub).
🔹 Knowledge of ML & AI-driven data processing solutions.
🔹 Certifications in Databricks, Snowflake, or cloud platforms (AWS/Azure/GCP).
Apply or share your updated CV at hr@anvicybernetics.com