CB-Data Engineer-L3-Bangalore at CoffeeBeans – Bengaluru, Karnātaka
CoffeeBeans
Bengaluru, Karnātaka, 560023, India
Posted on
NewIndustries:ConsultingEmployment Type:Full-TimeExperience Level:5-7 Years
New job! Apply early to increase your chances of getting hired.
About This Position
Location: Bangalore (with short/medium-term travel opportunities within India & foreign location)
Experience: 6–9 years
Job Type: Full-timeAbout the RoleWe’re looking for seasoned data engineers who can take end-to-end ownership of complex data engineering engagements—right from architecture and design to development and delivery. You’ll work with client stakeholders, lead data initiatives, and guide teams in building modern, scalable data platforms that support business intelligence, advanced analytics, and AI/ML applications.Key ResponsibilitiesLead the design and implementation of scalable, secure, and high-performance data architectures.
Architect and optimize data pipelines for batch and real-time processing across varied source systems.
Model and implement enterprise-grade data lakes, lakehouses, and warehouses to support analytical workloads.
Collaborate with client teams and internal stakeholders to define data strategies, tooling, and standards.
Set up and enforce best practices around code quality, pipeline observability, testing, and CI/CD.
Guide technical decision-making and mentor junior/mid-level engineers.
Troubleshoot complex data issues and provide architectural solutions across heterogeneous systems.
Must-Have SkillsStrong expertise in Python and advanced proficiency in SQL.
Deep understanding of data engineering architecture: ETL/ELT pipelines, data modeling (dimensional and normalized), warehousing, streaming systems, and data integration patterns.
Proven experience with distributed data processing frameworks (e.g., Apache Spark, Flink).
Proficient with orchestration tools like Airflow or similar, and production-grade data workflows.
Solid grounding in cloud-native data services (compute, storage, IAM, monitoring).
Experience leading technical implementations and collaborating with business stakeholders.Good to HaveMulti-cloud experience (AWS preferred; GCP or Azure familiarity is valued).
Hands-on experience with Databricks, dbt, and collaborative data development environments.
Awareness of data governance, security, lineage, and quality frameworks.
Certifications such as AWS Certified Data Analytics, Databricks Certified Data Engineer Professional, or equivalent.Other ExpectationsStrong leadership, ownership, and client communication skills.
Ability to balance hands-on contributions with technical guidance.
Adaptability to new tools, tech stacks, and problem domains.
Willingness to travel within India for short/medium-term client assignments.
Experience: 6–9 years
Job Type: Full-timeAbout the RoleWe’re looking for seasoned data engineers who can take end-to-end ownership of complex data engineering engagements—right from architecture and design to development and delivery. You’ll work with client stakeholders, lead data initiatives, and guide teams in building modern, scalable data platforms that support business intelligence, advanced analytics, and AI/ML applications.Key ResponsibilitiesLead the design and implementation of scalable, secure, and high-performance data architectures.
Architect and optimize data pipelines for batch and real-time processing across varied source systems.
Model and implement enterprise-grade data lakes, lakehouses, and warehouses to support analytical workloads.
Collaborate with client teams and internal stakeholders to define data strategies, tooling, and standards.
Set up and enforce best practices around code quality, pipeline observability, testing, and CI/CD.
Guide technical decision-making and mentor junior/mid-level engineers.
Troubleshoot complex data issues and provide architectural solutions across heterogeneous systems.
Must-Have SkillsStrong expertise in Python and advanced proficiency in SQL.
Deep understanding of data engineering architecture: ETL/ELT pipelines, data modeling (dimensional and normalized), warehousing, streaming systems, and data integration patterns.
Proven experience with distributed data processing frameworks (e.g., Apache Spark, Flink).
Proficient with orchestration tools like Airflow or similar, and production-grade data workflows.
Solid grounding in cloud-native data services (compute, storage, IAM, monitoring).
Experience leading technical implementations and collaborating with business stakeholders.Good to HaveMulti-cloud experience (AWS preferred; GCP or Azure familiarity is valued).
Hands-on experience with Databricks, dbt, and collaborative data development environments.
Awareness of data governance, security, lineage, and quality frameworks.
Certifications such as AWS Certified Data Analytics, Databricks Certified Data Engineer Professional, or equivalent.Other ExpectationsStrong leadership, ownership, and client communication skills.
Ability to balance hands-on contributions with technical guidance.
Adaptability to new tools, tech stacks, and problem domains.
Willingness to travel within India for short/medium-term client assignments.
Scan to Apply
Just scan this QR code to apply from your phone.
Job Location
Bengaluru, Karnātaka, 560023, India