We are seeking a Staff Data Engineer who will design and build data integration, storage systems, and data infrastructure that enable analytics and AI-powered applications. In this role, you will navigate complex enterprise data landscapes — building relationships to gain data access, handling undocumented schemas through investigation, and building robust, maintainable integration solutions that serve multiple teams.
What You’ll Do Responsibilities:
* Business: Immerse in operations until you think like an insider. Rapidly acquire domain expertise through direct observation, translate between business and engineering seamlessly, and mentor engineers in your area on immersion. Influence senior stakeholders effectively, manage complex stakeholder landscapes with competing agendas, and build trust rapidly with new stakeholders.
* Delivery: Lead rapid delivery initiatives across teams in your area, coach on prototype-first approaches for data solutions, and establish trust through consistent fast delivery. Navigate complex enterprise data landscapes, build relationships to gain data access, handle undocumented schemas through investigation, and define clear criteria for prototype-to-production transitions.
* People: Build high-performing teams across your area, navigate complex interpersonal dynamics, foster psychological safety, and create environments where diverse perspectives are valued. Influence through communication at all levels — from frontline to executive. Handle difficult conversations skillfully and train engineers in your area on effective communication.
* AI-Augmented Development: Optimise AI tool usage across teams in your area, train engineers on AI-augmented workflows, evaluate new AI development tools, and establish practices that balance AI speed with verification rigour.
* Scale: Design complex data architectures spanning multiple systems across teams. Make strategic trade-offs between consistency, performance, and maintainability. Establish and enforce quality standards across teams in your area, mentor engineers on effective code review, and drive testing strategies for data pipelines.
* Documentation: Define documentation standards across teams in your area, create documentation systems and templates for data pipelines, train engineers on spec-driven development, and ensure data lineage documentation quality across projects. Lead pattern generalization initiatives, defining criteria for when to generalize versus keep custom.
* Reliability: Define reliability standards across teams in your area, drive post-incident improvements systematically for data pipeline failures, design capacity planning processes for data infrastructure, and mentor engineers on SRE practices applied to data systems. Process: Lead lean transformations across teams in your area, design flow-optimised processes for data delivery, coach engineers on lean principles, and balance speed with sustainability. Establish metrics that drive improvement.Behaviour
* Own the Outcome: Drive accountability culture focused on outcomes not deliverables. Own business relationships and impact metrics across your function. Make trade-offs between custom data solutions and generalisable work. There is no \"I must run this by X.\" Ensure verification rigour for AI-generated code — particularly critical for data transformations.
* Be Polymath Oriented: Champion cross-disciplinary learning. Create holistic data solutions spanning technical and business domains. Embody the Renaissance Engineer ideal. Translate specialised data engineering knowledge into accessible explanations. Think like a business insider.
* Communicate with Precision: Create spec-driven development practices for data pipelines. Mentor others on precise communication around data definitions, quality expectations, and pipeline behaviour. Span C-level executives to frontline workers. Drive clarity as a core value across your function. Represent the organisation externally.
* Don't Lose Your Curiosity: Drive team curiosity through challenging questions about data quality and sources. Create environments where exploration and experimentation with data are encouraged. Model problem discovery orientation. Seek out ambiguity in data landscapes rather than avoiding it.
* Think in Systems: Shape systems design practices across your function for data architectures. Conduct chaos engineering experiments on data pipelines. Influence cross- team architecture decisions. Create clarity from complexity. Bridge technical data systems with business processes.
* Data Integration: You navigate complex enterprise data landscapes across teams, build relationships to gain data access, handle undocumented schemas through investigation, and build robust, maintainable integration solutions. You mentor engineers in your area on data integration challenges.
* Data Modeling: You design complex data architectures spanning multiple systems across teams. You make strategic trade-offs between consistency, performance, and maintainability. You mentor engineers in your area on data modelling best practices.
* Architecture & Design: You design complex multi-component systems end-to-end, evaluate architectural options for large initiatives across teams, guide technical decisions for your area, and mentor engineers on architecture. You balance elegance with delivery needs. Working-level Skills
* Code Quality & Review: You produce consistently high-quality, well-tested data pipeline code. You review AI-generated code critically and never ship code you don't fully understand. You identify edge cases and ensure adequate test coverage.
* Full-Stack Development: You build complete applications rapidly across any technology stack for teams in your area. You select the right tools for each problem, balance technical debt with delivery speed, and mentor engineers on full-stack development. DevOps & CI/CD: You build complete CI/CD pipelines end-to-end, manage infrastructure as code, implement monitoring, and design deployment strategies for your data services.
* Cloud Platforms: You design cloud-native data solutions, manage infrastructure as code, implement security best practices, and make informed service selections. You troubleshoot cloud-specific issues.
* Site Reliability Engineering: You design observability strategies for your data services, lead incident response, implement resilience testing, and conduct blameless post-mortems. You balance reliability investment with feature velocity.
* AI-Augmented Development: You integrate AI tools strategically into your development workflow. You review AI-generated code with the same rigour as human code and never ship code you don't fully understand.
* AI Evaluation & Observability: You design evaluation frameworks with custom evaluators tailored to your use case. You build golden datasets, establish annotation workflows with clear rubrics, and run experiments to compare prompt and model changes systematically.
* Problem Discovery: You seek out undefined data problems rather than avoiding them. You embed with users to discover latent data needs, coach engineers in your area on problem discovery techniques, and turn ambiguity into clear problem statements.
* Rapid Prototyping & Validation: You lead rapid delivery initiatives across teams in your area, coach on prototype-first approaches, establish trust through consistent fast delivery, and define clear criteria for prototype-to-production transitions.
* Business Immersion: You immerse in operations until you think like an insider. You rapidly acquire domain expertise through direct observation, translate between business and engineering seamlessly, and mentor engineers in your area on immersion.
* Stakeholder Management: You manage multiple stakeholders with different interests, navigate conflicting priorities diplomatically, and build trust through consistent delivery. You tailor communication to each audience.
* Multi-Audience Communication: You influence through communication at all levels — from frontline to executive. You handle difficult conversations skillfully, train engineers in your area on effective communication, and represent teams across the function.
Foundational-level Skills
* AI literacy, technical writing, collaboration, data analysis & knowledge management. Evaluates AI risks, documents systems, enables teams, analyzes data, and maintains searchable knowledge.
What You Bring
* Bachelor’s degree in computer science, Data Engineering, Software Engineering, or related field with 7+ years of relevant professional experience.
* Deep production experience with Python, SQL, and cloud data platforms (AWS preferred —including services such as Glue, S3, Redshift, Lambda, and EMR; Snowflake, Azure, or GCP also valued) is required.
* Proven expertise in large-scale data integration, distributed data processing (such as Spark), and enterprise data architecture is essential.
* Experience with enterprise data platforms (such as Salesforce Data Cloud or similar CRM data systems) is strongly valued.
* Demonstrable fluency with AI coding tools (such as Claude Code, Cursor, GitHub Copilot, or similar) and hands-on experience architecting data infrastructure that supports generative AI applications (data preparation for LLMs, vector databases, RAG systems, evaluation pipelines) is essential.
* Experience leading data initiatives across multiple teams, mentoring engineers, and navigating complex enterprise data landscapes with undocumented or messy data sources is required.
* Experience in an embedded, forward-deployed, or consulting-style engineering model is a strong plus.
Due to LinkedIn limitations, we are not able to post the complete job description. Detailed job description will be provided once we move ahead with your candidature.
Ready to Apply?
Let’s build the future of health together. Apply below or reach out to:
#J-18808-Ljbffr