PowerToFly
Recent searches
  • Events
  • Companies
  • Resources
  • Log in
    Don’t have an account? Sign up
Filters
Clear All
Advanced filters
Job type
  • Reset Show results
Date posted
  • Reset Show results
Experience level
  • Reset Show results
Company
  • Reset Show results
Skills
  • Reset Show results
Clear All
Cancel Show Results
Active filters:
Results 3437 Jobs

Wondering why you’re not getting hired?

Take our 3-min quiz and find out!

  • See what’s holding you back
  • Know exactly what to fix
  • Get a plan to move forward
Take the Quiz!
Loading...
Loading more jobs...

No more jobs to load

No more jobs to load

Principal AI Engineer
Save Job
Stryker

Principal AI Engineer

Onsite Bengaluru, India Full Time
Posted 7 hours ago
Save Job

Job Details

Vocera, now part of Stryker, is seeking a visionary and hands-on Principal Engineer – AI Test, Evaluation & Data Architecture

Watch this video to learn more about Stryker

to define and lead the enterprise-wide strategy for AI validation, model evaluation, and data governance across our speech and GenAI platforms.

This role serves as the AI Quality Architect for real-time speech systems, NLP pipelines, and LLM-powered applications deployed in mission-critical healthcare environments. You will establish scalable evaluation frameworks, design AI testing platforms, define data governance standards, and ensure production reliability of AI systems at scale.

This is a high-impact architectural leadership role requiring deep expertise in LLM validation, RAG evaluation, speech benchmarking, automation, MLOps, and AI lifecycle governance.


What You Will Do Enterprise AI Evaluation Architecture
  • Define and own the end-to-end AI evaluation architecture across speech, NLP, and GenAI platforms.

  • Establish standardized evaluation frameworks for:

    ASR systems (WER, latency, robustness, domain adaptation),

    NLP systems (intent accuracy, entity F1, confusion analysis),

    LLM systems (hallucination rate, groundedness, factual accuracy, consistency, safety)

  • Define measurable AI quality SLAs and release gating criteria.

  • Architect benchmarking standards across model versions, prompt changes, and retrieval updates.

  • Institutionalize regression evaluation pipelines for all AI releases.

LLM & RAG Reliability Strategy
  • Architect validation frameworks for:

    RAG-based systems,

    Prompt orchestration workflows,

    Multi-agent or multi-model AI pipelines

  • Define groundedness measurement strategies for enterprise RAG.

  • Establish adversarial testing, stress testing, and edge-case validation frameworks.

  • Implement hallucination detection standards and mitigation measurement.

  • Drive responsible AI practices, including bias detection and safety validation.

AI Testing Platform & Automation Architecture
  • Design and lead implementation of a scalable AI testing platform that includes:

    Offline evaluation pipelines,

    Golden dataset-driven regression systems,

    Synthetic data generation frameworks,

    Online A/B testing & shadow deployment strategies

  • Integrate AI validation workflows into CI/CD and MLOps pipelines.

  • Define drift detection and performance degradation monitoring strategies.

  • Establish real-time observability dashboards for AI quality metrics.

AI Data Governance & Lifecycle Management
  • Define enterprise-wide data governance strategy for AI systems, including:

    Data collection and curation standards,

    Annotation workflows and validation,

    Dataset versioning and reproducibility,

    Traceability across model iterations

  • Establish gold datasets for:

    Speech systems,

    NLP pipelines,

    Clinical and conversational workflows

  • Drive continuous learning loops between production telemetry and training data.

  • Ensure compliance with healthcare data privacy and regulatory standards.

Speech & Domain-Specific AI Validation
  • Define evaluation strategies for:

    Accent variability,

    Noisy clinical environments,

    Domain-specific vocabulary adaptation

  • Establish measurable latency and reliability benchmarks for real-time AI systems.

  • Lead failure mode analysis and systemic AI quality improvements.

Technical Leadership & Organizational Influence
  • Serve as the principal authority on AI testing and evaluation strategy.

  • Influence architecture decisions alongside Principal AI Architects and platform leaders.

  • Mentor senior engineers in AI validation, benchmarking, and data governance practices.

  • Drive AI quality maturity across multiple pods and engineering teams.

  • Partner with Product and Executive stakeholders to align AI quality metrics with business outcomes.

  • Shape long-term AI reliability roadmap for the organization.


Required Qualifications
  • Bachelor’s or Master’s degree in Computer Science, Engineering, AI, or related field.

  • 13+ years of experience in software engineering, AI engineering, or AI validation roles.

  • 5+ years of hands-on experience with LLM, RAG, NLP, or speech-based AI platforms.

  • Proven experience designing AI evaluation or testing frameworks at scale.

  • Strong expertise in:

    Hallucination detection,

    Golden dataset regression strategies,

    Adversarial and edge-case testing,

    Prompt validation and benchmarking

  • Strong proficiency in Python and data analysis for AI evaluation.

  • Experience building automated AI validation pipelines integrated with CI/CD.

  • Strong system design and distributed architecture understanding.

  • Experience leading cross-team technical initiatives.


Preferred / Strongly Desired Qualifications AI & GenAI
  • Experience in architecting evaluation frameworks for production RAG systems.

  • Familiarity with semantic search validation and retrieval benchmarking.

  • Experience designing LLM guardrails and structured output validation.

  • Knowledge of Responsible AI, fairness evaluation, and compliance auditing.

Speech & Voice Systems
  • Experience evaluating ASR/TTS systems in production environments.

  • Strong understanding of speech benchmarking metrics and domain adaptation strategies.

Cloud & Platform
  • Experience with Azure ML, Azure OpenAI, Azure AI Search.

  • Familiarity with MLOps and model lifecycle automation.

  • Experience designing scalable evaluation infrastructure in cloud-native environments.

Stryker is a global leader in medical technologies and, together with its customers, is driven to make healthcare better. The company offers innovative products and services in MedSurg, Neurotechnology, Orthopaedics and Spine that help improve patient and healthcare outcomes. Alongside its customers around the world, Stryker impacts more than 150 million patients annually.
Company Details
Stryker
 Kalamazoo, MI, United States
Work at Stryker

Stryker is a global leader in medical technologies and, together with our customers, we are driven to make healthcare better. We offer innovative... Read more

Did you submit an application for the Principal AI Engineer on the Stryker website?