Open to Opportunities

Himanshu
Kala

AI Quality Assurance Engineer  ·  LLM Evaluation  ·  AI Safety

Specializing in GPT-style LLM evaluation, hallucination detection, and AI safety validation. Ensuring reliable, safe, and high-performance AI products through rigorous testing.

// SYSTEM.STATUS
2+
Years Experience
🎯
300+
AI Responses / Week
🛡️
100%
Safety Focused

$ status: active

$ location: Meerut, India

Scroll

// Who I Am

About Me

I'm an AI-focused Quality Assurance Engineer with over 2 years of experience in manual and automation testing, with deep specialization in GPT-style LLM evaluation and AI safety validation.

My work sits at the intersection of AI quality and safety — I help ensure that conversational AI systems behave accurately, ethically, and reliably before they reach real users.

From hallucination detection to adversarial prompt design, I test the edges of AI systems so they're ready for the real world. I thrive in Agile environments and collaborate closely with cross-functional teams.

Email
LocationMeerut, India
🤖

LLM Evaluation Specialist

Validating 300+ model responses weekly for accuracy, relevance and logical consistency.

🛡️

AI Safety Engineer

Systematic detection of harmful, biased, and policy-violating AI outputs.

Adversarial Prompt Designer

Structured edge-case prompting to stress-test conversational AI robustness.

🔗

API and Automation Tester

REST API validation using Postman across authentication, payloads, and error handling.

// What I Know

Skills & Expertise

AI & LLM Testing
GPT-style LLM Evaluation Hallucination Detection AI Safety Evaluation Harmful Content Detection LLM Validation Conversational AI Testing Prompt Testing Manual Response Validation
Testing Methodologies
Manual Testing End-to-End Testing Smoke & Sanity Testing Regression Testing Test Case Design Defect Lifecycle Agile / Scrum
Tools & Platforms
Postman JIRA LogRocket Relay Retool Audacity Slack SQL

// Where I've Worked

Work Experience

Oct 2023 – Present
Simublade Technology (via Tavus.io)
AI Products · AI Video Personalization
Associate QA Engineer
AI & LLM Testing
  • Review and score 300+ AI model responses every week to check if they are accurate, relevant, and make logical sense.
  • Spot hallucinations — cases where the AI makes up facts or gives misleading answers — and flag them before they reach users.
  • Test AI responses for harmful, biased, or policy-breaking content to keep the product safe.
  • Write tricky and edge-case prompts to find weak spots in how the AI handles unusual or difficult inputs.
Video & Avatar QA
  • Test AI-generated videos to make sure lip movements match the voice, faces look natural, and the output is free of visual glitches.
  • Re-test conversational AI and digital avatar features after every update to make sure nothing is broken.
  • Check that the platform stays fast and stable after new AI model releases.
Voice Training & API Testing
  • Use Audacity to record, clean, and prepare voice audio samples used for training AI speech models.
  • Check audio quality — remove background noise, silences, and distortions so the training data is clean and usable.
  • Test REST APIs using Postman to verify that requests, responses, authentication, and error messages all work correctly.
General
  • Work closely with developers, designers, and product managers in an Agile team to ship AI features on time.

// Academic Background

Education

Bachelor of Computer Applications (BCA)
Chaudhary Charan Singh University
Computer Applications (General)
2022
12th Standard
Uttar Pradesh Board
English Medium
2019
10th Standard
CBSE Board
English Medium
2016
✅ Certified
Java Expert (JAVAEXPERT)
DUCAT India
Cert No: 31122022884189904 · Student ID: 31236/2022
Aug 2022 – Mar 2023

// What I've Built

Projects

Featured Project
AI Video Automation — Tavus AI Integration

Quality assurance for an AI-powered video platform that turns simple inputs into fully personalized, AI-generated videos.

  • Checked that the AI correctly turned user inputs into personalized videos without errors or quality issues.
  • Reviewed Phoenix AI model results to make sure faces looked natural and voices matched lip movements properly.
  • Re-tested AI chat and digital avatar features after each update to confirm nothing was broken.
  • Monitored the system after new releases to verify it stayed stable and performed well under real conditions.
Conversational AI
Digital Human Interaction Testing

Comprehensive validation of digital human and conversational AI interaction systems for production reliability.

  • Performed end-to-end testing of conversational AI workflows.
  • Validated digital avatar lip-sync and voice accuracy pipelines.
  • Ensured high response accuracy standards before production deployment.

// What I've Accomplished

Achievements

Efficiency
Improved AI Response Validation

Improved AI response validation efficiency by implementing a structured prompt testing strategy across teams.

AI Safety
Strengthened Safety Compliance

Strengthened AI safety compliance through systematic harmful-content detection workflows.

Recognition
Recognized for Debugging Skills

Recognized for analytical debugging skills and proactive communication during AI product releases.

Quality
100% Pre-Deployment Accuracy

Consistently ensured high model response accuracy and safety before every production deployment.

// Get in Touch

Contact Me

Let's connect

I'm open to new opportunities, collaborations, and conversations about AI quality and safety. Feel free to reach out any time.

✉️
📍Meerut, Uttar Pradesh, India
✅ Message sent! I'll get back to you soon.