Ela MCB - QA Lead

Ela MCB

AI-First Quality Engineer

Building the Validation Layer for Trustworthy AI Systems

About Me

I am an AI-First Quality Engineer dedicated to solving the fundamental challenge of reliability in generative AI and LLM applications. My expertise lies in building validation frameworks that ensure AI systems are safe, accurate, and trustworthy before they reach production.

Core Expertise

LLM Testing AI Model Validation Prompt Engineering AI Safety Protocols Model Context Protocol Test Automation AI-Augmented Testing

I create systematic approaches to test the unpredictable nature of AI systems. My frameworks catch hallucinations, prevent prompt injections, and measure model drift before they impact users.

Technical Skills

Test Automation

  • Playwright with TypeScript
  • Selenium WebDriver
  • Cypress
  • JMeter for Performance Testing
  • Appium for Mobile Testing

AI/ML Technologies

  • AI-Augmented Test Generation
  • NLP for Log Analysis
  • LLM-Based Application Testing
  • Predictive Test Selection
  • Visual Testing with AI

Programming Languages

  • TypeScript/JavaScript
  • Python
  • SQL
  • Java
  • Bash/Shell Scripting

AI Projects

Learn AI-First Development

Master AI-Assisted Development: Comprehensive guides showing exactly how this portfolio was built

What You'll Learn:

  • Prompt Engineering: Real prompts for 10x faster development
  • Workflow Integration: Daily development routines with AI
  • Advanced Techniques: Chain-of-thought and role-based prompting
  • Real Examples: Every technique demonstrated in this portfolio

LLMGuardian Testing Framework

Production Framework: Systematic LLM validation, safety testing, and performance monitoring

Real Capabilities:

  • Tests 100+ prompt/response pairs across LLM providers
  • Measures accuracy, toxicity, and bias automatically
  • Detects hallucinations and factual errors
  • Prevents prompt injection attacks
  • CI/CD pipeline integration for continuous testing

Recent Results:

Math Accuracy: 94% | Safety Score: 87% | Overall: 91%

Code Generation

Use Case: Generate Playwright test automation code

Input:

"Test login functionality with valid credentials"

AI Output:

test('login with valid credentials', async ({ page }) => {
  await page.goto('/login');
  await page.fill('[data-testid="email"]', 'user@example.com');
  await page.fill('[data-testid="password"]', 'password123');
  await page.click('[data-testid="login-button"]');
  await expect(page).toHaveURL('/dashboard');
});

Defect Analysis

Use Case: Analyze test failures and suggest fixes

Input:

Test failure: Element not found after 5s timeout

AI Output:

  • Check if element selector is correct
  • Verify page load timing
  • Add explicit wait conditions
  • Consider dynamic content loading

QA to Prompt Engineer Journey

Career Transition Project: 4-week structured plan to transition from QA Lead to Prompt Engineer role

What's Included:

  • Week 1: Core prompting & mental models
  • Week 2: Evaluation, data & iteration
  • Week 3: Domain transfer & advanced patterns
  • Week 4: Portfolio, interview prep & outreach

Key Deliverables:

  • Prompt A/B CLI tool
  • Evaluation frameworks & metrics
  • Case studies with measurable results
  • Portfolio site & interview stories

LLM & Gen AI Research Discovery

Automated Discovery System: Weekly scans for new LLMs and groundbreaking Gen AI research papers

Sources Monitored:

  • GitHub: New LLM repositories
  • ArXiv: Latest research papers
  • Hugging Face: New model releases
  • Papers with Code: SOTA updates

Features:

  • Automated weekly scans
  • Smart filtering for significant releases
  • Accumulated discoveries for review
  • Centralized discovery dashboard

AI Research

LLM Testing Methodologies

Research Focus: Comprehensive analysis of testing approaches for Large Language Models, including hallucination detection, bias measurement, and safety validation frameworks.

Key Contributions:

  • Hallucination Detection: Consistency-based framework for identifying factual errors
  • Bias Analysis: Multi-dimensional approach to measuring unfair responses
  • Safety Validation: Comprehensive framework for harmful content detection
  • Testing Pipeline: Integrated solution for production LLM validation

MCP in Software Testing

Research Focus: Exploring Model Context Protocol applications in software testing, examining how standardized AI-tool communication can revolutionize test automation and create context-aware testing frameworks.

Key Innovations:

  • Context-Aware Testing: Real-time application state integration
  • Dynamic Test Generation: AI-driven test creation based on live data
  • Self-Healing Tests: Automatic adaptation to application changes
  • Intelligent Debugging: Complete failure context analysis

Agentic Testing Integration

Research Focus: Investigating autonomous AI agents for software testing, from existing platform integration to specialized testing agent development and multi-agent orchestration systems.

Key Innovations:

  • Multi-Agent Systems: Coordinated autonomous testing workflows
  • Specialized Agents: Explorer, Executor, Analyzer, and Orchestrator agents
  • Platform Integration: Leveraging AutoGPT, LangChain, and Semantic Kernel
  • Autonomous QA: Self-improving testing systems with minimal human oversight

Evaluating AI Models for Testing

Research Focus: Comprehensive framework for evaluating AI models in software testing contexts, including benchmarking methodologies, performance metrics, ROI analysis, and production deployment strategies.

Key Contributions:

  • Evaluation Framework: Systematic approach to model assessment across 6 key dimensions
  • Benchmark Suite: Test generation, bug detection, and adversarial testing scenarios
  • Model Comparisons: Side-by-side analysis of GPT-4, Claude 3.5, CodeLlama, and Gemini
  • ROI Calculator: Production metrics and cost-benefit analysis for deployment decisions

Why Use AI Agents for Testing?

Research Focus: Practical healthcare case study answering why QA professionals should use AI agentic flows for software testing, demonstrating autonomous agents for test generation, security scanning, and compliance validation.

Key Insights:

  • Healthcare EHR Example: Patient portal with HIPAA compliance requirements
  • 7 Agent Types: Explorer, Generator, Security, Compliance, Orchestrator agents
  • Proven Results: 92% coverage, 88% faster tests, 487% ROI
  • Practical Implementation: Tech stack, adoption roadmap, code examples

Multi-Agent Orchestration Framework

Research Focus: Academic research comparing Manager-Worker, Collaborative Swarm, and Sequential Pipeline architectures for AI testing systems, with empirical results from 50 trials demonstrating optimal task decomposition strategies.

Key Findings:

  • Manager-Worker Architecture: 80.2% defect detection, 31% cost reduction
  • Comparative Analysis: 4 architectures across 5 specialized agent roles
  • ATAO Framework: Context-aware architecture selection system
  • Statistical Validation: ANOVA, Tukey HSD, effect size analysis

Featured Projects

QA-to-AI Transformation Roadmap Premium

A proven 6-12 month strategy for transitioning traditional QA teams to AI-augmented quality engineering. Transform your team into AI-first leaders.

Leadership Strategy Change Management ROI Modeling

Key Results:

  • 487% ROI (Healthcare case study)
  • 40-70% efficiency gains
  • 85%+ automation coverage
  • 32-week phased implementation

AI Test Generator

A tool that uses AI to automatically generate test cases based on application behavior and user stories.

Python OpenAI API Playwright

Playwright Framework

Production-ready test automation for elamcb.github.io with AI-powered testing via MCP. Validates critical functionality, navigation, and performance with 100% test success rate.

Playwright ES Modules MCP CI/CD

Job Search Automation Suite

An ethical automation system that demonstrates intelligent job matching, application tracking, and interview preparation using AI and test automation principles.

Python Playwright AI/ML Data Analytics

Automation Features:

  • āœ… Intelligent job matching (85% accuracy)
  • āœ… Application status tracking
  • āœ… Interview analytics dashboard
  • āœ… Resume optimization suggestions

AI IDE Collection - Gotta Code 'Em All

Interactive comparison of 10 AI-powered development environments tested over 100+ hours. S-Tier through B-Tier rankings with detailed pros, cons, and real-world performance insights.

Developer Tools AI Assistants Comparative Analysis Interactive UI

IDEs Tested:

  • S-Tier: Cursor (My Favorite), Windsurf
  • A-Tier: GitHub Copilot, Zed, Void, Continue.dev, Trae
  • B-Tier: Replit AI, CodeWhisperer, Tabnine
  • 11 IDEs tested over 100+ hours

Legacy-AI Bridge Framework

Practical solution for introducing AI capabilities into legacy enterprise systems without disruption. Addresses the #1 barrier to AI adoption in established companies.

Enterprise Integration Legacy Systems AI/ML Pipeline Risk Management

Real-World Results:

  • āœ… Banking System: 40% faster processing, 60% fraud reduction
  • āœ… Manufacturing ERP: 25% less downtime, 30% inventory optimization
  • āœ… Healthcare Records: Improved outcomes, maintained HIPAA compliance
  • āœ… Zero Downtime: Non-invasive integration approach

Algorithmic Trading System

A systematic mean reversion trading strategy with automated backtesting, risk management, and performance analytics. Demonstrates quantitative analysis and systematic decision-making.

Python pandas Statistical Analysis Risk Management

Trading Performance:

  • āœ… +127% total return (2020-2024)
  • āœ… 1.67 Sharpe ratio (risk-adjusted)
  • āœ… 64% win rate across 342 trades
  • āœ… -12.4% maximum drawdown

Bio-AI Analogies: Educational Content Generator

An innovative educational tool that explains complex AI concepts through biological analogies. Demonstrates interdisciplinary thinking connecting AI with biological systems, making advanced concepts accessible through familiar natural processes.

Python Educational AI BDH Model Content Generation

Key Features:

  • āœ… Explains neural networks through synaptic plasticity
  • āœ… Makes attention mechanisms accessible via selective vision
  • āœ… Compares reinforcement learning to animal training
  • āœ… Bridges AI and biological systems thinking

AI Innovations Discovery for QA Testing

Weekly automated discovery system that scans GitHub, Hacker News, and research sources to find cutting-edge AI tools, frameworks, and methodologies applicable to quality assurance and testing. Stay ahead of the latest innovations in AI testing.

Automated Discovery GitHub Actions Weekly Updates QA Innovation

Discovery Features:

  • āœ… Automatic weekly scans (every Monday)
  • āœ… 15+ targeted search keywords for QA/testing
  • āœ… Categorization by innovation type (test generation, execution, maintenance, etc.)
  • āœ… Filters duplicates and existing tools
  • āœ… Generates detailed reports with QA applications

Work In Progress WIP

Resources and projects currently in development. Includes ETL testing templates, AI innovations for data QA, and experimental tools.

ETL Testing Data QA AI Tools Templates

Current Contents:

  • āœ… ETL Test Plan Template
  • āœ… SQL & Python Test Cases
  • āœ… AI Innovations for ETL Testing
  • āœ… E2E Testing for DevOps & LLMOps
  • āœ… Data Quality Frameworks

Portfolio Impact

0

Unique Visitors

0

Total Views

10x

Development Speed

3

AI Systems Used

Interactive Zone

Test your skills and have some fun with AI-powered challenges!

AI vs Human: Guess Who Wrote This Code

Can you tell the difference between AI-generated and human-written code? Test your skills!

Score: 0/0 Accuracy: 0%

Challenge yourself: AI-generated code often has certain patterns, while human code shows personal style and creative problem-solving approaches.

Get In Touch

I'm always interested in discussing AI-augmented testing, test automation frameworks, or potential collaboration opportunities.

Contact Me