LLMs, MCP, agents & intelligent systems—from prototype to production
I am an AI-first AI engineer: I design and ship systems around large language models, agents, and modern tooling—with the rigor to make them reliable in the real world. That means evaluation, safety, and automation built in from day one, not bolted on at the end.
I build systematic ways to stress-test and observe AI behavior in the wild: catching hallucinations, hardening against prompt injection, and tracking model drift before users feel the pain.
Master AI-Assisted Development: Comprehensive guides showing exactly how this portfolio was built
Production Framework: Systematic LLM validation, safety testing, and performance monitoring
Math Accuracy: 94% | Safety Score: 87% | Overall: 91%
Use Case: Generate Playwright test automation code
"Test login functionality with valid credentials"
test('login with valid credentials', async ({ page }) => {
await page.goto('/login');
await page.fill('[data-testid="email"]', 'user@example.com');
await page.fill('[data-testid="password"]', 'password123');
await page.click('[data-testid="login-button"]');
await expect(page).toHaveURL('/dashboard');
});
Use Case: Analyze test failures and suggest fixes
Test failure: Element not found after 5s timeout
Career Transition Project: 4-week structured plan to move into prompt engineering and AI-first development
Automated Discovery System: Weekly scans for new LLMs and groundbreaking Gen AI research papers
Research Focus: A hybrid architecture combining Parallel-Agent Reinforcement Learning (PARL) with Input Domain-Aware Mixture of Experts (IDA-MoE) for dynamic AI testing. Bridges agent swarms and MoE for intelligent code quality assurance.
Research Focus: Comprehensive analysis of testing approaches for Large Language Models, including hallucination detection, bias measurement, and safety validation frameworks.
Research Focus: Exploring Model Context Protocol applications in software testing, examining how standardized AI-tool communication can revolutionize test automation and create context-aware testing frameworks.
Research Focus: Investigating autonomous AI agents for software testing, from existing platform integration to specialized testing agent development and multi-agent orchestration systems.
Research Focus: Comprehensive framework for evaluating AI models in software testing contexts, including benchmarking methodologies, performance metrics, ROI analysis, and production deployment strategies.
Research Focus: Practical healthcare case study answering why QA professionals should use AI agentic flows for software testing, demonstrating autonomous agents for test generation, security scanning, and compliance validation.
Research Focus: Academic research comparing Manager-Worker, Collaborative Swarm, and Sequential Pipeline architectures for AI testing systems, with empirical results from 50 trials demonstrating optimal task decomposition strategies.
Research Focus: Analysis of major AI breakthroughs in October-December 2025 and their implications for software testing, AI systems, and autonomous agents.
Research Focus: Living overview of AI testing trends and research discoveries for builders of AI systems. Updated monthly by the Research & Literary Agent—no manual publish needed.
Research Focus: Practical LLM safety and red-teaming for QA leaders—a first sprint teams can run before production: eight adversarial families, forty prompts, spreadsheet template, Pass/Conditional/Fail rubric, and numeric release gate.
Dedicated hub for QA leaders focused on LLM safety and red-teaming—parallel to the main AI research index, with its own cadence and article list.
Sister hub to the LLM Safety series: short articles on governance, principles, and operational ethics for the Ethical AI Frameworks LinkedIn group.
Research Focus: A human-centered approach to understanding and testing AI systems, exploring hallucinations, bias, and safety considerations through gentle observation and compassionate inquiry.
Research Focus: A research paper exploring model drift in machine learning systems: what it is, why it happens, how to detect it, and how to fix it. Written for the curious, not just the experts.
A proven 6-12 month strategy for transitioning traditional QA teams to AI-augmented ways of working. Transform your team into AI-first leaders.
Production-ready test automation for elamcb.github.io with AI-powered testing via MCP. Validates critical functionality, navigation, and performance with 100% test success rate.
An ethical automation system that demonstrates intelligent job matching, application tracking, and interview preparation using AI and test automation principles.
Interactive comparison of 10 AI-powered development environments tested over 100+ hours. S-Tier through B-Tier rankings with detailed pros, cons, and real-world performance insights.
Practical solution for introducing AI capabilities into legacy enterprise systems without disruption. Addresses the #1 barrier to AI adoption in established companies.
A systematic mean reversion trading strategy with automated backtesting, risk management, and performance analytics. Demonstrates quantitative analysis and systematic decision-making.
An innovative educational tool that explains complex AI concepts through biological analogies. Demonstrates interdisciplinary thinking connecting AI with biological systems, making advanced concepts accessible through familiar natural processes.
Weekly automated discovery system that scans GitHub, Hacker News, and research sources to find cutting-edge AI tools, frameworks, and methodologies applicable to quality assurance and testing. Stay ahead of the latest innovations in AI testing.
Resources and projects currently in development. Includes ETL testing templates, AI innovations for data QA, and experimental tools.
Modular unified agent system with multiple capabilities working 24/7 to maintain, monitor, and enhance this portfolio. Single workflow, shared utilities, easy to extend.
Unique Visitors
Total Views
Development Speed
AI Systems Used
Test your skills and have some fun with AI-powered challenges!
Can you tell the difference between AI-generated and human-written code? Test your skills!
Challenge yourself: AI-generated code often has certain patterns, while human code shows personal style and creative problem-solving approaches.
I'm always interested in discussing AI-augmented testing, test automation frameworks, or potential collaboration opportunities.
Contact Me