Activity (50)
Article (348)
Ask Me Anything (55)
Chat (72)
Club Topic (28)
Course (38)
Discussion Panel (44)
Feature Spotlight (58)
Term (699)
Into The MoTaverse Episode (27)
Certification (91)
Collection (279)
Masterclass (92)
Moment (4389)
Newsletter Issue (243)
News (333)
Satellite (3219)
Solution (81)
Talk (251)
MoTaCon Session (873)
Testing Planet Session (105)
Insight (234)
TWiQ Episode (315)
An LLM is an AI system which is basically a massive, high-speed pattern-recognition engine trained on a mountain of text. It’s not "thinking" in the way you or I do. It tries to produce what it thinks the answer would look like based on everything it’s ever read and the instructions and context you give it. For us, it’s like having a pair programmer who has read every technical manual and Stack Overflow thread in existence, but sometimes forgets to check whether the advice is actually useful, practical, or just 'out there'.
Developers use LLMs primarily as a productivity multiplier. It’s brilliant at the "boilerplate" stuff that usually bores us to tears. For example, you can ask an LLM to "Write a Python function to sort a list of dictionaries by a specific key," and it’ll spit out a working version in seconds. But it can also "hallucinate" (make things up). It might suggest a library that doesn't exist or uses a deprecated method with security vulnerabilities. You still need to be the "Adult in the room" to review the code. Rahul wrote a great piece on this subject, "Human in the loop vs AI in the loop."
For a Quality Engineer, or tester of any kind, an LLM can be a powerful tool for generating test ideas and data, as long as you don't let it drive the bus. For example, you could feed a set of requirements into an LLM and ask, "What are ten edge cases for this login feature?" It might suggest things you hadn't considered, like handling emojis in usernames or SQL injection attempts. But if you use it to generate your automated tests, it might create "brittle" code that looks right but fails the moment your UI changes.
The biggest risk with LLMs, as with many things, is a loss of context. The model doesn't know your specific business logic, your security constraints, or your "unwritten" team rules, so be careful how you use it.
Use it to bounce ideas off, draft documentation, or create code snippets. It’s an assistant, not a replacement for the critical thinking and scepticism that a human brings to the party. Just because the LLM gave you an answer doesn't mean it is right, or that you can stop being a thought worker.
JIT Testing is the practice of testing only what has changed, when it changed, to keep the feedback loop as tight as possible. It means creating temporary disposable tests or selecting from an existing automation suite. More tools are using LLMs (large language models) to support this type of testing, and even security tools are using similar practices. It is important to note that this isn't really about replacing or doing less testing. It's a strategy that aims to reduce effort in some areas. Where continuous quality or testing practices are used, JIT testing aligns well.
Y2K simply means the 'year 2000' or 'year 2000 problem'. In the late 1990s, it was a massive wake-up call about the long-term cost of technical debt. To save on expensive memory, early developers stored years as just two digits (e.g., "99" instead of "1999"). It was a clever shortcut that worked, until time nearly ran out, and it didn't anymore.The Problem: When the year 2000 was reached, computers would see "00" and assume it was 1900. This threatened to send global banking, utilities, and transport systems into a proper meltdown. There were claims that 'planes would drop from the sky.' And 'Nuclear weapons would launch'. The Outcome: It wasn’t really a "hoax" as some people thought. The reason the world didn't end is that thousands of teams spent years finding, fixing, and testing every scrap of code they could find. When we finally reached the year 2000, there was little real disruption. A couple of examples were a video store that tried to charge tens of thousands for a '100-year-old' return. Some children were registered as being born in 1900, and the US Naval Observatory showed a date of 19100 on its website for a while. The Lesson: Quality isn't just about making things work today. It's about ensuring your "temporary" hacks don't become tomorrow’s disasters.
Harness engineering is an emerging AI methodology focused on creating reliable, structured environments ("harnesses") that enable AI agents to function securely and effectively in production. It involves designing feedback loops, constraints, and validation systems rather than just relying on model improvements.
Core Principles of Harness Engineering (as of 2026)
Agent Control & Reliability: Moving from "model-first" to "harness-first" by building scaffolding that allows agents to work on complex tasks for hours or days.
Mechanical Enforcement: Translating documentation into hard code constraints (guardrails) to ensure compliance, rather than relying on manual, human-driven review.
Context Engineering: Curating the knowledge base and designing the codebase for agent legibility, ensuring agents know what to do and how to do it.
Feedback Loops: Implementing automated systems that verify agent output, correct mistakes, and manage multi-agent workflows across repositories.
Key Components in a Harness
Grounding: Ensuring the agent knows its position, constraints, and the current state of the project.
Architecture & Design: Structured documentation (e.g., AGENT.md, PLANS.md) and directory structures designed for AI, not just humans.
Evaluation: Using tools to continuously test the agent’s work (e.g., using Playwright for browser automation).
Industry Applications
AI Agent Development: Used by organizations like OpenAI and Anthropic to make AI coding agents reliable.
CI/CD Optimization: Integrating agent workflows into CI/CD pipelines to manage deployment and security.
QA and Testing: Automating the software development lifecycle from building to deployment.
Community as a Service (CaaS) is the intentional design, facilitation, and stewardship of a community so that its outcomes: learning, connection, network building, trust, knowledge, insight, opportunities and change are reliably delivered as an ongoing part of the infrastructure and service.In a nutshell, it is about offering the support, services and product so that people can get the best out of their community to help them achieving their own personal credibility goals, with ease.
CaaS means:
Acknowledging that we have to work closely with people to help elevate them.
Too many people see community as an intangible, it's time we made it tangible and visible.
Being dedicated to research, because if we aren't actively researching, how can we know what our people need?
Research leads to community knowledge management, which helps us build things with the people
We are actively teaching, advising and supporting people how to become good community members, and how this is more than just showing up in a forum
Markdown is a lightweight markup language for formatting plain text documents. It uses simple symbols for converting easily to HTML for web display. Its syntax stays readable as plain text, even without rendering.Tech people rely on Markdown for README files, wikis and technical documentation.
Markdown files use a '.md' or '.markdown' file extension, for example, README.md or AGENTS.md.
An AGENTS.md is a Markdown file used in software projects to provide structured instructions for AI coding agents.Often seen as a companion to the human-focused README.md, offering machine-readable guidance on project setup, coding styles, testing commands, and contribution rules to help AI tools generate consistent, compliant code.The .md extension stands "Markdown", a lightweight markup language for creating formatted plain-text documents that render as HTML, commonly used for documentation.
Joyscrolling is the act of scrolling online for content that feels positive, uplifting and invigorating. It's the opposite of doomscrolling which involves reading negative, distressing and worrying news and posts. The MoTaverse platform is a good example of where someone can joyscroll as most content is encouraging, thoughtful and invigorating.
Slowly and indirectly building support and alignment.A way to approach to delivering value more reliably and sustainably look like https://www.ministryoftesting.com/moments/what-does-delivering-value-more-reliably-and-sustainably-look-like From Wikipedia:Nemawashi (根回し) is an informal Japanese business process of laying the foundation for some proposed change or project by talking to the people concerned and gathering support and feedback before a formal announcement.
A Schrödinbug is a bug that works purely by accident until someone examines or modifies the code. The software behaves correctly even though the logic is wrong, usually because of lucky conditions like memory being zeroed or timing working out just right.The moment someone refactors the code, adds logs, or changes compiler settings, the behavior breaks. The problem was always there, just hidden, nothing new was introduced. These bugs rely on undefined behavior, which makes them fragile and unpredictable.They are dangerous because the code looks stable and builds false confidence. Apparent correctness hides serious flaws that can surface after a compiler update, platform change, or even a small modification. If something works only because of luck, it’s already broken, it just hasn’t failed yet.
A hiccup error is a brief, self-resolving failure that appears once and then disappears. It usually happens because of momentary instability, a network glitch, a timing issue, a short resource spike, or a system running close to its limits. When the same action is retried, it succeeds, which makes the issue easy to ignore.For example, a payment API returns a 503 error and then works on the next attempt. The transaction completes, so no one investigates. Or a database query times out during a CPU spike, but the retry finishes in milliseconds. These errors leave little trace and rarely point to a clear code defect, they expose fragile interactions between systems.Hiccup errors are dangerous because teams treat them as flukes until they become frequent. A weekly hiccup turns daily, then hourly. By the time it's taken seriously, the system is already degraded. What looks like noise at first is often an early warning of a deeper reliability problem.
Bug convergence is the point in testing where the number of new bugs being found starts to drop and flatten out. At first glance, it looks like the product is becoming stable. But this slowdown doesn't always mean the system is high quality, it often means testers are hitting the same areas repeatedly, or the test approach has stopped uncovering new risks.For example, a team finds 50 bugs in week one, 20 in week two, 8 in week three. Leadership celebrates the progress, but testers have only been exercising the login and checkout flows. The entire admin panel, bulk operations, and error handling paths remain untested. The system still contains serious defects, just not in the places currently being tested.Bug convergence can happen because coverage is limited, test data is repetitive, or exploration has stalled. If bug reports cluster in the same few modules, or if testers struggle to think of new test scenarios, convergence is likely artificial, a sign that the test strategy has been exhausted, not the bugs.This is why convergence should be treated as a signal to change strategy, not a reason to stop testing. Shifting to different user personas, testing edge cases, varying test data, or exploring less-traveled system paths can reveal the defects that routine testing missed.
🚀 Accelerate releases with Test Impact Analysis.
With servers in >250 cities around the world, check your site for localization problems, broken GDPR banners, etc.
Jira Issue Connect brings live, real-time Jira data directly into TestRail, eliminating tab-switching and stale fields.
Catch the on-demand session with gaming legend John Romero and see how we’re redefining software quality at AI speed.