name: mflux-testing description: Run tests in mflux (fast/slow/full), preserve image outputs, and handle golden image diffs safely.
mflux testing
This repo uses pytest with image-producing tests. Always preserve outputs for inspection and never update reference images unless explicitly asked.
When to Use
- You need to run tests (fast/slow/full) or debug failing tests.
- There are image/golden mismatches and you need to report paths/output for review.
Instructions
- Prefer the Makefile test targets:
make test-fast(fast tests, no image generation)make test-slow(slow tests, image generation)make test(full suite)
- Always keep
MFLUX_PRESERVE_TEST_OUTPUT=1on test runs (already built into the Makefile test targets). - If a change affects defaults, config resolution, metadata fields, or CLI behavior, add or update tests that cover the changed behavior directly instead of relying only on manual verification.
- If tests fail:
- Summarize the failing test names and the key assertion output.
- Point to any generated images/artifacts on disk for manual review.
- Do not regenerate/replace reference (“golden”) images unless the user explicitly requests it.
Manual validation (config resolution + local model paths)
Use when a change touches model config resolution, mflux-save, or the model’s generate CLI, or when a PR fixes local model-path handling for the model under investigation. Refer to the mflux-cli skill to find the correct generate command for the model you are testing.
- Run a local-path quantize/save:
- Use the
mflux-cliskill to look up the correct command and flags. - Verify CLI usage with the command’s
--helpbefore running it. - Save to a known location (e.g., Desktop) to make follow-up steps explicit.
- Use the
- Run generation from the saved model using the correct model-specific generate CLI:
- Use the
mflux-cliskill to find the generate command and required flags. - Verify CLI usage with the command’s
--helpbefore running it.
- Use the
- If the model has multiple size variants, repeat the above for each variant to confirm the correct overrides are applied.
- Do not commit output artifacts; delete or leave them untracked.
TDD Red-Green-Refactor
Testing
Skill qui guide Claude a travers le cycle TDD complet.
Audit d'Accessibilité Web
Testing
Réalise un audit d'accessibilité web complet selon les normes WCAG.
Générateur de Tests UAT
Testing
Génère des cas de test d'acceptation utilisateur structurés et complets.