Non-finetuned release record

Qwen3.6 AEON RYS 15/20: what we tested, what won, and what did not.

This is the long-form audit trail for the base, non-finetuned AEON RYS 15/20 GGUF release. It records the source model choice, layer-window selection, quantization evidence, runtime work, rejected variants, implementation notes, and the practical tests that shaped the final default file.

15,20RYS window. Source layers 15..19 are duplicated into output layers 20..24.
IQ4_NLMain deployment artifact. BF16 exists for exploration, not normal inference.
-0.75%Mixed four-probe mean change from RYS BF16 to released IQ4_NL.
custom forkRequired runtime path; faster in the internal comparison, but quality preservation is the main claim.

Short Read

The public claim is narrow: this exact AEON RYS 15/20 branch, served through the custom ik-llama fork, produced a practical Q4-class release whose mixed probe score stayed close to its BF16 source. The fork also mattered for speed in our internal runtime comparison, but the primary claim is quality preservation under compression, not a speed benchmark. The broader aim is practical local work: one capable worker on a 24 GB-class GPU, or more parallel workers when more GPUs are available.

What This Page Claims

This page is intentionally written as a record, not a leaderboard page. It explains why the release default became the non-MTP IQ4_NL GGUF and where the evidence is strong, weak, or mixed.

Project Positioning

The broader goal of this model series is practical local work: capable, functional models that fit realistic hardware and can simply get tasks done. Built on the AEON uncensored base, this release is meant to stay low-friction: less lecturing, less getting in the way, and more focus on completing the task in front of it.

RYS and IQ4_NL are the practical part of that goal: preserving useful reasoning and coding behavior in a compact form factor. One 24 GB-class GPU can be enough for a serious local worker. More GPUs naturally mean more independent workers in parallel, not a different claim about one model instance.

Claimed

On the mixed BF16-vs-IQ4_NL snapshot, this RYS 15/20 branch lost less than 1% relative mean score after compression to the released practical GGUF.

Claimed

The custom ik-llama fork is the intended runtime path. It also measured faster in the internal comparison, 39.37 tok/s versus 22.51 tok/s for the patched upstream-style comparison, but that speed note is secondary to preserving quality in the Q4_NL file.

Not Claimed

15/20 is not presented as universally best. A later 11/14 long-reasoning comparison was cleaner on repetition, and that remains an important caveat.

Explanatory route map from AEON base to RYS 15/20, strict scan, IQ4_NL quantization, and release default.
The release story is a path, not a single score. The practical claim depends on the RYS build, strict selection, quantization survival, and tested runtime path together.
Four caveat cards distinguishing strong evidence, the 11/14 repetition caveat, experimental MTP, and implementation footnotes.
The page separates strong claims, real caveats, experimental branches, and known tradeoffs so readers do not need to reconstruct that hierarchy from the tables.

Released Artifacts

The public Hugging Face repo contains the practical inference file, an experimental MTP file, a BF16 GGUF reference, and the BF16 safetensors folder for continued work.

FilePurposeSizeDecision
Qwen3.6-27B-AEON-RYS-MaxThinkCoder-IQ4_NL-ik-llama-custom-mixed.gguf Main non-finetuned inference artifact. 16,554,834,080 bytes Default release file.
Qwen3.6-27B-AEON-RYS-MaxThinkCoder-SpeedBoosted-IQ4_NL-MTP-Experimental.gguf MTP-capable IQ4_NL artifact with MTP-tail imatrix coverage. 16,794,473,728 bytes Experimental; not the default.
Qwen3.6-27B-AEON-RYS-MaxThinkCoder-BF16.gguf Source-quality GGUF reference for inspection, conversion, and comparison. 57,597,296,608 bytes Exploration artifact.
bf16-safetensors/ HF-format checkpoint for Transformers, LoRA, SFT, continued training, or conversion work. 11 shards Training/workflow artifact.

Decision

For normal users, the intended file is the non-MTP IQ4_NL GGUF. The BF16 files exist so people can inspect or continue the work; they are not the small-form deployment claim.

Experiment Timeline

The release moved through source selection, RYS construction, strict scanning, quantization screening, runtime work, and practical validation. Later fine-tune work built on this base, but is separate from the non-finetuned record here.

1. Source Branch

Use AEON-7/Qwen3.6-27B-AEON-Ultimate-Uncensored as the source branch instead of the earlier official-base line or other abliterated candidates.

2. RYS Build

Build safetensors-first RYS checkpoints with correct tensor remapping and per-layer metadata remapping for Qwen3.6 hybrid attention.

3. Strict Window Scan

Compare candidate RYS windows against AEON baseline on strict math and EQ validation files. Select 15/20 as the balanced winner.

4. Quantization Screen

Convert to GGUF, build imatrix calibration, quantize to IQ4_NL, and compare BF16 against the quantized release candidate.

5. Runtime Work

Patch and package the ik-llama fork for the custom mixed GGUF layout, Qwen3.6 hybrid handling, graph split, Jinja, and DeepSeek reasoning format.

6. Release Decision

Publish the non-MTP IQ4_NL as the default, keep MTP as experimental, keep BF16 as exploration/reference, and document known caveats.

Testing ladder showing source selection, mechanical RYS build, short scan, strict scan, quantization survival, runtime path, MTP research, and practical agent checks.
The ladder view shows the full release process without asking readers to parse every table first.

Test Coverage Snapshot

This is the compact count of the evidence behind the release. The public values used for the argument are reproduced on this page; the evidence ledger gives provenance filenames.

StageCoverageWhat It AnsweredDecision Impact
Early AEON short scan13 candidate mappings across math_16 and eq_16.Which windows looked promising before the strict pass?Exploratory only; not final selection basis.
AEON strict scan6 candidate mappings across math_120 and historical eq_140 files.Which RYS window had the best balanced strict score?Selected 15,20.
15/20 BF16 vs IQ4_NLFour mixed probes: math_16, eq_16, math_4, gsm8k_5; 41 prompt items total.Did the release quant survive compression?Confirmed IQ4_NL as the practical target.
11/14 quant caveatsReasoning slice plus no-think math/EQ quant comparison files.Was 11/14 obviously better after quantization?No; it was promising but volatile and not strict-balanced winner.
Long fair comparison2 IQ4_NL branches, 21 numeric questions, 2048-token reasoning budget.Did 15/20 have repetition weaknesses against 11/14?Yes; documented as a real caveat.
Mixed-Q8 probe2 quant variants, 9-item quick paired eval.Would protecting the RYS window in Q8 improve the default?No replacement.
MTP workMTP-tail imatrix, 512/2048 quality checks, short and long speed matrices.Should the MTP file become the default?No; keep experimental.
Practical agent checksBase row from the 5-task production matrix, plus the AEON RYS attempt/retry rows from a later 5-run canvas comparison.Could the compressed model act as a coding-agent base?Yes with caveats. SignalLatch and Unsloth comparison rows are downstream context, not part of this base-release claim.

Spreadsheet Source

The chart data is collected into an OnlyOffice workbook: qwen36_aeon_rys_stats.xlsx. It contains the short scan, strict scan, quantization snapshot, 11/14 long-reasoning comparison, MTP speed notes, and duplicate-window cost sheet.

Complete Candidate Lists

The first version of this page summarized the scan coverage and listed the decision rows, but did not spell out every tested mapping. This section is the explicit candidate appendix from the AEON scan files.

Single-Layer Note

The early short scan contains one true single-layer duplication candidate: blocks:20,21, which duplicates source layer 20 only. The file named aeon_single_blocks_15_20_math120.pkl is misleadingly named for this question: it contains one entry for the 15,20 block candidate, not a full single-layer sweep.

Early short-scan candidatemath_16eq_16MeanRead
blocks:24,320.8211090.7140060.767558Best short-scan mean; not final strict winner.
blocks:15,200.8185870.7123400.765463Strong short-scan candidate; later strict-balanced winner.
blocks:11,140.8029400.7156090.759275Strong math/reasoning branch, later caveated.
blocks:31,340.7907850.7133170.752051Close strict-scan runner-up later.
blocks:30,350.8000020.7015060.750754Official-base winner did not transfer as AEON winner.
blocks:0,00.7943870.7065060.750447AEON baseline.
blocks:11,14;30,350.7710050.7090060.740006Two-window mesh; did not beat simpler candidates.
blocks:15,270.7569660.7144230.735695Wider mid-window candidate.
blocks:28,360.7667330.7005770.733655Late-window candidate.
blocks:9,170.7308460.7235260.727186Good EQ, weaker math.
blocks:30,340.7133820.7165060.714944Late-window candidate.
blocks:8,170.6640310.7247760.694403High EQ but poor math balance.
blocks:20,210.6803630.7035260.691944Only true single-layer duplication candidate in this short scan.
Strict-scan candidatemath_120eq_140MeanDecision role
blocks:15,200.9714410.6473470.809394Winner.
blocks:31,340.9772690.6407500.809010Very close runner-up.
blocks:11,140.9814980.6349290.808214Best strict math, not combined winner.
blocks:24,320.9718230.6285590.800191Near baseline.
blocks:0,00.9701810.6297100.799945AEON baseline.
blocks:30,350.9648410.6275210.796181Official-base winner, not AEON winner.
Bar chart of AEON strict scan mean scores for 15/20, 31/34, 11/14, 24/32, baseline, and 30/35.
The strict mean chart uses the same values as the table above. The score axis is intentionally narrowed and labelled because these candidates are close together.

Decision From The Full List

The page should not imply that every possible single-layer duplication across all 64 layers was run. The documented evidence supports: 13 early AEON short-scan mappings, 6 strict AEON mappings, one true single-layer candidate in the short scan, and a separate single-entry strict file for 15,20.

Source Model And RYS Build

The source branch for this release is AEON-7/Qwen3.6-27B-AEON-Ultimate-Uncensored. The RYS operation was a checkpoint transformation: duplicate a trained layer window and insert it back into the stack.

RYS 15/20 insert mapping showing source layers 15 to 19 duplicated into output layers 20 to 24.
For blocks:15,20, source layers 15..19 are duplicated after source layer 19. Output layers 20..24 are the copied window, and source layer 20 resumes at output layer 25.

Build Rule

Every output layer must copy its weights and execution metadata from the same source layer. For Qwen3.6 hybrid models, remapping text_config.layer_types with the same output-to-source map was execution-critical.

Failure Mode

A broken RYS checkpoint can load and still generate badly. The key trap was treating the copied tensors and the per-layer execution plan as separate ledgers.

Layer-Window Selection

The selected window was not chosen because it had the highest score on every possible slice. It was chosen because it won the balanced AEON strict scan and later survived the practical Q4_NL compression screen better than the stronger math-leaning branch.

Strict probe chart comparing AEON RYS candidate windows.
The strict scan used AEON validation files, not the older official-base scan. The official-base winner did not transfer as the AEON winner.
Specmath_120eq_140CombinedDelta vs AEON baselineRead
blocks:15,200.9714410.6473470.809394+1.181%Best balanced strict candidate.
blocks:31,340.9772690.6407500.809010+1.133%Very close second.
blocks:11,140.9814980.6349290.808214+1.034%Best strict math, not combined winner.
blocks:24,320.9718230.6285590.800191+0.031%Near baseline.
blocks:0,00.9701810.6297100.799945baselineAEON baseline.
blocks:30,350.9648410.6275210.796181-0.471%Official-base winner, not AEON winner.

Final Materialized 15/20 Check

After the materialized 15/20 artifact was built, strict checks recorded 0.983317798162 on artifact_15_20_strict_math120.pkl with 120/120, and 0.647756674780 on artifact_15_20_strict_eq140.pkl with 139/139. Combined: 0.815537236471.

Short Scan Caveat

The earlier 13-candidate short scan was not the final selection basis. In that short math_16 + eq_16 view, 24,32 was slightly higher than 15,20, so the release choice depends on the stricter balanced scan plus quantization survival.

Decision

Select 15,20 for the AEON release branch because it had the best balanced strict result. Keep 11,14 in mind as a math-leaning research branch, not the default release target.

Quantization Survival

This is the main reason the public release exists as a Q4-class model. The RYS BF16 did not produce a huge headline gain by itself. The useful result was that the selected branch held up unusually well after compression.

Quantization survival cards comparing RYS candidate behavior after quantization.
The quantization result is the core release argument: make the small file useful, not just make a larger BF16 variant.
Visual comparison of BF16 reference artifact versus released IQ4_NL deployment target.
The BF16 files are for exploration and continued work. The named project target is the compressed IQ4_NL release path.
ProbeRYS BF16Released IQ4_NLChange
Mixed four-probe mean0.7298990.724435-0.005465 / -0.7487%
math_160.8421430.789686down
eq_160.7123400.711090near flat
math_40.4851160.516963up
gsm8k_50.8800000.880000flat
Grouped bar chart comparing BF16 and released IQ4_NL scores across the mixed mean, math_16, eq_16, math_4, and gsm8k_5 probes.
The released IQ4_NL stayed within -0.0055 absolute of the BF16 mixed four-probe mean, while the individual probes moved in different directions.

Decision

Keep the main release centered on IQ4_NL because the mixed four-probe drop was small relative to the size reduction: about 57.6 GB BF16 GGUF to 16.6 GB IQ4_NL. Runtime speed helped the deployment story, but preserving useful reasoning/coding quality through compression is the core result.

Important Caveat

The 11,14 quant headline is not the same scoreboard. Its two-probe reasoning slice went from 0.854324 BF16 to 0.729260 IQ4_NL, a much larger drop, but that file only used math_4 and gsm8k_5. A separate no-think math_16 + eq_16 file showed a smaller drop, 0.759275 to 0.738754. Defensible read: 11/14 was strong on some probes but more volatile, and it did not win the strict-balanced scan.

Runtime Profile

The model is released for the custom AEON ik-llama fork. That is not packaging trivia; it is part of the tested artifact. The fork carries the Qwen3.6 hybrid and graph-split work needed for this line. In the internal runtime comparison, the recommended custom path decoded at 39.37 tok/s versus 22.51 tok/s for the patched upstream-style comparison; useful speed, but still secondary to the quality-preservation result that made the IQ4_NL file worth releasing.

./build/bin/llama-server \
  -m /path/to/Qwen3.6-27B-AEON-RYS-MaxThinkCoder-IQ4_NL-ik-llama-custom-mixed.gguf \
  -c 65536 \
  -ngl 999 \
  -np 1 \
  -fa on \
  -sm graph \
  --temp 0.7 \
  --jinja \
  --reasoning-format deepseek \
  --reasoning-budget 0 \
  -cram 0 \
  --ctx-checkpoints 0

Why the Fork Exists

The project needed custom mixed-GGUF support, Qwen3.6/Qwen3.5 hybrid handling, graph-split stability work, Jinja chat formatting, DeepSeek reasoning extraction, and May 2026 duplicate tool-call filtering.

Long Context

The public profile starts with -c 65536. The same family was also used for 131072 context comparisons, and default/FP16 KV was separately tested to about 160k context without the earlier failure pattern. FP32 KV was a conservative validation setting, not a requirement.

Practical Single-GPU Deployment

The Q4_NL release is small enough for practical single-GPU deployment. In an observed 24 GB-class GPU reference profile, roughly 160k context with default/FP16 KV fit at about 20.3 GiB total VRAM on an RTX 3090-class card. Treat this as a practical deployment reference point, not a guaranteed cross-hardware memory benchmark.

Runtime and implementation cards for the required fork, GGUF representation, long-context KV caveat, and experimental MTP branch.
The runtime cards collect the practical constraints that matter before someone treats the file as a stock GGUF.
Runtime CheckSetupDecodeDecision
Recommended custom pathGraph split, long-context deployment profile, FP32 KV validation snapshot.39.37 tok/sPublic runtime path; faster in this internal comparison, with the release claim still anchored on quality preservation.
Patched upstream-style comparisonInternal standard-typed comparison path, shorter context, layer-style comparison.22.51 tok/sNot released as the public target.

MTP And Speed Work

The MTP file is real and structurally valid, but it did not replace the default. We kept it because it is useful for runtime research, not because it beat the normal non-MTP file.

CheckResultRead
MTP metadataqwen35.nextn_predict_layers = 1MTP tail exists.
MTP tensor coverage8 quantizable blk.69 tensorsPatched imatrix collection covered the MTP tail.
MTP-aware imatrix80 chunks, PPL 4.3762 +/- 0.07886Calibration path completed.
512-token quality0.166138 score with MTPEffectively identical to the old MTP GGUF in that suite.
2048-token no-MTP quality on MTP fileworst repeat 59Worse repeat penalty than the practical non-MTP default.
Speed CheckDecodePromptRead
No-MTP graph split reference48.6795 tok/s214.9709 tok/sFinal clean matrix reference; still fastest.
Naive MTP draft-138.1550 tok/snoted in matrixAcceptance 225/345 = 65.217%.
Adaptive MTP short check45.3589 tok/s208.79 tok/sAcceptance 158/218 = 72.477%; closer, still not enough to replace default.
No-MTP long 768-token check48.7140 tok/slong checkLonger generation still favored no-MTP.
Adaptive MTP long 768-token check46.9539 tok/slong checkAcceptance 38/54 = 70.37%; close but still behind.
Horizontal bar chart comparing no-MTP, naive MTP, and adaptive MTP decode-speed notes.
These are internal same-machine speed notes used for the release decision, not a normalized public speed benchmark. In the tested paths, no-MTP stayed slightly ahead.

Decision

Publish the MTP GGUF as experimental. Keep the non-MTP IQ4_NL file as the default because practical quality and speed still favored it.

Rejected Or Non-Default Variants

Several useful experiments did not become the public default. They are included here because the negative results explain the release shape.

VariantTestResultDecision
11,14 IQ4_NL Long-reasoning fair comparison, math_16 + gsm8k_5, 21 questions, 2048 max tokens. Same final/any exact rates as 15/20, but much cleaner repetition: worst 4gram repeat 7 vs 47 for 15/20. This resolves the long-reasoning Q4_NL setting in favor of cleaner 11/14 repetition, but does not replace the selected release without matching evidence across the mixed quant suite, runtime packaging, and practical agent tests.
RYS-window mixed-Q8 Force 46 attention/SSM tensors in layers 15..24 to Q8_0. File size +3.35%; mean best-rel 0.87299 vs 0.89998 baseline in a 9-item quick paired eval. Do not replace IQ4_NL. Consider narrower Q8 variants later.
Standard llama.cpp-style public file Internal patched upstream-style comparison path. Still required special runtime assumptions and was not the main tested target. Do not present as stock llama.cpp support.
MTP default MTP graph split, graph reuse, adaptive gate, MTP-tail imatrix. Technically valid but slower or less clean than no-MTP in tested paths. Publish as experimental only.
Two-panel chart showing 11/14 versus 15/20 composite score and worst 4-gram repeat in the long-reasoning caveat test.
The 11/14 branch deserves the caveat: it tied final/any exact rates in this slice and had much cleaner repetition. It still was not rerun through the full release decision ladder.

Practical Agent Tests

These checks use the released non-finetuned AEON RYS 15/20 IQ4_NL artifact unless a row explicitly says otherwise. They are practical coding-agent checks, not broad benchmarks. The numbers needed to interpret them are reproduced here; the file paths in the ledger are provenance only.

Scope Boundary

Some practical rows were collected during later SignalLatch and Unsloth comparison work. This page uses only the AEON RYS IQ4_NL base rows. SignalLatch strength-sweep and clean-pass claims belong to the separate fine-tune page.

Five-task Matrix

Base run: AEON RYS IQ4_NL. Setting: temp 0.7, graph split, flash attention, Jinja/DeepSeek, 65536 context. Result: strict pass 1/5, mean 0.550, task scores 0.75, 1.00, 0.25, 0.25, 0.50, timeout-like tasks 4.

Canvas Attempt 1

Base run: AEON RYS IQ4_NL attempt 1. Setting: temp 0.7, 131072 context, FP32 KV, graph split, flash attention. Result: rc=1, 337s, verifier 0.0417 / false, root files: none.

Canvas Retry

Base run: AEON RYS IQ4_NL retry 1. Same task and runtime family. Result: rc=0, 803s, verifier 1.0 / true, complete app files. Read: the base can complete the task, but first-attempt reliability remains a caveat.

Production Matrix TaskScorePassRC / TimeVerifier Read
github_mcp_commits_fix_repeat0.75Norc=1, 260sBuild, branch schema, branch output, and README checks passed; request path/branch parameter checks failed.
github_mcp_pr_details_fix1.00Yesrc=124, 600sCorrectly used the PR detail endpoint and detail additions/deletions/changed-files fields, but still hit the full timeout.
local_search_kill_excess_fix0.25Norc=124, 600sBuild passed; targeted process-kill behavior was not implemented.
local_search_search_timeout_fix0.25Norc=124, 600sBuild passed; timeout schema and handler propagation were not implemented.
local_search_web_search_race_fix0.50Norc=124, 600sMultiple engines remained, but the first-success race behavior was not implemented.
Canvas Harness DetailValue Included Here
PromptBuild an isolated Krita-like raster canvas app with layers, brush/eraser, transforms, opacity, and a local AI image-generation stub.
Shared settingsTemp 0.7, context 131072, FP32 K/V cache, flash attention, graph split, Jinja, DeepSeek reasoning format, CLAW_MAX_TOKENS=1800, TIMEOUT_SECONDS=900.
Retry root filesindex.html 3,450 bytes; styles.css 7,062 bytes; app.js 17,551 bytes; README.md 1,033 bytes.
Excluded comparison rowsThe full later comparison also included SignalLatch IQ4_NL, Unsloth IQ4_NL, and Unsloth Q8_0 rows. Those are not used as evidence for this non-finetuned base release.

Decision

Keep this release claim narrow: AEON RYS 15/20 IQ4_NL is a viable compressed coding-agent base with practical competence and documented reliability caveats. SignalLatch is a later behavior-finetuned attempt to improve that reliability, not part of the non-finetuned release score.

Implementation Footnotes

The duplicate-window representation is recorded here for transparency, but it is not the headline claim of the release. The main claim remains quantization survival and practical runtime behavior for the tested IQ4_NL file.

GGUF file size16,554,834,080 bytes. This is the released default IQ4_NL file size.
Materialized duplicate spanOutput layers 20..24 occupy 1,067,475,584 bytes / 0.994 GiB as copied RYS-window tensors.
File-level footprintThe duplicate span is about 6.448% of the released file, or 7.220% of transformer block tensor bytes.
Logical layer count69 logical layers: 64 source layers plus 5 inserted RYS layers.
Long-context KV scalingAt 131,072 tokens, the extra KV/cache estimate is about +1.0 GiB FP16 / +2.0 GiB FP32. Around 160k tokens, it scales to about +1.2 GiB FP16 / +2.4 GiB FP32; at 163,840 tokens, about +1.25 GiB FP16 / +2.5 GiB FP32. The inserted window contains two full-attention layers, so KV cost scales with context length; all five inserted layers still add compute.

Why This Version Stayed Materialized

Materialized tensors kept the HF checkpoint, GGUF conversion, quantization, and downstream fine-tune/LoRA workflows explicit and stable for the tested release.

Future Optimization

A procedural or aliased RYS runtime could reuse source-layer weight buffers and save duplicate-weight memory. That is possible future runtime work, not a change to this already tested release artifact.

Evidence Ledger

This ledger records the source filenames used to reconstruct the page. It is a provenance map, not required reading: the public numeric data needed to understand the release is reproduced above in the tables, charts, captions, and artifact rows.

Workspace status qwen36_rys_work/README.md records current defaults, directory map, RYS semantics, strict top-6 table, quantization snapshot, GGUF/imatrix paths, and serving notes.
Strict AEON scan qwen36_aeon_validation/results/aeon_strict_math120.pkl and qwen36_aeon_validation/results/aeon_strict_eq140.pkl.
15/20 BF16 vs IQ4_NL qwen36_aeon_validation/results/quant_compare_15_20_q4_vs_bf16_math16_eq16_reasoning_math4_gsm8k5_20260426.json.
11/14 quant caveat qwen36_aeon_validation/results/quant_compare_11_14_q4_vs_bf16_reasoning_math4_gsm8k5_20260426.json.
11/14 vs 15/20 long-reasoning comparison qwen36_rys_work/rys_fair_compare_11_14_vs_15_20/results/q4nl_reason2048_ctx32768_bwrap0349_20260430_123403/.
Mixed-Q8 probe qwen36_rys_work/aeon_rys_15_20_gguf/mixed_q8_ryswin/README_results.md and latest_quick_eval_results.json.
MTP-aware imatrix and quality qwen36_rys_work/aeon_rys_15_20_mtp_gguf/imatrix_mtp_iq4nl/README_results.md.
MTP speed search qwen36_rys_work/mtp_speed_probe/autoresearch_3090/RESULTS_3090_MTP_SPEED_20260501.md and qwen36_rys_work/mtp_rebuild_20260501_151130/final_results.md.
Practical production matrix base row docs/ckpt386-s010-testing-process/evidence/base_q4nl_summary.md. The base task scores, pass flags, return codes, and timings are reproduced in the Practical Agent Tests section.
Practical canvas comparison qwen36_rys_work/aeon_rys_15_20_signallatch_gguf/evidence/canvas_unsloth_comparison_20260505_summary.md. Only the AEON RYS IQ4_NL attempt/retry rows are used for this base-release page; SignalLatch and Unsloth rows are downstream or external comparison context.
Published model card Qwen3.6-27B-AEON-RYS-15-20-GGUF.