Return to the chalkboard.
Yes, it happened again. It's time to return to the chalkboard and make some new components.
Building from linear, baseline 20% accuracy on cifar10, Blues Brothers reference, it's midnight, and we're wearing sunglasses.
Hit it.
We're going clean in, and making these omega tokens sing. Every single will have a purpose, and a tracking point as to WHY they have a purpose.
Nothing ambiguous. Nothing left to up another mechanism.
MLP testing
I'm extracting and preserving the 4096 omega tokens [b, 4, 64, 64] from the fairly tanky Freckles-4096
With these extracted cifar10 features it will allow for much more rapid prototyping.
Freckles-4096 epoch 1 battery is ran
Check the results in the directory https://huggingface.co/AbstractPhil/geolip-SVAE/tree/main/v41_freckles_256
This is a patchwork tokenization size of 4096 related directly to the cross-attention, so be aware I'm not bloating numbers here.
This is literally how many patches exist, and this many patches exist specifically in the spectrum of 256x256. As per the scaling rule applied by the internal architectural resonance.
The sizes are internally very very different, but the actual information and parameters are identical.
enc_in [262144, 384]
enc_block_0 [262144, 384]
enc_block_1 [262144, 384]
enc_block_2 [262144, 384]
enc_block_3 [262144, 384]
enc_out_raw [262144, 192]
cross_attn_0_qkv [64, 4096, 12]
cross_attn_0_in [64, 4096, 4]
cross_attn_0_out [64, 4096, 4]
cross_attn_1_qkv [64, 4096, 12]
cross_attn_1_in [64, 4096, 4]
cross_attn_1_out [64, 4096, 4]
dec_in [262144, 384]
dec_block_0 [262144, 384]
dec_block_1 [262144, 384]
dec_block_2 [262144, 384]
dec_block_3 [262144, 384]
dec_out [262144, 48]
boundary_in [64, 3, 256, 256]
boundary_out [64, 3, 256, 256]
svd_U [64, 4096, 48, 4]
svd_S_orig [64, 4096, 4]
svd_S [64, 4096, 4]
svd_Vt [64, 4096, 4, 4]
svd_M [64, 4096, 48, 4]
recon [64, 3, 256, 256]
input [64, 3, 256, 256]
Keep in mind when I test Freckles with images, Freckles has never trained on a single image.
Freckles has ONLY SEEN NOISE. 16 types of noise. That's it, historically Freckles is image ignorant.
Meaning when I train heads and classification models with Freckles features, I'm training models with what Freckles turns pure images into.
Freckles has zero knowledge of any real data, ever. Just noise.
Freckles-4096 epoch 1 is complete
The MSE shows roughly the same as the freckles-256
Battery testing begins soon.
Little freckles couldn't beat pixel-in transformers
With just her 256 patches condensed into 64 dim features.
However... We're just getting started. The little 64 dim features were good, but not the full capacity of freckles, not by a long shot.
Next up is the 384 dim hidden features, no more pulling punches.
======================================================================
HEAD-TO-HEAD COMPARISON
======================================================================
Omega Processor (Freckles features): 72.7%
Baseline (raw patches): 77.2%
Delta: -4.5%
Random chance: 10.0%
======================================================================
Well fought little bottlenecked Freckles, it's time to unlock the limit gates now.
First Omega Processor Tests
Omega Processor Baseline (raw)
Final accuracy: 92.8% 82.0%
Params: 837,008 867,824
Epoch time: 34s 19s
PER-CLASS FINAL:
Omega Baseline
gaussian 100% 86%
uniform 99% 51% ← baseline can't separate
uniform_sc 100% 49% ← from its sibling
poisson 94% 97%
pink 33% 84% ← baseline does BETTER here
brown 64% 16% ← but worse here
salt_pepper 100% 100%
sparse 100% 100%
block 100% 100%
gradient 100% 100%
checker 100% 78%
mixed 95% 46% ← massive gap
structural 100% 100%
cauchy 100% 100%
exponential 100% 100%
laplace 100% 100%
UNSOLVABLE PAIRS:
Baseline: uniform/uniform_sc (51/49), pink/brown (84/16), mixed (46%), checker (78%)
Omega: pink/brown (33/64) — one remaining confusion pair
The omega processor features in this state are more robust than dataset introduced features. The convergence is faster, the data learns faster, and the output is clearly pretrained to a direct extreme.
Now... lets try Cifar10 shall we? I'll just feed it into our little Johanna-256, no image training ever seen.
4/9/2026 Omega Processer Prototype
Freckles encoder (frozen)
↓
SVD: U, S, Vt (exact, frozen)
↓
Feature Extractor (tiny, learned):
scalar_features(S) → 16 dims
relational_features(S, grid) → 16 dims
basis_features(U, Vt) → 32 dims
──────────────────────────────────────
concat → 64 dims per patch
↓
LayerNorm + Linear(64, d_model) → project to transformer dim
↓
Standard transformer encoder/decoder
↓
Task head (classification, generation, text, etc.)
Due to the model being so deviant from transformers, Claude and I have come up with a potential scaffold.
It aligns with the geometry and has a very small learning curve, which should coalesce into the necessary behavior.
This SHOULD allow any model to be able to utilize the information without needing a massive architectural change.
4/9/2026 Freckles - Resolution Scaling!?!
Freckles handles noise at massive ratios if I piecemeal the information together.
I'm going to try massive sizes, and test every single noise spectrum.
Currently Freckles handles MSE 0.000005 noise error from 16 types of noise, the tests show rigidity in the patches so they can't be slid around.
The manifold shows you can just... aim it at any noise target and it will solve via the patches, and the patches can be easily calculated to which is most likely the noise type with very low compute cost.
Noise type, noise location within structure, structure of that encapsulated system's image space... Alright yeah this is, well beyond expectation.
These are resolution independent models, resolution means nothing to them. I believe as well, that Freckles being 2m params, is substantially more powerful than each of the others at their 17m params.
The Omega Processor
The plan now is to wrap the center of the VAE structure with a memory bank driven constellation observer with hundreds of thousands of anchors.
We will see, exactly what this model is doing, why this model is doing that behavior, what we need to do to replicate this model, what steps we need to take to build the necessary optimizations to build the scaffolding for those tools, and everything between here and there.
We have TOUCHED Omega now, and I will NOT let this vapor dissolve. I will not let these numbers simply get mashed into a transformer without full observation and understanding of WHAT we are seeing. I MUST know.
IF THIS IS NOT TRULY THE SELF SOLVING FRAME, we will know VERY soon. VERY VERY soon. I will redact my claims of Omega and continue towards the legitimate self-solving quantum state that we are all looking for.
I TRULY BELIEVE this has ENOUGH potential for me to spend a good amount of time investigating.
If I can even find a SINGLE PIECE of truly high-dimensional scaffolding that doesn't directly conform to the dimensions; that somehow survived the battery of CM volume, that somehow formed knots upon knots, upon knots of impossible conjecture - that can be turned into a viable utility. Then I will have finished my task. Everything will have been proven and I can simply build.
All we need is JUST ONE. JUST ONE, and it will be enough to create the universe.
4/8/2025 Prototype V13, 14, 15, etc
Johanna - Noise variant 48:1
Patchworks Available:
- 64x64 0.06 mse
- 128x128 0.002 mse
- 256x256 0.0002 mse
17m params all
48:1 compression
[16, 8, 8] omega tokens
fair mse decoding on 16 noise types
This model can handle noise, images, audio, whatever - semi okay.
This is the one you want to finetune if you have a task you want to try.
Fresnel - Image variant 48:1
Patchworks Available:
- 64x64 0.00001~ mse
- 128x128 0.000002 mse
- 256x256 0.000001 mse
17m params
48:1 compression
[16, 8, 8] omega tokens
0.000002 mse on clean image latents
Don't feed fresnel noise, the model does not understand noise
Johanna is the noise model.
Grandmaster - Denoise Variant 48:1
Finetuned Johanna-128
Patchworks Available:
- 128x128 mse 0.0042 mse
17m params
48:1 compression
[16, 8, 8] omega tokens
This model accepts a noisy image input, and returns denoised output.
Trained using Fresnel's omega tokens as training targets, has a fair SNR denoising capacity already at 0.0042 mse.
This is part of a prototype meant to skip diffusion steps entirely.
Freckles - Noise Variant 12:1
Patchworks Available:
- 4x4 0.0002 mse
- gaus=0.001 unif=0.000 unif=0.001 pois=0.000
- pink=0.000 brow=0.000 salt=0.004 spar=0.000
- bloc=0.000 grad=0.000 chec=0.000 mixe=0.000
- stru=0.000 cauc=0.002 expo=0.001 lapl=0.001
2.7m params
12:1 compression
Uhhhh.. yeah I uhh... didn't expect this one to recon so well. I didn't set the measures to register lower than that.
I'll report later.
Indev:
Alexandria - Text variant
17m params
48:1 compression
DOES NOT WORK YET!
Prototype v12 - Patch 16 - 128x128 images - 1.2m imagenet images
Well the 64x64 image set worked just fine, so it's time to upgrade and test the limits of the architecture.
Can it simply... scale? ooooor do we need more solvers along the way to compensate?
benjamin-paine/imagenet-1k-128x128
Can we actually solve it
YES WE DID!
It seems geometric manifolds learn... differently than standard manifolds, don't they.
Prototype V11 - Patch16 - MSE 0.0005 - 64x64 tiny imagenet
I'd say it works.
The images show it works. It works.
Using geolip-core SVD (fp64 Gram+eigh (FL=available, N<=12))
PatchSVAE - 16 patches of 16×16
Dataset: tiny_imagenet (64×64, 200 classes)
Per-patch: (256, 16) = 4096 elements, rows on S^15
Encoder/Decoder: hidden=768, depth=4 (residual blocks)
Cross-attention: 2 layers on S vectors (2,272 params)
Soft hand: boost=1.5x near CV=0.125, penalty=0.3 far
Total params: 16,942,419
===============================================================================================
ep | loss recon t/ep | t_rec | S0 SD ratio erank | row_cv prox rw | S_delta
-----------------------------------------------------------------------------------------------
1 | 0.2595 0.1806 12.2 | 0.1024 | 5.036 3.254 1.55 15.87 | 0.2007 0.905 1.45 | 0.09694 a:0.0242/0.0247
2 | 0.1216 0.0845 12.3 | 0.0675 | 5.071 3.298 1.54 15.88 | 0.2018 0.885 1.44 | 0.17411 a:0.0251/0.0257
3 | 0.0847 0.0587 12.3 | 0.0470 | 5.093 3.312 1.54 15.88 | 0.2046 0.869 1.43 | 0.19894 a:0.0258/0.0265
4 | 0.0623 0.0432 12.3 | 0.0430 | 5.115 3.323 1.54 15.88 | 0.2006 0.864 1.43 | 0.20848 a:0.0264/0.0272
6 | 0.0359 0.0248 12.3 | 0.0198 | 5.129 3.332 1.54 15.88 | 0.2006 0.907 1.45 | 0.21832 a:0.0273/0.0281
8 | 0.0225 0.0155 12.2 | 0.0196 | 5.149 3.341 1.54 15.87 | 0.2017 0.876 1.44 | 0.22351 a:0.0279/0.0287
10 | 0.0170 0.0116 12.3 | 0.0100 | 5.151 3.352 1.54 15.88 | 0.2035 0.924 1.46 | 0.22671 a:0.0283/0.0290
12 | 0.0141 0.0096 12.3 | 0.0114 | 5.159 3.354 1.54 15.88 | 0.2009 0.909 1.45 | 0.22924 a:0.0285/0.0293
14 | 0.0121 0.0082 12.3 | 0.0073 | 5.156 3.362 1.53 15.88 | 0.2018 0.855 1.43 | 0.23137 a:0.0288/0.0296
16 | 0.0105 0.0072 12.3 | 0.0108 | 5.161 3.363 1.53 15.88 | 0.2003 0.860 1.43 | 0.23316 a:0.0290/0.0298
18 | 0.0094 0.0064 12.3 | 0.0055 | 5.158 3.365 1.53 15.88 | 0.2017 0.879 1.44 | 0.23467 a:0.0292/0.0300
20 | 0.0086 0.0058 12.3 | 0.0050 | 5.157 3.367 1.53 15.88 | 0.2023 0.805 1.40 | 0.23601 a:0.0293/0.0301
22 | 0.0079 0.0054 12.4 | 0.0045 | 5.157 3.369 1.53 15.88 | 0.1996 0.872 1.44 | 0.23726 a:0.0295/0.0303
24 | 0.0074 0.0050 12.2 | 0.0064 | 5.146 3.380 1.52 15.88 | 0.2044 0.879 1.44 | 0.23848 a:0.0296/0.0305
26 | 0.0068 0.0046 12.4 | 0.0039 | 5.155 3.372 1.53 15.88 | 0.2036 0.884 1.44 | 0.23955 a:0.0297/0.0306
28 | 0.0063 0.0042 12.3 | 0.0036 | 5.155 3.378 1.53 15.88 | 0.2077 0.841 1.42 | 0.24057 a:0.0299/0.0307
30 | 0.0058 0.0038 12.3 | 0.0038 | 5.155 3.380 1.53 15.88 | 0.2027 0.911 1.46 | 0.24149 a:0.0300/0.0309
32 | 0.0055 0.0036 12.2 | 0.0032 | 5.150 3.383 1.52 15.88 | 0.2045 0.807 1.40 | 0.24239 a:0.0301/0.0310
34 | 0.0054 0.0036 12.3 | 0.0037 | 5.145 3.388 1.52 15.88 | 0.1996 0.875 1.44 | 0.24329 a:0.0302/0.0311
36 | 0.0049 0.0032 12.3 | 0.0031 | 5.154 3.385 1.52 15.88 | 0.2054 0.828 1.41 | 0.24409 a:0.0303/0.0312
38 | 0.0046 0.0030 12.3 | 0.0027 | 5.152 3.390 1.52 15.88 | 0.2038 0.847 1.42 | 0.24490 a:0.0304/0.0313
40 | 0.0044 0.0029 12.3 | 0.0032 | 5.155 3.392 1.52 15.89 | 0.2046 0.855 1.43 | 0.24566 a:0.0305/0.0314
42 | 0.0043 0.0028 12.3 | 0.0024 | 5.152 3.395 1.52 15.89 | 0.2064 0.905 1.45 | 0.24637 a:0.0305/0.0315
44 | 0.0042 0.0027 12.3 | 0.0023 | 5.150 3.395 1.52 15.89 | 0.2084 0.844 1.42 | 0.24705 a:0.0306/0.0316
46 | 0.0039 0.0025 12.3 | 0.0022 | 5.149 3.400 1.51 15.89 | 0.2057 0.868 1.43 | 0.24776 a:0.0307/0.0317
48 | 0.0037 0.0024 12.3 | 0.0024 | 5.152 3.403 1.51 15.89 | 0.2138 0.831 1.42 | 0.24843 a:0.0308/0.0318
50 | 0.0038 0.0024 12.3 | 0.0025 | 5.149 3.406 1.51 15.89 | 0.2078 0.810 1.40 | 0.24906 a:0.0309/0.0319
52 | 0.0034 0.0021 12.3 | 0.0019 | 5.154 3.405 1.51 15.89 | 0.2082 0.872 1.44 | 0.24965 a:0.0309/0.0320
54 | 0.0033 0.0020 12.2 | 0.0019 | 5.156 3.406 1.51 15.89 | 0.2085 0.894 1.45 | 0.25022 a:0.0310/0.0320
56 | 0.0033 0.0020 12.4 | 0.0019 | 5.150 3.412 1.51 15.89 | 0.2058 0.866 1.43 | 0.25079 a:0.0311/0.0321
58 | 0.0031 0.0019 12.3 | 0.0033 | 5.147 3.416 1.51 15.89 | 0.2071 0.774 1.39 | 0.25135 a:0.0311/0.0322
60 | 0.0030 0.0018 12.4 | 0.0017 | 5.153 3.415 1.51 15.89 | 0.2134 0.840 1.42 | 0.25187 a:0.0312/0.0323
62 | 0.0030 0.0018 12.3 | 0.0016 | 5.155 3.416 1.51 15.89 | 0.2080 0.764 1.38 | 0.25235 a:0.0313/0.0323
64 | 0.0028 0.0017 12.2 | 0.0014 | 5.156 3.416 1.51 15.89 | 0.2100 0.666 1.33 | 0.25285 a:0.0313/0.0324
66 | 0.0028 0.0017 12.3 | 0.0017 | 5.151 3.419 1.51 15.89 | 0.2101 0.865 1.43 | 0.25333 a:0.0314/0.0324
68 | 0.0026 0.0015 12.3 | 0.0014 | 5.158 3.419 1.51 15.89 | 0.2078 0.838 1.42 | 0.25381 a:0.0314/0.0325
70 | 0.0025 0.0015 12.4 | 0.0021 | 5.160 3.422 1.51 15.89 | 0.2112 0.806 1.40 | 0.25428 a:0.0315/0.0326
72 | 0.0026 0.0015 12.2 | 0.0013 | 5.158 3.422 1.51 15.89 | 0.2126 0.835 1.42 | 0.25471 a:0.0316/0.0326
74 | 0.0024 0.0014 12.2 | 0.0015 | 5.154 3.427 1.50 15.89 | 0.2143 0.838 1.42 | 0.25514 a:0.0316/0.0327
76 | 0.0024 0.0014 12.2 | 0.0012 | 5.161 3.424 1.51 15.89 | 0.2151 0.847 1.42 | 0.25553 a:0.0317/0.0327
78 | 0.0023 0.0013 12.3 | 0.0014 | 5.157 3.428 1.50 15.89 | 0.2121 0.686 1.34 | 0.25592 a:0.0317/0.0328
80 | 0.0024 0.0013 12.3 | 0.0012 | 5.160 3.428 1.51 15.89 | 0.2068 0.824 1.41 | 0.25630 a:0.0317/0.0328
82 | 0.0027 0.0016 12.2 | 0.0015 | 5.146 3.394 1.52 15.89 | 0.2065 0.899 1.45 | 0.25687 a:0.0318/0.0329
84 | 0.0022 0.0013 12.2 | 0.0013 | 5.156 3.405 1.51 15.89 | 0.2092 0.875 1.44 | 0.25709 a:0.0319/0.0329
86 | 0.0022 0.0012 12.3 | 0.0022 | 5.154 3.413 1.51 15.89 | 0.2091 0.835 1.42 | 0.25726 a:0.0319/0.0329
88 | 0.0021 0.0012 12.3 | 0.0014 | 5.154 3.417 1.51 15.89 | 0.2074 0.840 1.42 | 0.25740 a:0.0319/0.0329
90 | 0.0022 0.0013 12.3 | 0.0030 | 5.147 3.416 1.51 15.89 | 0.2191 0.848 1.42 | 0.25753 a:0.0319/0.0329
92 | 0.0021 0.0011 12.3 | 0.0012 | 5.157 3.418 1.51 15.89 | 0.2123 0.775 1.39 | 0.25766 a:0.0319/0.0329
94 | 0.0021 0.0011 12.3 | 0.0010 | 5.156 3.419 1.51 15.89 | 0.2117 0.710 1.35 | 0.25779 a:0.0319/0.0329
96 | 0.0020 0.0011 12.3 | 0.0013 | 5.154 3.420 1.51 15.89 | 0.2166 0.940 1.47 | 0.25793 a:0.0319/0.0330
98 | 0.0019 0.0011 12.2 | 0.0010 | 5.156 3.421 1.51 15.89 | 0.2143 0.762 1.38 | 0.25807 a:0.0320/0.0330
100 | 0.0020 0.0010 12.3 | 0.0009 | 5.155 3.422 1.51 15.89 | 0.2173 0.642 1.32 | 0.25821 a:0.0320/0.0330
102 | 0.0020 0.0010 12.2 | 0.0009 | 5.156 3.423 1.51 15.89 | 0.2165 0.868 1.43 | 0.25835 a:0.0320/0.0330
104 | 0.0019 0.0010 12.4 | 0.0009 | 5.157 3.423 1.51 15.89 | 0.2125 0.788 1.39 | 0.25850 a:0.0320/0.0330
106 | 0.0019 0.0009 12.3 | 0.0009 | 5.156 3.424 1.51 15.89 | 0.2219 0.666 1.33 | 0.25866 a:0.0320/0.0331
108 | 0.0019 0.0009 12.3 | 0.0009 | 5.153 3.425 1.50 15.89 | 0.2202 0.671 1.34 | 0.25881 a:0.0321/0.0331
110 | 0.0020 0.0009 12.3 | 0.0011 | 5.153 3.427 1.50 15.89 | 0.2163 0.726 1.36 | 0.25896 a:0.0321/0.0331
112 | 0.0019 0.0009 12.4 | 0.0009 | 5.155 3.427 1.50 15.89 | 0.2205 0.837 1.42 | 0.25911 a:0.0321/0.0331
114 | 0.0019 0.0008 12.3 | 0.0008 | 5.155 3.427 1.50 15.89 | 0.2220 0.803 1.40 | 0.25926 a:0.0321/0.0332
116 | 0.0018 0.0008 12.3 | 0.0009 | 5.155 3.427 1.50 15.89 | 0.2211 0.852 1.43 | 0.25942 a:0.0321/0.0332
118 | 0.0019 0.0008 12.3 | 0.0008 | 5.153 3.429 1.50 15.89 | 0.2207 0.694 1.35 | 0.25957 a:0.0321/0.0332
120 | 0.0018 0.0008 12.2 | 0.0008 | 5.156 3.429 1.50 15.89 | 0.2271 0.664 1.33 | 0.25972 a:0.0322/0.0332
122 | 0.0018 0.0008 12.3 | 0.0008 | 5.154 3.429 1.50 15.89 | 0.2266 0.658 1.33 | 0.25986 a:0.0322/0.0333
124 | 0.0017 0.0007 12.3 | 0.0008 | 5.154 3.431 1.50 15.89 | 0.2201 0.771 1.39 | 0.26000 a:0.0322/0.0333
126 | 0.0017 0.0007 12.4 | 0.0008 | 5.157 3.430 1.50 15.89 | 0.2253 0.862 1.43 | 0.26014 a:0.0322/0.0333
128 | 0.0018 0.0007 12.3 | 0.0008 | 5.153 3.431 1.50 15.89 | 0.2222 0.638 1.32 | 0.26027 a:0.0322/0.0333
130 | 0.0018 0.0007 12.3 | 0.0007 | 5.155 3.431 1.50 15.89 | 0.2255 0.786 1.39 | 0.26040 a:0.0322/0.0333
132 | 0.0018 0.0007 12.3 | 0.0007 | 5.153 3.432 1.50 15.89 | 0.2325 0.778 1.39 | 0.26053 a:0.0323/0.0333
134 | 0.0017 0.0007 12.3 | 0.0007 | 5.154 3.433 1.50 15.89 | 0.2250 0.876 1.44 | 0.26065 a:0.0323/0.0334
136 | 0.0017 0.0006 12.3 | 0.0007 | 5.157 3.432 1.50 15.89 | 0.2269 0.866 1.43 | 0.26077 a:0.0323/0.0334
138 | 0.0017 0.0006 12.3 | 0.0007 | 5.156 3.433 1.50 15.89 | 0.2247 0.760 1.38 | 0.26088 a:0.0323/0.0334
140 | 0.0016 0.0006 12.3 | 0.0006 | 5.157 3.433 1.50 15.89 | 0.2242 0.725 1.36 | 0.26099 a:0.0323/0.0334
142 | 0.0016 0.0006 12.3 | 0.0006 | 5.157 3.433 1.50 15.89 | 0.2241 0.909 1.45 | 0.26109 a:0.0323/0.0334
144 | 0.0017 0.0006 12.3 | 0.0006 | 5.157 3.433 1.50 15.89 | 0.2287 0.815 1.41 | 0.26119 a:0.0323/0.0334
146 | 0.0016 0.0006 12.3 | 0.0007 | 5.158 3.434 1.50 15.90 | 0.2205 0.722 1.36 | 0.26128 a:0.0324/0.0334
148 | 0.0016 0.0006 12.3 | 0.0008 | 5.157 3.434 1.50 15.90 | 0.2286 0.691 1.35 | 0.26137 a:0.0324/0.0335
150 | 0.0016 0.0006 12.3 | 0.0006 | 5.158 3.434 1.50 15.90 | 0.2259 0.845 1.42 | 0.26146 a:0.0324/0.0335
152 | 0.0017 0.0006 12.3 | 0.0006 | 5.158 3.434 1.50 15.90 | 0.2295 0.757 1.38 | 0.26154 a:0.0324/0.0335
154 | 0.0016 0.0005 12.3 | 0.0006 | 5.159 3.435 1.50 15.90 | 0.2304 0.751 1.38 | 0.26162 a:0.0324/0.0335
156 | 0.0018 0.0005 12.3 | 0.0006 | 5.159 3.435 1.50 15.90 | 0.2264 0.796 1.40 | 0.26169 a:0.0324/0.0335
158 | 0.0017 0.0005 12.3 | 0.0006 | 5.160 3.434 1.50 15.90 | 0.2282 0.788 1.39 | 0.26176 a:0.0324/0.0335
160 | 0.0017 0.0005 12.3 | 0.0005 | 5.161 3.434 1.50 15.90 | 0.2291 0.766 1.38 | 0.26183 a:0.0324/0.0335
162 | 0.0016 0.0005 12.3 | 0.0005 | 5.161 3.434 1.50 15.90 | 0.2282 0.716 1.36 | 0.26189 a:0.0324/0.0335
164 | 0.0016 0.0005 12.3 | 0.0005 | 5.161 3.435 1.50 15.90 | 0.2344 0.792 1.40 | 0.26196 a:0.0324/0.0335
166 | 0.0016 0.0005 12.3 | 0.0006 | 5.162 3.434 1.50 15.90 | 0.2305 0.707 1.35 | 0.26202 a:0.0324/0.0335
168 | 0.0016 0.0005 12.3 | 0.0005 | 5.162 3.434 1.50 15.90 | 0.2353 0.816 1.41 | 0.26207 a:0.0324/0.0335
170 | 0.0016 0.0005 12.3 | 0.0005 | 5.163 3.434 1.50 15.90 | 0.2296 0.756 1.38 | 0.26213 a:0.0325/0.0335
172 | 0.0018 0.0005 12.4 | 0.0005 | 5.163 3.434 1.50 15.90 | 0.2391 0.742 1.37 | 0.26218 a:0.0325/0.0336
174 | 0.0016 0.0005 12.4 | 0.0005 | 5.163 3.434 1.50 15.90 | 0.2307 0.863 1.43 | 0.26224 a:0.0325/0.0336
176 | 0.0016 0.0005 12.3 | 0.0005 | 5.163 3.434 1.50 15.90 | 0.2329 0.854 1.43 | 0.26228 a:0.0325/0.0336
178 | 0.0017 0.0005 12.3 | 0.0005 | 5.164 3.434 1.50 15.90 | 0.2287 0.803 1.40 | 0.26233 a:0.0325/0.0336
180 | 0.0017 0.0005 12.3 | 0.0005 | 5.164 3.434 1.50 15.90 | 0.2361 0.819 1.41 | 0.26237 a:0.0325/0.0336
182 | 0.0018 0.0005 12.3 | 0.0005 | 5.164 3.434 1.50 15.90 | 0.2360 0.729 1.36 | 0.26241 a:0.0325/0.0336
184 | 0.0016 0.0005 12.3 | 0.0005 | 5.164 3.434 1.50 15.90 | 0.2395 0.774 1.39 | 0.26245 a:0.0325/0.0336
186 | 0.0015 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2325 0.777 1.39 | 0.26248 a:0.0325/0.0336
188 | 0.0018 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2430 0.669 1.33 | 0.26250 a:0.0325/0.0336
190 | 0.0016 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2293 0.781 1.39 | 0.26252 a:0.0325/0.0336
192 | 0.0017 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2246 0.763 1.38 | 0.26254 a:0.0325/0.0336
194 | 0.0019 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2324 0.764 1.38 | 0.26255 a:0.0325/0.0336
196 | 0.0016 0.0005 12.2 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2301 0.868 1.43 | 0.26256 a:0.0325/0.0336
198 | 0.0016 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2332 0.696 1.35 | 0.26256 a:0.0325/0.0336
200 | 0.0017 0.0005 12.3 | 0.0005 | 5.165 3.434 1.50 15.90 | 0.2397 0.808 1.40 | 0.26256 a:0.0325/0.0336
==========================================================================================
FINAL ANALYSIS
==========================================================================================
PatchSVAE: 16 patches × (256, 16)
Target CV: 0.125
Recon MSE: 0.000489 +/- 0.000743
Row CV: 0.2397
Cross-attention S delta: 0.26256
Learned alpha per mode (coordination strength):
Layer 0: mean=0.0327 max=0.0336 min=0.0323
α[ 0]: 0.0324 ######################################
α[ 1]: 0.0323 ######################################
α[ 2]: 0.0327 ######################################
α[ 3]: 0.0326 ######################################
α[ 4]: 0.0325 ######################################
α[ 5]: 0.0326 ######################################
α[ 6]: 0.0332 #######################################
α[ 7]: 0.0336 #######################################
α[ 8]: 0.0326 ######################################
α[ 9]: 0.0324 ######################################
α[10]: 0.0326 ######################################
α[11]: 0.0325 ######################################
α[12]: 0.0328 #######################################
α[13]: 0.0324 ######################################
α[14]: 0.0331 #######################################
α[15]: 0.0326 ######################################
Layer 1: mean=0.0323 max=0.0327 min=0.0315
α[ 0]: 0.0324 #######################################
α[ 1]: 0.0326 #######################################
α[ 2]: 0.0323 #######################################
α[ 3]: 0.0326 #######################################
α[ 4]: 0.0327 #######################################
α[ 5]: 0.0326 #######################################
α[ 6]: 0.0320 #######################################
α[ 7]: 0.0315 ######################################
α[ 8]: 0.0324 #######################################
α[ 9]: 0.0327 #######################################
α[10]: 0.0325 #######################################
α[11]: 0.0324 #######################################
α[12]: 0.0321 #######################################
α[13]: 0.0324 #######################################
α[14]: 0.0317 ######################################
α[15]: 0.0322 #######################################
Coordinated singular value profile:
S[ 0]: 5.1650 cum= 9.2% #############################
S[ 1]: 4.9525 cum= 17.6% ############################
S[ 2]: 4.8142 cum= 25.6% ###########################
S[ 3]: 4.6335 cum= 32.9% ##########################
S[ 4]: 4.5199 cum= 40.0% ##########################
S[ 5]: 4.4203 cum= 46.7% #########################
S[ 6]: 4.3376 cum= 53.2% #########################
S[ 7]: 4.2448 cum= 59.3% ########################
S[ 8]: 4.1641 cum= 65.3% ########################
S[ 9]: 4.0915 cum= 71.1% #######################
S[10]: 4.0086 cum= 76.6% #######################
S[11]: 3.9144 cum= 81.9% ######################
S[12]: 3.7995 cum= 86.8% ######################
S[13]: 3.6926 cum= 91.5% #####################
S[14]: 3.5949 cum= 95.9% ####################
S[15]: 3.4336 cum=100.0% ###################
Saving reconstruction grid...
Saved to /content/svae_patch_recon.png
4/7/2025 Prototype V10.3 - Patch16 - The VIT size.
Might need some tweaks but I don't think so. We're approaching actual vit prototype accuracy now.
Lets see how the SVAE performs.
Prototype V10.2 - Patch32 - Patchwork Cross-Attention with Edge Smoothing
This eliminates the edge cutting of the last version, and in the process the recon accuracy has gone up.
Model still escapes the discharge within 2 epochs and has robust recon.
Prototype V10.1 Patchwork Cross-Attention - Stabilized
The patchwork has stabilized, and the output is more accurate than the original now that it supports SVD 32 with more accuracy and higher speed
Epoch 28 hit the unstable point, but the gradient clipped attention was the ticket that ensured solidity.
The discharge recovered immediately.
Give or take 97% accurate recall, lets get those numbers up before we move onto more powerful image sets. Roughly 28m params.
174 | 0.0471 0.0315 8.0 | 0.0318 | 4.038 2.092 1.93 31.52 | 0.1271 0.995 1.50 | 0.26864 a:0.0471/0.0476
176 | 0.0471 0.0314 7.9 | 0.0318 | 4.038 2.092 1.93 31.52 | 0.1343 1.000 1.50 | 0.26874 a:0.0471/0.0476
178 | 0.0471 0.0314 7.9 | 0.0318 | 4.038 2.093 1.93 31.52 | 0.1313 0.995 1.50 | 0.26883 a:0.0471/0.0477
180 | 0.0471 0.0314 8.0 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1312 1.000 1.50 | 0.26892 a:0.0471/0.0477
182 | 0.0471 0.0314 7.9 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1310 0.995 1.50 | 0.26899 a:0.0471/0.0477
184 | 0.0471 0.0314 7.9 | 0.0317 | 4.038 2.092 1.93 31.52 | 0.1350 0.993 1.50 | 0.26906 a:0.0472/0.0477
186 | 0.0470 0.0314 7.9 | 0.0317 | 4.038 2.092 1.93 31.52 | 0.1338 1.000 1.50 | 0.26911 a:0.0472/0.0477
188 | 0.0470 0.0314 8.0 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1305 0.999 1.50 | 0.26916 a:0.0472/0.0477
190 | 0.0470 0.0314 8.0 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1358 0.999 1.50 | 0.26919 a:0.0472/0.0477
192 | 0.0470 0.0314 8.0 | 0.0317 | 4.038 2.092 1.93 31.52 | 0.1354 0.999 1.50 | 0.26922 a:0.0472/0.0477
194 | 0.0470 0.0314 7.9 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1301 0.992 1.50 | 0.26923 a:0.0472/0.0477
196 | 0.0470 0.0314 7.9 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1330 0.998 1.50 | 0.26924 a:0.0472/0.0477
198 | 0.0470 0.0314 7.9 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1312 1.000 1.50 | 0.26925 a:0.0472/0.0477
200 | 0.0470 0.0314 7.9 | 0.0317 | 4.038 2.093 1.93 31.52 | 0.1300 1.000 1.50 | 0.26925 a:0.0472/0.0477
==========================================================================================
FINAL ANALYSIS
==========================================================================================
PatchSVAE: 4 patches × (256, 32)
Target CV: 0.125
Recon MSE: 0.031701 +/- 0.024789
Row CV: 0.1300
Cross-attention S delta: 0.26925
Learned alpha per mode (coordination strength):
Layer 0: mean=0.0471 max=0.0477 min=0.0466
α[ 0]: 0.0470 #######################################
α[ 1]: 0.0473 #######################################
α[ 2]: 0.0474 #######################################
α[ 3]: 0.0473 #######################################
α[ 4]: 0.0471 #######################################
α[ 5]: 0.0474 #######################################
α[ 6]: 0.0469 #######################################
α[ 7]: 0.0472 #######################################
α[ 8]: 0.0470 #######################################
α[ 9]: 0.0475 #######################################
α[10]: 0.0467 #######################################
α[11]: 0.0471 #######################################
α[12]: 0.0477 #######################################
α[13]: 0.0466 #######################################
α[14]: 0.0471 #######################################
α[15]: 0.0472 #######################################
α[16]: 0.0472 #######################################
α[17]: 0.0471 #######################################
α[18]: 0.0470 #######################################
α[19]: 0.0475 #######################################
α[20]: 0.0466 #######################################
α[21]: 0.0477 #######################################
α[22]: 0.0470 #######################################
α[23]: 0.0469 #######################################
α[24]: 0.0472 #######################################
α[25]: 0.0472 #######################################
α[26]: 0.0471 #######################################
α[27]: 0.0471 #######################################
α[28]: 0.0474 #######################################
α[29]: 0.0472 #######################################
α[30]: 0.0466 #######################################
α[31]: 0.0475 #######################################
Layer 1: mean=0.0472 max=0.0477 min=0.0466
α[ 0]: 0.0474 #######################################
α[ 1]: 0.0472 #######################################
α[ 2]: 0.0470 #######################################
α[ 3]: 0.0471 #######################################
α[ 4]: 0.0474 #######################################
α[ 5]: 0.0470 #######################################
α[ 6]: 0.0474 #######################################
α[ 7]: 0.0473 #######################################
α[ 8]: 0.0473 #######################################
α[ 9]: 0.0470 #######################################
α[10]: 0.0477 #######################################
α[11]: 0.0472 #######################################
α[12]: 0.0466 #######################################
α[13]: 0.0477 #######################################
α[14]: 0.0473 #######################################
α[15]: 0.0471 #######################################
α[16]: 0.0472 #######################################
α[17]: 0.0471 #######################################
α[18]: 0.0476 #######################################
α[19]: 0.0470 #######################################
α[20]: 0.0475 #######################################
α[21]: 0.0470 #######################################
α[22]: 0.0472 #######################################
α[23]: 0.0475 #######################################
α[24]: 0.0472 #######################################
α[25]: 0.0471 #######################################
α[26]: 0.0475 #######################################
α[27]: 0.0474 #######################################
α[28]: 0.0472 #######################################
α[29]: 0.0469 #######################################
α[30]: 0.0476 #######################################
α[31]: 0.0466 #######################################
Coordinated singular value profile:
S[ 0]: 4.0376 cum= 5.3% #############################
S[ 1]: 3.9321 cum= 10.3% #############################
S[ 2]: 3.8501 cum= 15.1% ############################
S[ 3]: 3.7785 cum= 19.8% ############################
S[ 4]: 3.7092 cum= 24.2% ###########################
S[ 5]: 3.6414 cum= 28.5% ###########################
S[ 6]: 3.5771 cum= 32.7% ##########################
S[ 7]: 3.5158 cum= 36.7% ##########################
S[ 8]: 3.4554 cum= 40.6% #########################
S[ 9]: 3.3961 cum= 44.3% #########################
S[10]: 3.3371 cum= 48.0% ########################
S[11]: 3.2788 cum= 51.5% ########################
S[12]: 3.2230 cum= 54.8% #######################
S[13]: 3.1681 cum= 58.1% #######################
S[14]: 3.1141 cum= 61.2% #######################
S[15]: 3.0607 cum= 64.3% ######################
S[16]: 3.0088 cum= 67.2% ######################
S[17]: 2.9568 cum= 70.1% #####################
S[18]: 2.9075 cum= 72.8% #####################
S[19]: 2.8572 cum= 75.5% #####################
S[20]: 2.8067 cum= 78.0% ####################
S[21]: 2.7584 cum= 80.5% ####################
S[22]: 2.7075 cum= 82.9% ####################
S[23]: 2.6574 cum= 85.2% ###################
S[24]: 2.6060 cum= 87.4% ###################
S[25]: 2.5535 cum= 89.5% ##################
S[26]: 2.4991 cum= 91.5% ##################
S[27]: 2.4413 cum= 93.5% ##################
S[28]: 2.3770 cum= 95.3% #################
S[29]: 2.2906 cum= 97.0% #################
S[30]: 2.2012 cum= 98.6% ################
S[31]: 2.0926 cum=100.0% ###############
Saving reconstruction grid...
Saved to /content/svae_patch_recon.png
Prototype V10 Patchwork Cross-Attention - Unstable
Tiny Imagenet can't draw enough information from a single monotonic MLP projection, so I'm breaking the structure into quadrant-based mlp patches with cross-attention for a prototype.
Each patch is 32x32 and they have svd 24 independently represented each with patchwork cross-attention. Similar to a vit, so I'm building it to a full vit structure over time to ensure solidity and solidarity.
Current proto is more stable but requires a bit more oomph.
The CV is enjoying it's drift a BIT too much
I'll try attention alpha rather than rigid alpha. 4 patches is a bit unstable, so lets get some stability.
Prototype V9 prod
Should run on colab. Install the necessary repos.
https://huggingface.co/AbstractPhil/geolip-SVAE/blob/main/prototype_v9_prod.py
Prototype V8 Soft Hand Loss
Stable prototype found. Scaling with the CV ratio within this band is a stable attractor to the structural response.
The soft hand loss is acting like a stable attractant. Correct utilization of this behavior can directly attenuate a model's structural internals to align to certain trajectory-based routes.
The alignment can be directly tuned at runtime, shifted to learn implicit rules, altered to teach specific behaviors, and more.
0.034 mse, which is a different gauge of loss entirely.

Prototype V7
Normalized spherical without magnitude, expected considerably faster with less accuracy at first stages.
What happens if you train with the wrong CV value?
Using geolip-core SVD (Gram + eigh)
SVAE - V=96, D=24 (Validated: CV=0.3668)
Matrix: (96, 24) = 2304 elements
SVD: geolip-core Gram+eigh
Losses: recon + CV(w=0.1, target=0.3668)
Params: 6,036,736
=====================================================================================
ep | loss recon cv_l t/ep | t_rec | S0 SD ratio erank | row_cv
-------------------------------------------------------------------------------------
1 | 0.4174 0.4169 0.0037 7.3 | 0.2843 | 5.39 1.977 2.73 23.15 | 0.3039
2 | 0.2492 0.2489 0.0031 7.3 | 0.2286 | 5.43 1.978 2.75 23.14 | 0.3148
3 | 0.2096 0.2093 0.0030 7.3 | 0.1946 | 5.50 1.982 2.77 23.13 | 0.3352
4 | 0.1858 0.1855 0.0021 7.2 | 0.1812 | 5.48 1.980 2.77 23.13 | 0.3460
6 | 0.1586 0.1581 0.0046 7.3 | 0.1541 | 5.31 1.873 2.83 23.09 | 0.3938
8 | 0.1419 0.1407 0.0096 7.3 | 0.1377 | 5.33 1.815 2.93 23.03 | 0.4565
10 | 0.1314 0.1283 0.0385 7.3 | 0.1279 | 5.42 1.778 3.05 22.97 | 0.5373
12 | 0.1226 0.1160 0.0599 7.2 | 0.1162 | 5.67 1.738 3.26 22.86 | 0.6060
14 | 0.1189 0.1087 0.0847 7.1 | 0.1109 | 5.78 1.705 3.39 22.79 | 0.6643
16 | 0.1175 0.1014 0.1935 7.1 | 0.0996 | 6.17 1.701 3.63 22.67 | 0.7598
18 | 0.1170 0.0952 0.2238 7.2 | 0.0974 | 6.50 1.671 3.89 22.52 | 0.8211
20 | 0.1173 0.0905 0.1539 7.2 | 0.0907 | 6.69 1.649 4.06 22.43 | 0.8383
22 | 0.1200 0.0852 0.3335 7.1 | 0.0903 | 7.11 1.655 4.30 22.30 | 0.9128
24 | 0.1233 0.0817 0.2770 7.2 | 0.0831 | 7.51 1.646 4.56 22.15 | 0.9654
26 | 0.1286 0.0785 0.3243 7.1 | 0.0778 | 7.71 1.646 4.68 22.09 | 1.0196
28 | 0.1328 0.0752 0.4244 7.2 | 0.0780 | 7.84 1.636 4.80 22.02 | 1.1002
30 | 0.1373 0.0726 0.8786 7.1 | 0.0752 | 8.24 1.631 5.05 21.87 | 1.1243
32 | 0.1437 0.0703 0.6946 7.2 | 0.0704 | 8.52 1.631 5.23 21.76 | 1.2061
34 | 0.6025 0.6020 0.0062 7.1 | 0.5194 | 28.25 10.261 2.75 23.14 | 0.2935
36 | 0.4995 0.4990 0.0062 7.2 | 0.4949 | 29.82 10.939 2.73 23.15 | 0.2982
38 | 0.4947 0.4942 0.0058 7.2 | 0.4915 | 28.37 10.433 2.72 23.15 | 0.2988
40 | 0.4579 0.4574 0.0053 7.2 | 0.4557 | 26.31 9.585 2.74 23.14 | 0.3041
42 | 0.4333 0.4328 0.0051 7.1 | 0.4259 | 22.03 7.996 2.75 23.14 | 0.2984
44 | 0.4057 0.4054 0.0038 7.1 | 0.3880 | 21.15 7.656 2.76 23.15 | 0.3177
46 | 0.3670 0.3667 0.0024 7.2 | 0.3634 | 19.33 6.943 2.78 23.13 | 0.3280
48 | 0.3495 0.3493 0.0005 7.1 | 0.3457 | 18.34 6.569 2.79 23.13 | 0.3336
50 | 0.3341 0.3340 0.0010 7.3 | 0.3326 | 17.55 6.298 2.79 23.13 | 0.3424
52 | 0.3205 0.3204 0.0003 7.2 | 0.3182 | 16.94 6.069 2.79 23.12 | 0.3549
SNAP. right there at epoch 34. The tension was too strong, the model simply snapped. I had it set to around 0.366, and it requires that value there where it snapped to. 0.2935
The actual value as of the bulk embedding tests show; CV=0.2992 is the stable attractor, almost precisely where the model snapped to.
The effect was so strong, that the entire model had a forced reset when it realized the fundamental invalidity.
Why? I don't know yet.
Models
V4 5m SVD+EIGH 100 epochs 48x24
V4 111m 200x24 SVD+EIGH KL_DIV - Undercooked, needs more epochs -> sequel faulty, collapse
V3 v1024 - SVD 24
V2 16 modes












