{"name":"SymbioGPT-Gemma-Fused","version":"1.0.0","description":"SymbioGPT-10M with Gemma LoRA knowledge projected via PCA. The evolved LoRA adapter (r=44, PPL 61.4 on philosophy) was decomposed and blended into the 11.6M-param SymbioGPT architecture.","architecture":"4-organelle decoder (CausalConv + Monarch + LongConv + Attention) + OrganelleGate + SwiGLU + RoPE + RMSNorm","model":{"d_model":320,"n_layers":8,"n_heads":5,"context_length":256,"vocab_size":2000,"params":"11.1M"},"fusion":{"source":"Gemma-3-270M LoRA (symbiogenesis-evolved, r=44)","method":"PCA projection + blend","blend_alpha":0.3},"organelles":["causal_conv","monarch","long_conv","attention"],"endpoints":["/v1/models","/v1/chat/completions"],"features":["streaming","OpenAI-compatible","top-k","top-p"],"github":"https://github.com/DavinciDreams/SymbioGPT"}