Model Description
David Collective is a geometric-simplex deep learning system that distills Stable Diffusion 1.5's knowledge into an ultra-efficient pentachoron-based architecture. This model was continued from epoch 20 to epoch 105, achieving remarkable performance with full pattern supervision.
Architecture Highlights
- Geometric Foundation: Uses 5D pentachora (5-vertex simplices) instead of traditional attention
- Multi-Scale Learning: Extracts features from all 9 SD1.5 UNet blocks
- Crystal Navigation: 1000-class supervision (100 timesteps × 10 geometric patterns)
- Parameter Efficiency: Ultra-compact architecture with shared geometric structures
- Full Supervision: Every sample supervised by both timestep and geometric pattern
Training Details
Continuation Training:
- Starting epoch: 20
- Final epoch: 105
- Total prompts trained: 600,500~ samples, 120,500~ prompts
- All prompts included:
prompts_all_epochs.jsonlcontains every prompt with metadata - Dataset: Symbolic caption synthesis (complexity 1-5)
- Batch size: 128
- Learning rate: 1e-4 with cosine annealing
- Optimizer: AdamW (weight_decay=0.01)
Final Metrics (Epoch 105):
- Total Loss: 0.2923
- Timestep Accuracy: 66.98%
- Pattern Accuracy: 100.00%
- Full Accuracy: 66.98%
- Pattern Diversity: -0.221
Active Blocks
David learns from all 9 SD1.5 UNet blocks:
down_0,down_1,down_2,down_3: Coarse semantic featuresmid: Bottleneck representationsup_0,up_1,up_2,up_3: Fine reconstruction details
Loss Components
- Feature Similarity (0.5): Cosine similarity with teacher
- Rose Loss (0.3): Geometric alignment with crystal centroids
- Cross-Entropy (0.2): 1000-class classification
- Pattern Diversity (0.05): Encourages balanced pattern usage
Usage
Loading the Model
import torch
from geovocab2.train.model.core.david_diffusion import DavidCollective, DavidCollectiveConfig
from safetensors.torch import load_file
# Load configuration
config = DavidCollectiveConfig(
num_timestep_bins=100,
num_feature_patterns_per_timestep=10,
active_blocks=['down_0', 'down_1', 'down_2', 'down_3', 'mid', 'up_0', 'up_1', 'up_2', 'up_3'],
david_sharing_mode='fully_shared',
david_fusion_mode='deep_efficiency',
use_belly=True,
belly_expand=1.5
)
# Create model
model = DavidCollective(config)
# Load weights from safetensors
state_dict = load_file("model.safetensors")
model.load_state_dict(state_dict)
model.eval()
# Inference
with torch.no_grad():
outputs = model(teacher_features, timesteps)
Training Data
This model includes prompts_all_epochs.jsonl - every single prompt used during training with full metadata:
{"timestamp": "2025-10-27T01:30:00", "epoch": 21, "batch": 0, "global_step": 6250, "sample_idx": 0, "timestep": 453, "timestep_bin": 45, "prompt": "a woman wearing red dress, against mountain landscape"}
Total prompts: 120,500 approximately
You can use this to:
- Analyze training data distribution
- Reproduce training
- Study prompt complexity vs model performance
- Generate similar synthetic datasets
Technical Details
Crystal System
- Architecture: Pentachoron-based geometric deep learning
- Centroids: 100 timestep bins × 10 patterns = 1000 anchors
- Navigation: Samples assigned to nearest pattern within timestep bin
- Diversity: Regularization prevents mode collapse
Progressive Training
- Started with early blocks (down_0, down_1)
- Progressively activated all 9 blocks
- Each block warmed up for 2 epochs
Pattern Supervision
Unlike traditional timestep-only supervision, David learns:
- When (timestep bin 0-99)
- How (geometric pattern 0-9 within that bin)
- Combined (full 1000-class space)
This provides 10x finer-grained supervision of the diffusion process.
Training History
Trained continuously from epoch 20 to epoch 105. See metrics:
- Timestep accuracy improved from ~60.3% to 66.98%
- Pattern accuracy maintained at 100.00%
- Loss decreased from 0.3431 to 0.2923
Citation
@misc{david-collective-sd15,
title={David Collective: Geometric Deep Learning for Diffusion Distillation},
author={AbstractPhil},
year={2025},
publisher={HuggingFace},
howpublished={\url{https://huggingface.co/AbstractPhil/david-collective-sd15-geometric-distillation}}
}
License
MIT License - See repository for details.
Acknowledgments
Built on the geometric deep learning research by AbstractPhil, using:
- Stable Diffusion 1.5 (teacher model)
- Pentachoron-based geometric algebra
- Crystalline consciousness architectures
- Symbolic caption synthesis
For more information, visit the geovocab2 repository.
- Downloads last month
- 4