No description
Train Qwen3-0.6B-Base (596M params) on 36K folksy proverb pairs using full SFT with HuggingFace TRL. 3 epochs, 11 min on RTX 4090. Results: train_loss=0.954, eval_loss=1.032, test_loss=1.031 Model checkpoint at folksy-model/final/ (not committed — 1.2 GB) Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com> |
||
|---|---|---|
| corpus | ||
| data | ||
| examples | ||
| schemas | ||
| scripts | ||
| .gitignore | ||
| CORPUS_GENERATION_SPEC.md | ||
| CORPUS_QUALITY_REVIEW.md | ||
| EVALUATION.md | ||
| folksy_generator.py | ||
| FOLKSY_GENERATOR_SPEC.md | ||
| GPU_TRAINING_REQUIREMENTS.md | ||
| GRAPH_ENHANCEMENT_SPEC.md | ||