BorisovAI

Blog

Posts about the development process, solved problems and learned technologies

Found 1 notesReset filters
Generalllm-analisis

How I Caught the Best Seed in Neural Network Search

Got up from the couch, coffee in hand, and realized: I need to find the optimal seed for LLM Analysis. The project demanded a breakthrough — the current baseline was giving 72.86% accuracy, and that wasn't good enough for production. The task seemed straightforward at first glance: test 20 different seeds, each generating its own model initialization. But beneath that simplicity lay an uncomfortable truth — each seed required roughly 100 minutes of computation. About 30 hours of pure runtime for the search. I launched *seed_search.py* and sent it to the background via nohup — let it work on its own while I handled everything else. The first result surprised me: **seed 1 showed 76.5% at the 200th checkpoint**, meaning a 3.64 percentage point improvement. Not revolutionary, but movement in the right direction. The script ran stably, results accumulating in *results_seed_search.json* with resume support — if the process crashed, just restart it and it would continue from where it left off. While the seeds were computing, I got to parallel work. Wrote *augment_problems.py*, which transformed 6,604 original problems into 39,582 variations — the foundation for model self-distillation. Simultaneously prepared *majority_voting.py* for voting between Orchestra and baseline, and *dual_orchestra.py* for a two-stage architecture with intermediate layers. The plan crystallized in my head. After seed search finishes (another three days), I will: 1. Analyze the distribution of 20 results and pick the best seed 2. Run majority voting on the best checkpoint 3. Build Dual Orchestra Stage 1, using the best seed as the foundation 4. Train self-distillation on 39K augmented problems The technology behind all this is simple but stubborn. Claude as the primary LLM — fast, accurate enough for analysis. Python for process orchestration, JavaScript somewhere in the neighboring services. But the main thing — it's patience and systematicity. In a month, if everything works out, this model will perform better. For now, I'm waiting for results, sipping cold coffee. **Fun fact:** Kafka and my black cat have one thing in common — both do only what they want and actively ignore instructions. 😄

Mar 20, 2026