Introducing the Model Zoo Paradigm: From Rapid Prototyping to Systematic Asset Preservation

During the early exploratory phase of the ViT polynomial replacement project, my primary engineering objective was agility. At that time, I adopted a minimalistic, overwrite-based workflow: all training outputs were written directly into a temporary build directory. This design was a local optimum for that stage—it dramatically reduced engineering overhead and allowed me to validate mathematical hypotheses as quickly as possible. In that phase, model weights had a very short lifecycle; they served mainly as transient intermediates for validating code paths, rather than long-term assets worth preserving.

However, as the scope of the research expanded, my disk began to accumulate multiple key results produced under different strategies—from hybrid architectures on CIFAR-10 to fully polynomial replacements on CIFAR-100. At this point, the value proposition of model weights underwent a qualitative shift. These files were no longer disposable artifacts, but high-value assets representing substantial compute cost and tuning effort. The original practice of mixing all outputs in a single build directory began to expose clear management risks: mature SOTA models were intermingled with discarded debugging artifacts, making them difficult to distinguish and vulnerable to accidental overwriting by subsequent experiments.

To address this asset management bottleneck, I decided to adopt a well-established concept from the deep learning community and introduce a local Model Zoo architecture. The concept was first popularized by the Caffe framework [1] and later standardized through platforms such as Hugging Face [2]. Its core principle is the strict physical separation between the experimental workspace and archived assets. In my revised setup, the original build directory continues to function as a high-churn construction site, allowing frequent writes and cleanup, while the Model Zoo is implemented as a read-only persistent repository dedicated to hosting validated, high-value models.

When implementing this standard, it quickly became clear that storing binary weight files alone was insufficient. To ensure long-term usability and reproducibility, I established a strict local archiving protocol. Every model admitted into the Model Zoo must be packaged as an atomic unit containing complete contextual information. In addition to the trained weights, each package includes an immutable snapshot of the full configuration required to reproduce the model, along with a documentation file recording its final performance metrics. By forcibly binding parameters, recipes, and results together, this approach eliminates the risk of configuration–weight mismatches over time.

This architectural evolution resulted in a clear one-way workflow. Experiments iterate freely within the build directory; once a result is deemed valuable, it is immediately distilled and sealed into the Model Zoo via a scripted promotion step. This mechanism fundamentally changed my research experience. High-accuracy baseline models and experimental models operating under extreme arithmetic constraints are now safely isolated in an archival zone. This systematic certainty frees me from concerns about losing historical results and allows greater focus on future algorithmic development. Ultimately, this is not merely a reorganization of the file system, but a deliberate alignment with mature model management practices from industrial machine learning.


References

[1] Jia, Y., Shelhamer, E., Donahue, J., Karayev, S., Long, J., Girshick, R., et al. (2014). Caffe: Convolutional Architecture for Fast Feature Embedding. Proceedings of the 22nd ACM International Conference on Multimedia.
(The concept of the “Model Zoo” was popularized by the Caffe community to share pretrained models.)

[2] Wolf, T., Debut, L., Sanh, V., Chaumond, J., Delangue, C., Moi, A., et al. (2019). HuggingFace’s Transformers: State-of-the-art Natural Language Processing. arXiv:1910.03771.
(Hugging Face established the modern standard of coupling model weights with configuration files and model cards.)