Network Architectures

TinyMind provides a range of neural network architectures, all as header-only C++ templates with both fixed-point and floating-point support.

Architecture Memory (Q8.8, trainable) Key Advantage
LSTM & GRU 952 / 808 bytes Sequential data, temporal patterns
Kolmogorov-Arnold Networks 1,192 bytes Learnable activation functions
Conv & Pooling Layers 1,825 bytes (1D pipeline) 1D time-series + 2D spectrogram/image feature extraction, MobileNet-style separable blocks
Linear Self-Attention ~6 KB (mid-range) Sequence dependency modeling without O(N^2)
FFT Layer 768 bytes (64-pt Q8.8) Frequency-domain feature extraction
Quantized Networks 128 bytes (packed binary) 32-64x weight compression

Table of contents


Back to top

Dan McLeran — danmcleran@gmail.com — MIT License

This site uses Just the Docs, a documentation theme for Jekyll.