Show HN: AI Accel,Tension-based pruning framework(40% sparsity, 1.5-2x speedups)
wwes369 Friday, December 26, 2025I built a PyTorch framework that achieves ~40% effective parameter reduction with 1.58x training and 2.05x inference speedups on mid-sized models—while keeping accuracy almost intact.Key ideas: Dynamic tension thresholds aggressively prune low-importance weights (with rollback for stability) Vibration-based deferral skips low-signal computations Entropy scheduling + sparse conversion for hardware gains
It's a drop-in replacement for nn.Linear (CurvatureTuner) and works out-of-the-box on MLPs. Planning Transformer tests next.Benchmark on ~400k param MLP (synthetic data, 5 epochs): Baseline: Train 2.45s / Inf 0.0045s Enhanced: Eff params ~281k (40% reduction) / Train 1.55s (1.58x) / Inf 0.0022s (2.05x)
Repo (MIT): https://github.com/wwes4/AI_Accel_1.5x Feedback, forks, and real-dataset tests very welcome! Inspired by unconventional efficiency ideas.