Posts
All the articles I've posted.
SNN视频流任务调研
Updated: at 15:06Published: at 13:42学习一下视频stream上任务的一些工作,大概计划一下后续的工作。
SpikeZIP-TF: Conversion is All You Need for Transformer-based SNN
Updated: at 15:06Published: at 14:19游康师兄的工作,ANN2SNN的Transformer。
SpikingJelly: An open-source machine learning infrastructure platform for spike-based intelligence
Updated: at 15:06Published: at 19:49北大惊蛰,非常有影响力的SNN框架,实现了从数据编码、数据集整合到训练、硬件部署的全流程,SNN的torch级别的工作。发表在Science Advanced上。
I-LLM: Efficient Integer-Only Inference for Fully-Quantized Low-Bit Large Language Models
Updated: at 15:06Published: at 17:11LLM的Interger-Only PTQ量化工作。
程序语言理论笔记
Updated: at 15:06Published: at 15:12程序语言理论课程的复习笔记。
The Minimum Equivalent DNF Problem and Shortest Implicants
Updated: at 15:06Published: at 06:55证明MIN-DNF问题是完全的
I-ViT: Integer-only Quantization for Efficient Vision Transformer Inference
Updated: at 15:06Published: at 15:56对ViT的纯整型量化,W8A8,中科院2023 ICCV
Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Updated: at 15:06Published: at 16:28EAGL,声称只要用CPU在3秒内就能完成对ResNet的量化,效率远高于HAWQ等其他传统的方法
Towards spike-based machine intelligence with neuromorphic computing
Updated: at 15:06Published: at 18:43Nature上关于SNN的综述
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Updated: at 15:06Published: at 13:27Flash Attention,利用硬件结构加速Attention计算速度、减少内存占用的算法。核心是Tiling,Online Softmax和Kernel Fusion。