kFsN
kFsN, short for k-Feature Sparse Network, is a class of artificial neural networks designed to impose structured sparsity on hidden representations. In a kFsN, each layer outputs a sparse vector by keeping only the k features with the highest activation values for each input example; the remaining activations are set to zero. The value of k is a hyperparameter and can be fixed or scheduled during training.
Implementation approaches include hard top-k masking, where a binary mask is applied to the activations, and
The concept arises from broader efforts to reduce computation and memory requirements in neural networks while
Advantages of kFsN include reduced arithmetic operations and memory usage at inference time, potential improvements in
Related concepts encompass sparse neural networks, pruning methods, and activation-based sparsification techniques such as top-k pooling