Dimension Reduction is a technique that involves reducing the number of variables or features in a dataset while preserving its essential information. The goal is to simplify the dataset, making it more manageable and potentially improving the performance of certain algorithms. Dimension reduction is particularly useful when dealing with datasets with a large number of features, as it can help mitigate the “curse of dimensionality” and improve computational efficiency.

Posts

Improving neural network robustness through neighborhood preserving layers

One major source of vulnerability of neural nets in classification tasks is from overparameterized fully connected layers near the end of the network. In this paper, we propose a new neighborhood preserving layer which can replace these fully connected layers to improve the network robustness. Networks including these neighborhood preserving layers can be trained efficiently. We theoretically prove that our proposed layers are more robust against distortion because they effectively control the magnitude of gradients. Finally, we empirically show that networks with our proposed layers are more robust against state-of-the-art gradient descent-based attacks, such as a PGD attack on the benchmark image classification datasets MNIST and CIFAR10.