I have created a dataset which has rather large number of features for example-100,000. Is it too large for a decent computer to handle ( I have a 1080ti )?
1 Answers
It highly depends on your data. If it's image, I guess it is somehow logical but if not I recommend you constructing covariance matrix and tracking whether features have correlation or not. If you see many features are correlated, it is better to discard correlated features. You also can employ PCA
to do this. Correlated features cause larger number of parameters for neural network.
Also I have to say that maybe you can reduce the number of parameters if your inputs are images by resizing them. In popular nets the length and height of input images are usually less than three hundred which makes the number of input features 90000
. Also you can employ max-pooling after some convolution layers, if you are using convolutional nets, to reduce the number of parameters. Refer here which maybe helpful.

- 14,058
- 9
- 57
- 98
-
Considering 3 values for R G and B it will be 270000 instead of 90000 – Afshin Amiri Jul 14 '18 at 15:06
-
226x226x3
, which is 153k 32-bit features per instance. Of course, the batch size has to be pretty small, but it is possible to train a CNN on a 1080ti. – Maxim Jan 12 '18 at 20:49