What do the features given by a feature selection method mean in a binary classifier which has a cross validation accuracy of 0?

StackOverflow https://stackoverflow.com/questions/21560540

Pergunta

So I know that given a binary classifier, the farther away you are from an accuracy of 0.5 the better your classifier is. (I.e. A binary classifier that gets everything wrong can be converted to one which gets everything right by always inverting its decisions.)

However, I have an inner feature selection process, which provides me "good" features to use (I'm trying out recursive feature elimination, and another based on Spearman's rank correlation coefficient). Given that the classifier using these "good" features gets a cross validation accuracy of 0, can I still conclude that the features selected are useful and are predictive of the class in this binary prediction problem?

Foi útil?

Solução

To simplify, let's assume you're testing on some balanced set. Half the testing data is positive and half the testing data is negative.

I would say that something strange is happening that is flipping the sign of your decision. That classifier you're evaluating is very useful, but you would need to flip the decision it makes. You should probably check your code to make sure you're not flipping the class of the training data. Some libraries (LIBSVM for example) require that the first training example is from the positive class.

To summarize: It seems the features you're selecting are useful, but it seems you have a bug that is flipping the classes.

Licenciado em: CC-BY-SA com atribuição
Não afiliado a StackOverflow
scroll top