Skip to main content
Hima Varsha's user avatar
Hima Varsha's user avatar
Hima Varsha's user avatar
Hima Varsha
  • Member for 9 years, 5 months
  • Last seen more than 6 years ago
suggested
Approve
Loading…
Loading…
Loading…
comment
Naive Bayes: Divide by Zero error
Hey, I used it in python 2.7 and it seems to be working fine. Maybe some changes in the versions are giving rise to the warnings. Isn't the predictions working still?
asked
Loading…
comment
Picking training data
use train_test_split of sklearn, it shuffles and splits
comment
How to choose a classifier after cross-validation?
Oh sorry, yes - the one with the better test accuracy normally is what I use. But say your data is vary varied and you get very fluctuating validation accuracies, comparing the average of accuracies can also be tried.
comment
How to choose a classifier after cross-validation?
once we have used cross-validation to select the better performing model(for instance you have 2 models-linear regression or neural network), we train that model (whether it be the linear regression or the neural network) on all the data
comment
How to choose a classifier after cross-validation?
when you say test, you mean validation dataset's test right?
Loading…
Loading…
comment
Obtaining frequently occurring phrases using Word2Vec
do you want the frequency of the phrase in the document? If that is the case, you can just use a dictionary counter right?
comment
In general, when does TF-IDF reduce accuracy?
So when you say you don't use tf-idf, what are your features then? Just a simple BOG?
comment
In general, when does TF-IDF reduce accuracy?
Maybe the dataset you contain is vary varied and the frequencies of the words don't matter to come to the conclusion?
comment
Example of binary classifier with numerical features using deep learning
@Richard, sentiment analysis for 2 classes is also a well received binary classification problem, did you try it with your datasets?
Loading…
answered
Loading…
comment
What does Prec@1 in fastText mean?
Sorry for that. I mean prec@1 is essentially the precision i'll obtain if I use scikit's metrics library and calculate for multiclass classification?
1
6 7
8
9 10
14