Function selection can be applied to dummy variables
What is the best 2-class classifier for your application? [closed]
Would you like to improve this question? Update the question so it can be answered with facts and quotes by editing this post.
Closed 3 years ago.
- One classifier per answer
- Vote if you agree
- Reject / remove duplicates.
- Enter your application in the comment
- easily captures complicated structure / nonlinear relationship
- invariable to the scale of the variables
- No dummy variables need to be created for categorical predictors
- Not much variable selection is required
- relatively difficult to cover
- Fast and powerful on most data sets
- Almost no parameters to match
- processes both discrete / continuous functions
- The model is easy to interpret
- (Not really limited to binary classifications)
Regularized discriminant for monitored problems with noisy data
- Computationally efficient
- Robust against noise and outliers in data
- Both linear discriminant (LD) and quadratic discriminant (QD) classifiers can be obtained from the same implementation by changing the regularization parameters' [lambda, r] 'to' [1 0] 'for LD classifiers and' [0 0] 'Can be set for QD classifier - very useful for reference purposes.
- Model is easy to interpret and export
- Works well for sparse and 'wide' data sets where the class covariance matrices may not be well defined.
- An estimate of the rear class probability can be estimated for each sample by applying the softmax function to the discriminant values for each class.
Link to the original work by Friedman et al . From 1989 here. There are also very good explanations by Kuncheva in her book "Combining Pattern Classifiers".
- At least as accurate as HF in many applications
- Seamlessly includes missing values
- Var importance
- Partial dependency diagrams
- GBM versus randomForest in R: handles MUCH larger data sets
Gaussian Process Classifier - It gives probabilistic predictions (useful when the frequency of your operational relative classes is different from that in your training set, or when the false-positive / false-negative costs are unknown or variable). It also provides an estimate of the uncertainty in model predictions due to the uncertainty in "estimating the model" from a finite data set. The co-variance function corresponds to the kernel function in an SVM and can therefore also be applied directly to non-vectorial data (e.g. character strings or diagrams, etc.). The math framework is neat too (but don't use the Laplace approximation). Automated model selection by maximizing the marginal probability.
Essentially combines good properties of logistic regression and SVM.
L1-regulated logistic regression.
- It's computationally fast.
- It has an intuitive interpretation.
- There is only one easy-to-understand hyperparameter that can be automatically adjusted through cross-validation, which is often a good way to go.
- The coefficients are piecewise linear and their relationship to the hyperparameter is immediately and easily visible in a simple diagram.
- This is one of the less dubious methods of variable selection.
- It has a really cool name too.
- Toll roads in Colorado 1
- Will Fitbit make a tape device
- Is it really worth paying for financial coaching
- Why is alkane called paraffin
- What is the best golf training aid
- What will forever be impossible
- Why do Saenger sign multiple record labels?
- What are bungee chords
- What's the Best Ursa Jungle Build
- What restricts the freedom of mankind?
- How do I fight foreclosure
- What impact will Ethereum have
- Who invented the magnetic eyelashes?
- What exactly is horse starch?
- Why is the term unofficial princesses used
- Organizational culture can best be explained as an organization
- Can I use WiFi without a password?
- What is the scope of performance management
- When was the last tennis game
- What are you ashamed of
- Is tea tree oil good for acne
- What is a hotel reservation management system
- What engineers made ICs
- Can I find true love here