In: Computer Science
Advanced Database:
Every answer should be minimum if 4 to 5 lines
Underfitting:
A statistical model or a machine learning algorithm is said to have
underfitting when it cannot capture the underlying trend of the
data. (It’s just like trying to fit undersized pants!) Underfitting
destroys the accuracy of our machine learning model. Its occurrence
simply means that our model or the algorithm does not fit the data
well enough. It usually happens when we have less data to build an
accurate model and also when we try to build a linear model with a
non-linear data. In such cases the rules of the machine learning
model are too easy and flexible to be applied on such a minimal
data and therefore the model will probably make a lot of wrong
predictions. Underfitting can be avoided by using more data and
also reducing the features by feature selection.
Overfitting:
A statistical model is said to be overfitted, when we train it with
a lot of data (just like fitting ourselves in an oversized pants!).
When a model gets trained with so much of data, it starts learning
from the noise and inaccurate data entries in our data set. Then
the model does not categorize the data correctly, because of too
much of details and noise. The causes of overfitting are the
non-parametric and non-linear methods because these types of
machine learning algorithms have more freedom in building the model
based on the dataset and therefore they can really build
unrealistic models. A solution to avoid overfitting is using a
linear algorithm if we have linear data or using the parameters
like the maximal depth if we are using decision trees.
Decision tree pruning
Pruning is a technique in machine learning and search algorithms
that reduces the size of decision trees by removing sections of the
tree that provide little power to classify instances. Pruning
reduces the complexity of the final classifier, and hence improves
predictive accuracy by the reduction of overfitting.
Cross-validation
Cross validation (CV) is one of the technique used to test the
effectiveness of a machine learning models, it is also a
re-sampling procedure used to evaluate a model if we have a limited
data. To perform CV we need to keep aside a sample/portion of the
data on which is do not use to train the model, later us this
sample for testing/validating. There are many methods
it sometimes called rotation estimation or out-of-sample testing,
is any of various similar model validation techniques for assessing
how the results of a statistical analysis will generalize to an
independent data set.
The purpose of the activation function is to introduce non-linearity into the output of a neuron.
Activation function decides, whether a neuron should be activated or not by calculating weighted sum and further adding bias with it.
Most popular types of Activation functions -
Sigmoid or Logistic.
Tanh — Hyperbolic tangent.
ReLu -Rectified linear units.