Category «Machine Learning»

Heteroscedasticity (Heteroskedasticity)

Heteroscedasticity refers to the condition in which the variability of a variable is unequal across the range of values of a second variable that predicts it. A scatterplot of these variables will often create a cone-like shape, as the scatter (or variability) of the dependent variable (DV) widens or narrows as the value of the independent …

Decision Tree

Decision Trees (DTs) are a non-parametric supervised learning method used for classification and regression. It works for both categorical and continuous input and output variables. It is a non-parametric algorithm. It uses the following algorithms. 1- ID3: ID3 uses Entropy and Information gain to construct a decision tree. 2- Gini Index (, if we select two items from a population …

KNN

KNN is another supervised classification technique, which makes no assumption of distribution of data. It is a non-parametric algorithm.It is the simplest technique of classification, the algorithm used for this technique is as: Algorithm Let m be the number of training data samples. Let p be an unknown point. Store the training samples in an …

CNN

CNN stands for convolutional neural network. It is mostly used for developed for object recognition tasks such as handwritten digit recognition. There are four types of layers in a Convolutional Neural Network: 1. Convolutional Layers. 2. Pooling Layers. 3. Fully-Connected Layers. 4.Dropout layers. CNN is just a function which operates on another function in the below …

Naive Bayes

The Naive Bayes Classifier technique is based on the so-called Bayesian theorem and is particularly suited when the dimensionality of the inputs is high. Naive Bayes is a classi cation algorithm for binary and multiclass classi cation problems. The technique is easiest to understand when described using binary or categorical input values. It is called naive Bayes or …

Gradient Descent

Gradient Descent Gradient Descent is an optimization algorithm that optimize the cost of the function.The goal is to continue to try different values for the coefficients, evaluate their cost and select new coefficients that have a slightly better (lower) cost. https://www.hackerearth.com/blog/machine-learning/3-types-gradient-descent-algorithms-small-large-data-sets/ https://medium.com/@zhaoyi0113/python-implementation-of-batch-gradient-descent-379fa19eb428 https://www.analyticsvidhya.com/blog/2017/03/introduction-to-gradient-descent-algorithm-along-its-variants/