Machine Learning

Given two kernels of the feature representations $K=k(x,x)$ and $L=l(y,y)$, HSIC is defined as12 $$ …

Using different learning rates in different layers of our artificial neural network.

For numerical stability we can use the log-sum-exp trick to calculate some loss such as cross entropy

ELBO is an very important concept in variational methods

Ask for valid confidence: “Valid”: validate for test data, train data, or the generating …

Hierarchical Classification Problem Hierarchical classification labels involves hierarchical class …

Artificial neuron that separates the state space

The loss calculated on all the data points

The loss calculated on all the whole population

Tutorials on machine learning and data science productivity articles

Latent variable models brings us new insights on identifying the patterns of some sample data.

We can set the parameters in a for loop. We take some of the initialization methods from Lippe1. To …

Pandas Groupby Does Not Guarantee Unique Content in Groupby Columns, it also considers the …

The Gini impurity is a measurement of the impurity of a set.

The information is a measurement of the entropy of the dataset.

During feature engineering, we have to deal with missing values.