A personal point of view

There is a huge mismatch between what most companies need and what ML practitioners want to do

Some background context

I started my career in 2010 after a master in computer science. During my studies I got hooked by software engineering. It can be really fascinating! You get to build stuff out of nothing. You start from a virtual blank sheet and you can end up with an empire. But I think what really hooked me the most is this sense of personal growth that I got day after day as my craft improved.

Anyway, after 7 years I didn’t always feel that I was learning new things on a daily basis. I was still learning but clearly not like…

One of the key ingredients of CatBoost explained from the ground up

Photo by Chandler Cruttenden on Unsplash

CatBoost is a “relatively” new package developed by Yandex researchers. It is pretty popular right now, especially in Kaggle competitions where it generally outperforms other gradient tree boosting libraries.

Among other ingredients, one of the very cool feature of CatBoost is that it handles categorical variables out of the box (hence the name of the algorithm).

When using an implementation, it is important to really understand how it works under the hood. This is the goal of this article. We are going to take an in-depth look at this technique called Ordered Target Statistics. It is presumed that you have…

A friendly but rigorous explanation

Photo by Bill Jelen on Unsplash

This article is intended for either students trying to break into data science or professionals in need of a refresher on boosting and gradient boosting. There are already quite a lot of materials regarding this topic on the web but not many include graphical visualization of the learning process.

So we are going to start with the Boosting principle and end with Gradient Boosting. We are also going to understand the transition from the former to the latter. We will start with graphical intuitions and then explain the algorithm written with pseudo-code.


Gradient Tree Boosting is one of the best…

A complete tutorial

Photo by Peter Bond on Unsplash

In my last article, I wrote a detailed explanation of the Gaussian Mixture Model (GMM) and the way it is trained using the Expectation-Maximization (EM) algorithm. This time, I wanted to show that a mixture model is not necessarily a mixture of Gaussian densities. It can be a mixture of any distribution. In this example, we are going to use a mixture of multinomial distributions.

Also, the idea is, for once, not to solely focus on the mathematical and computer science aspects of a data science project but on the business side too. Therefore we are going to use a…

This article is an extension of “Gaussian Mixture Models and Expectation-Maximization (A full explanation)”. If you didn’t read it, this article might not be very useful.

The goal here is to derive the closed-form expressions necessary for the update of the parameters during the Maximization step of the EM algorithm applied to GMMs. This material was written as a separate article in order not to overload the main one.

Ok so recall that during the M-Step, we want to maximize the following lower bound with respect to Θ :

The lower bound is defined to be a concave function easy…

Photo by NASA on Unsplash

In the previous article, we described the Bayesian framework for linear regression and how we can use latent variables to reduce model complexity.

In this post, we will explain how latent variables can also be used to frame a classification problem, namely the Gaussian Mixture model (or GMM in short) that allows us to perform soft probabilistic clustering.

This model is classically trained by an optimization procedure named the Expectation-Maximization (or EM in short) for which we will have a thorough review. At the end of this article, we will also see why we do not use traditional optimization methods.

Over the past year, I have taken more and more interest in Bayesian statistics and probabilistic modeling. Along this journey, I have encountered the latent probabilistic models.

We will start by explaining the concept of a latent variable. And to properly understand its benefits we need to make sure that you are familiar with the Bayesian framework for linear regression and this is what this first article is about.

As a quick note, I am no expert in Bayesian statistics. But I wanted to share the current state of my knowledge because I know it can help some people (myself…

Adrien Biarnes

I am a machine learning engineer at Dailymotion. I love to learn and share my passion for data science — https://www.linkedin.com/in/adrien-biarnes-81975717

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store