article thumbnail

Changing assignment weights with time-based confounders

The Unofficial Google Data Science Blog

One reason to do ramp-up is to mitigate the risk of never before seen arms. A ramp-up strategy may mitigate the risk of upsetting the site’s loyal users who perhaps have strong preferences for the current statistics that are shown. For example, imagine a fantasy football site is considering displaying advanced player statistics.

article thumbnail

Explaining black-box models using attribute importance, PDPs, and LIME

Domino Data Lab

This dataset classifies customers based on a set of attributes into two credit risk groups – good or bad. This is to be expected, as there is no reason for a perfect 50:50 separation of the good vs. bad credit risk. but it generally relies on measuring the entropy in the change of predictions given a perturbation of a feature.

Modeling 139
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

ML internals: Synthetic Minority Oversampling (SMOTE) Technique

Domino Data Lab

This renders measures like classification accuracy meaningless. This carries the risk of this modification performing worse than simpler approaches like majority under-sampling. The use of multiple measurements in taxonomic problems. Chawla et al. Indeed, in the original paper Chawla et al. Machine Learning, 57–78. Quinlan, J.

article thumbnail

Variance and significance in large-scale online services

The Unofficial Google Data Science Blog

Well, it turns out that depending on what it cares to measure, an LSOS might not have enough data. The practical consequence of this is that we can’t afford to be sloppy about measuring statistical significance and confidence intervals. Being dimensionless, it is a simple measure of the variability of a (non-negative) random variable.

article thumbnail

Using Empirical Bayes to approximate posteriors for large "black box" estimators

The Unofficial Google Data Science Blog

Posteriors are useful to understand the system, measure accuracy, and make better decisions. Methods like the Poisson bootstrap can help us measure the variability of $t$, but don’t give us posteriors either, particularly since good high-dimensional estimators aren’t unbiased.

KDD 40