Remove 2017 Remove Blog Remove Data mining Remove Risk
article thumbnail

The trinity of errors in financial models: An introductory analysis using TensorFlow Probability

O'Reilly on Data

In this blog post, we explore three types of errors inherent in all financial models, with a simple example of a model in TensorFlow Probability (TFP). They trade the markets using quantitative models based on non-financial theories such as information theory, data science, and machine learning. Finance is not physics.

Modeling 134
article thumbnail

Transforming Credit and Collection with Predictive Analytics

BizAcuity

is delinquent as of June 30th, 2017. With Big Data, it is possible to acquire and segregate data with laser sharp focus with respect to one singular debtor. By clubbing various techniques like data mining, machine learning, artificial intelligence and statistical modelling, it makes predictions about events in the future.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

ML internals: Synthetic Minority Oversampling (SMOTE) Technique

Domino Data Lab

This carries the risk of this modification performing worse than simpler approaches like majority under-sampling. note that this variant “performs worse than plain under-sampling based on AUC” when tested on the Adult dataset (Dua & Graff, 2017). Data mining for direct marketing: Problems and solutions. Chawla et al.

article thumbnail

Changing assignment weights with time-based confounders

The Unofficial Google Data Science Blog

One reason to do ramp-up is to mitigate the risk of never before seen arms. A ramp-up strategy may mitigate the risk of upsetting the site’s loyal users who perhaps have strong preferences for the current statistics that are shown. Proceedings of the 13th ACM SIGKDD international conference on Knowledge discovery and data mining.

article thumbnail

Towards optimal experimentation in online systems

The Unofficial Google Data Science Blog

This blog post discusses such a comprehensive approach that is used at Youtube. Risk and Robustness Our estimates $widehat{beta}$ of the "true'' coefficients $beta$ of our model (1) depend on the random data we observe in experiments, and they are therefore random or uncertain.

article thumbnail

Explaining black-box models using attribute importance, PDPs, and LIME

Domino Data Lab

For this demo we’ll use the freely available Statlog (German Credit Data) Data Set, which can be downloaded from Kaggle. This dataset classifies customers based on a set of attributes into two credit risk groups – good or bad. Conference on Knowledge Discovery and Data Mining, pp. References.

Modeling 139
article thumbnail

Top 10 Analytics And Business Intelligence Trends For 2020

datapine

This is one of the major trends chosen by Gartner in their 2020 Strategic Technology Trends report , combining AI with autonomous things and hyperautomation, and concentrating on the level of security in which AI risks of developing vulnerable points of attacks. It’s an extension of data mining which refers only to past data.