Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.

Historically, probabilistic models for decision support have focused on discrimination, e.g., minimizing the ranking error of predicted outcomes. Unfortunately, these models ignore another important aspect, calibration, which indicates the magnitude of correctness of model predictions. Using discrim...

Full description

Bibliographic Details
Main Authors: Xiaoqian Jiang, Aditya Menon, Shuang Wang, Jihoon Kim, Lucila Ohno-Machado
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2012-01-01
Series:PLoS ONE
Online Access:http://europepmc.org/articles/PMC3490990?pdf=render
id doaj-702b048a508840cbac6cc616ec488eda
record_format Article
spelling doaj-702b048a508840cbac6cc616ec488eda2020-11-24T20:50:07ZengPublic Library of Science (PLoS)PLoS ONE1932-62032012-01-01711e4882310.1371/journal.pone.0048823Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.Xiaoqian JiangAditya MenonShuang WangJihoon KimLucila Ohno-MachadoHistorically, probabilistic models for decision support have focused on discrimination, e.g., minimizing the ranking error of predicted outcomes. Unfortunately, these models ignore another important aspect, calibration, which indicates the magnitude of correctness of model predictions. Using discrimination and calibration simultaneously can be helpful for many clinical decisions. We investigated tradeoffs between these goals, and developed a unified maximum-margin method to handle them jointly. Our approach called, Doubly Optimized Calibrated Support Vector Machine (DOC-SVM), concurrently optimizes two loss functions: the ridge regression loss and the hinge loss. Experiments using three breast cancer gene-expression datasets (i.e., GSE2034, GSE2990, and Chanrion's datasets) showed that our model generated more calibrated outputs when compared to other state-of-the-art models like Support Vector Machine (p=0.03, p=0.13, and p<0.001) and Logistic Regression (p=0.006, p=0.008, and p<0.001). DOC-SVM also demonstrated better discrimination (i.e., higher AUCs) when compared to Support Vector Machine (p=0.38, p=0.29, and p=0.047) and Logistic Regression (p=0.38, p=0.04, and p<0.0001). DOC-SVM produced a model that was better calibrated without sacrificing discrimination, and hence may be helpful in clinical decision making.http://europepmc.org/articles/PMC3490990?pdf=render
collection DOAJ
language English
format Article
sources DOAJ
author Xiaoqian Jiang
Aditya Menon
Shuang Wang
Jihoon Kim
Lucila Ohno-Machado
spellingShingle Xiaoqian Jiang
Aditya Menon
Shuang Wang
Jihoon Kim
Lucila Ohno-Machado
Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
PLoS ONE
author_facet Xiaoqian Jiang
Aditya Menon
Shuang Wang
Jihoon Kim
Lucila Ohno-Machado
author_sort Xiaoqian Jiang
title Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
title_short Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
title_full Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
title_fullStr Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
title_full_unstemmed Doubly Optimized Calibrated Support Vector Machine (DOC-SVM): an algorithm for joint optimization of discrimination and calibration.
title_sort doubly optimized calibrated support vector machine (doc-svm): an algorithm for joint optimization of discrimination and calibration.
publisher Public Library of Science (PLoS)
series PLoS ONE
issn 1932-6203
publishDate 2012-01-01
description Historically, probabilistic models for decision support have focused on discrimination, e.g., minimizing the ranking error of predicted outcomes. Unfortunately, these models ignore another important aspect, calibration, which indicates the magnitude of correctness of model predictions. Using discrimination and calibration simultaneously can be helpful for many clinical decisions. We investigated tradeoffs between these goals, and developed a unified maximum-margin method to handle them jointly. Our approach called, Doubly Optimized Calibrated Support Vector Machine (DOC-SVM), concurrently optimizes two loss functions: the ridge regression loss and the hinge loss. Experiments using three breast cancer gene-expression datasets (i.e., GSE2034, GSE2990, and Chanrion's datasets) showed that our model generated more calibrated outputs when compared to other state-of-the-art models like Support Vector Machine (p=0.03, p=0.13, and p<0.001) and Logistic Regression (p=0.006, p=0.008, and p<0.001). DOC-SVM also demonstrated better discrimination (i.e., higher AUCs) when compared to Support Vector Machine (p=0.38, p=0.29, and p=0.047) and Logistic Regression (p=0.38, p=0.04, and p<0.0001). DOC-SVM produced a model that was better calibrated without sacrificing discrimination, and hence may be helpful in clinical decision making.
url http://europepmc.org/articles/PMC3490990?pdf=render
work_keys_str_mv AT xiaoqianjiang doublyoptimizedcalibratedsupportvectormachinedocsvmanalgorithmforjointoptimizationofdiscriminationandcalibration
AT adityamenon doublyoptimizedcalibratedsupportvectormachinedocsvmanalgorithmforjointoptimizationofdiscriminationandcalibration
AT shuangwang doublyoptimizedcalibratedsupportvectormachinedocsvmanalgorithmforjointoptimizationofdiscriminationandcalibration
AT jihoonkim doublyoptimizedcalibratedsupportvectormachinedocsvmanalgorithmforjointoptimizationofdiscriminationandcalibration
AT lucilaohnomachado doublyoptimizedcalibratedsupportvectormachinedocsvmanalgorithmforjointoptimizationofdiscriminationandcalibration
_version_ 1716804689811472384