论文标题
准确且可解释的机器学习,用于透明的健康保险计划定价
Accurate and Interpretable Machine Learning for Transparent Pricing of Health Insurance Plans
论文作者
论文摘要
健康保险公司通过商业雇主赞助的健康计划覆盖美国人口的一半,并每年支付12万亿美元,以支付其成员的医疗费用。一家健康保险公司的精算师和承销商职务用于评估要承担哪些风险以及如何定价这些风险以确保组织的盈利能力。尽管贝叶斯分层模型是行业中估计风险的当前标准,但对机器学习的兴趣作为改进这些现有方法的一种方式正在增加。医疗保健分析公司Lumiata与美国一家大型健康保险公司进行了一项研究。我们评估了机器学习模型在下一个续签期内预测雇主群体每月每月每月成本成本的能力,尤其是那些将不到精算模型预测的成本不到95%的群体(具有“让步机会”的组)。我们开发了两个模型,一个个体患者级别和一个雇主组级模型,以预测基于1400万患者的人口,雇主群体的每位每月允许每年允许的每年允许的每年。我们的模型的执行时间比保险公司现有的定价模型好20%,并确定了84%的特许权机会。这项研究表明,机器学习系统在计算健康保险产品的准确而公平的价格上的应用,并分析了可解释的机器学习模型如何超越精算模型的预测准确性,同时保持可解释性。
Health insurance companies cover half of the United States population through commercial employer-sponsored health plans and pay 1.2 trillion US dollars every year to cover medical expenses for their members. The actuary and underwriter roles at a health insurance company serve to assess which risks to take on and how to price those risks to ensure profitability of the organization. While Bayesian hierarchical models are the current standard in the industry to estimate risk, interest in machine learning as a way to improve upon these existing methods is increasing. Lumiata, a healthcare analytics company, ran a study with a large health insurance company in the United States. We evaluated the ability of machine learning models to predict the per member per month cost of employer groups in their next renewal period, especially those groups who will cost less than 95\% of what an actuarial model predicts (groups with "concession opportunities"). We developed a sequence of two models, an individual patient-level and an employer-group-level model, to predict the annual per member per month allowed amount for employer groups, based on a population of 14 million patients. Our models performed 20\% better than the insurance carrier's existing pricing model, and identified 84\% of the concession opportunities. This study demonstrates the application of a machine learning system to compute an accurate and fair price for health insurance products and analyzes how explainable machine learning models can exceed actuarial models' predictive accuracy while maintaining interpretability.