Emerging Technologies and their Impact on Actuarial Science

Link:https://www.soa.org/resources/research-reports/2021/emerging-technologies-and-their-impact-on-actuarial-science/

Full report: https://www.soa.org/globalassets/assets/files/resources/research-report/2021/2021-emerging-technologies-report.pdf

Graphic:

Excerpt:

Technologies that have reached widespread adoption today:
o Dynamic Collaboration Tools – e.g., Microsoft Teams, Slack, Miro – Most companies are now using this
type of technology. Some are using the different functionalities (e.g., digital whiteboarding, project
management tools, etc.) more fully than others at this time.
• Technologies that are reaching early majority adoption today:
o Business Intelligence Tools (Data Visualization component) – e.g., Tableau, Power BI — Most
respondents have started their journey in using these tools, with many having implemented solutions.
While a few respondents are lagging in its adoption, some companies have scaled applications of this
technology to all actuaries. BI tools will change and accelerate the way actuaries diagnose results,
understand results, and communicate insights to stakeholders.
o ML/AI on structured data – e.g., R, Python – Most respondents have started their journey in using
these techniques, but the level of maturity varies widely. The average maturity is beyond the piloting
phase amongst our respondents. These are used for a wide range of applications in actuarial functions,
including pricing business, modeling demand, performing experience studies, predicting lapses to
support sales and marketing, producing individual claims reserves in P&C, supporting accelerated
underwriting and portfolio scoring on inforce blocks.
o Documentation Generators (Markdown) – e.g., R Markdown, Sphinx – Many respondents have started
using these tools, but maturity level varies widely. The average maturity for those who have started
amongst our respondents is beyond the piloting phase. As the use of R/Python becomes more prolific
amongst actuaries, the ability to simultaneously generate documentation and reports for developed
applications and processes will increase in importance.
o Low-Code ETL and Low-Code Programming — e.g., Alteryx, Azure Data Factory – Amongst respondents
who provided responses, most have started their journey in using these tools, but the level of maturity
varies widely. The average maturity is beyond the piloting phase with our respondents. Low-code ETL
tools will be useful where traditional ETL tools requiring IT support are not sufficient for business
needs (e.g., too difficult to learn quickly for users or reviewers, ad-hoc processes) or where IT is not
able to provision views of data quickly enough.
o Source Control Management – e.g., Git, SVN – A sizeable proportion of the respondents are currently
using these technologies. Amongst these respondents, solutions have already been implemented.
These technologies will become more important in the context of maintaining code quality for
programming-based models and tools such as those developed in R/Python. The value of the
technology will be further enhanced with the adoption of DevOps practices and tools, which blur the
lines between Development and Operations teams to accelerate the deployment of
applications/programs

Author(s):

Nicole Cervi, Deloitte
Arthur da Silva, FSA, ACIA, Deloitte
Paul Downes, FIA, FCIA, Deloitte
Marwah Khalid, Deloitte
Chenyi Liu, Deloitte
Prakash Rajgopal, Deloitte
Jean-Yves Rioux, FSA, CERA, FCIA, Deloitte
Thomas Smith, Deloitte
Yvonne Zhang, FSA, FCIA, Deloitte

Publication Date: SOA

Publication Site: October 2021

Interpretable Machine Learning: A Guide for Making Black Box Models Explainable

Link: https://christophm.github.io/interpretable-ml-book/

Graphic:

Excerpt:

Machine learning has great potential for improving products, processes and research. But computers usually do not explain their predictions which is a barrier to the adoption of machine learning. This book is about making machine learning models and their decisions interpretable.

After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME.

All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.

Author(s): Christoph Molnar

Publication Date: 2021-06-14

Publication Site: github

Idea Behind LIME and SHAP

Link: https://towardsdatascience.com/idea-behind-lime-and-shap-b603d35d34eb

Graphic:

Excerpt:

In machine learning, there has been a trade-off between model complexity and model performance. Complex machine learning models e.g. deep learning (that perform better than interpretable models e.g. linear regression) have been treated as black boxes. Research paper by Ribiero et al (2016) titled “Why Should I Trust You” aptly encapsulates the issue with ML black boxes. Model interpretability is a growing field of research. Please read here for the importance of machine interpretability. This blog discusses the idea behind LIME and SHAP.

Author(s): ashutosh nayak

Publication Date: 22 December 2019

Publication Site: Toward Data Science

Opinion: Amid a Pandemic, a Health Care Algorithm Shows Promise and Peril

Excerpt:

In the midst of the uncertainty, Epic, a private electronic health record giant and a key purveyor of American health data, accelerated the deployment of a clinical prediction tool called the Deterioration Index. Built with a type of artificial intelligence called machine learning and in use at some hospitals prior to the pandemic, the index is designed to help physicians decide when to move a patient into or out of intensive care, and is influenced by factors like breathing rate and blood potassium level. Epic had been tinkering with the index for years but expanded its use during the pandemic. At hundreds of hospitals, including those in which we both work, a Deterioration Index score is prominently displayed on the chart of every patient admitted to the hospital.

The Deterioration Index is poised to upend a key cultural practice in medicine: triage. Loosely speaking, triage is an act of determining how sick a patient is at any given moment to prioritize treatment and limited resources. In the past, physicians have performed this task by rapidly interpreting a patient’s vital signs, physical exam findings, test results, and other data points, using heuristics learned through years of on-the-job medical training.

Ostensibly, the core assumption of the Deterioration Index is that traditional triage can be augmented, or perhaps replaced entirely, by machine learning and big data. Indeed, a study of 392 Covid-19 patients admitted to Michigan Medicine that the index was moderately successful at discriminating between low-risk patients and those who were at high-risk of being transferred to an ICU, getting placed on a ventilator, or dying while admitted to the hospital. But last year’s hurried rollout of the Deterioration Index also sets a worrisome precedent, and it illustrates the potential for such decision-support tools to propagate biases in medicine and change the ways in which doctors think about their patients.

Author(s): VISHAL KHETPAL, NISHANT SHAH

Publication Date: 27 May 2021

Publication Site: Undark Magazine

Towards Explainability of Machine Learning Models in Insurance Pricing

Link: https://arxiv.org/abs/2003.10674

Paper: https://arxiv.org/pdf/2003.10674.pdf

Citation:


arXiv:2003.10674
 [q-fin.RM]

Graphic:

Abstract:

Machine learning methods have garnered increasing interest among actuaries in recent years. However, their adoption by practitioners has been limited, partly due to the lack of transparency of these methods, as compared to generalized linear models. In this paper, we discuss the need for model interpretability in property & casualty insurance ratemaking, propose a framework for explaining models, and present a case study to illustrate the framework.

Author(s): Kevin Kuo, Daniel Lupton

Publication Date: 24 March 2020

Publication Site: arXiv

Ethics and use of Data Sources for Underwriting ft. Neil Raden and Kevin Pledge -NSNA(Ep.4)

Video:

Description:

The video features Neil Raden who is the author of ethical use of AI for Actuaries. Alongside him , it features Kevin Pledge who is CEO of Acceptiv , FSA,FIA and chair of Innovation and Research Committee of SOA. We discuss about the issue of ethics and about the use of new data sources in the recent Emerging issues in Underwriting Survey Report by IfOA.

Authors: Harsh Jaitak, Kevin Pledge, Neil Raden

Publication Date: 17 March 2021

Publication Site: TBD Actuarial at YouTube