Machine learning algorithms in insurance: solvency, textmining, anonymization and transparency.

Authors Publication date
2019
Publication type
Thesis
Summary In the summer of 2013, the term "Big Data" made its appearance and aroused strong interest among companies. This thesis studies the contribution of these methods to actuarial sciences. It addresses both theoretical and practical issues on high-potential topics such as textit{Optical Character Recognition} (OCR), text analysis, data anonymization or model interpretability. Starting with the application of machine learning methods in the calculation of economic capital, we then try to better illustrate the frontality that can exist between machine learning and statistics. Putting forward some advantages and different techniques, we then study the application of deep neural networks in the optical analysis of documents and text, once extracted. The use of complex methods and the implementation of the General Data Protection Regulation (GDPR) in 2018 led us to study the potential impacts on pricing models. By applying anonymization methods on pure premium calculation models in non-life insurance, we explored different generalization approaches based on unsupervised learning. Finally, as regulations also impose criteria in terms of model explanation, we conclude with a general study of methods that now allow for better understanding of complex methods such as neural networks In summer 2013, the term "Big Data" appeared and attracted a lot of interest from companies.
Topics of the publication
Themes detected by scanR from retrieved publications. For more information, see https://scanr.enseignementsup-recherche.gouv.fr