The insurance industry is wrestling with the tremendous growth of data sources at its disposal. Traditional ETL processes are expensive, time-consuming, and complicated by the variety of data structures and formats. In contrast, Hadoop platforms provide a clean, safe, and manageable format for data wrangling, the critical first step of the data analysis process.
Forward-thinking insurance companies have embraced data wrangling as more than janitorial work. For them, wrangling is just as important a component as the final results. Properly executed, wrangling provides data insights that improve both analytical inquiries and the quality of the results.
“Discovering is something of an umbrella term for the entire process; in it, you learn what is in your data and what might be the best approach for productive analytic explorations.”
In the insurance world, it’s extremely important to have a more accurate understanding of available data. For example, a common term such as “household” is dynamic, given the growth of millennials living with their parents. Accordingly, the lifetime value of a household is now a better metric than the lifetime value of a customer. Discoveries such as these give us a good understanding of how to proceed.
“Structuring is needed because data comes in all shapes and sizes.”
Today, the insurance industry is able to pull data from not only structured sources, but unstructured sources as well. Call center transcripts, or unstructured text, can indicate whether a customer has problems with their current policies. These transcripts are an important data point, but it’s necessary to find a self-service tool, such as Trifacta, that offers an efficient way to structure this type of unstructured data for use.
“Cleaning involves taking out data that might distort the analysis, such as a null value.”
Data cleaning involves more than reformatting null values or fields: it’s also an opportunity to validate the trustworthiness of the data. Yet regulatory compliance within the insurance industry demands a complete data lineage. Trifacta enables this lineage by tracking all changes from raw data to completion.
“Enriching allows you to ask questions about other data that might be useful in the analysis, or new data that you can derive from existing data.”
Insurance data can be enriched by joining it with other data. For example, the what, when, and how of individual product purchases can be paired with marketing data. Once joined, the effectiveness of outreach efforts can be gauged, and tailored marketing efforts can target customers with greater success.
“Validating is the activity that surfaces data quality and consistency issues, or verifies that they have been properly addressed by applied transformations.”
When reviewing a total risk profile, it’s important to fact-check all known variables. The number of policies with a customer and subsidiary, for instance, should be contrasted with the exact timeline of changing systems. By validating the data, it’s possible to begin visualizing how all of these elements work together, and to ensure the data has been validated.
“Publishing delivers the output of data wrangling efforts for downstream project needs.”
Publishing enables use of wrangled data, whether by loading the data in a particular analysis package, or by documenting the transformation logic for future needs. In the insurance industry, it’s extremely important that wrangled data is published for the actuarial department, or for downstream business users that will do the analysis for risk analysis and underwriting, claims, or customer analytics.
As the insurance industry relies on Hadoop to support new analytical endeavors, proper data wrangling should remain a critical component of the process. Without it, opportunities for enhanced risk selection, customer growth, and the mitigation of claim leakage are lost. Data wrangling is a key process that explores all possible data, thereby providing the best possible understanding of it.
More about Trifacta and Hortonworks: https://hortonworks.com/partner/trifacta/