2020 GUIDE TO DATA-DRIVEN LENDING - How data transformation and machine learning are creating a new profit center for credit unions - Blue Orange ...
←
→
Page content transcription
If your browser does not render page correctly, please read the page content below
2020 GUIDE TO DATA-DRIVEN LENDING How data transformation and machine learning are creating a new profit center for credit unions. blueorange.digital
Table of Contents 2.4 Predictive Analytics: Transforming Decision Making 19 Section 1: Introduction 05 2.4.1 Prediction is Now Cheap 19 Summary 05 2.4.2 Advanced Analytics Examples 20 Where We’re Going 06 2.4.3 Prediction Can Change Business Models 21 2.4.4 What Does This Mean for Credit Unions 22 2.1 Distilling the Buzzwords 08 Section 3: Data Driven Loan Pricing 25 2.1.1 AI Lowers the Cost of Prediction 08 3.1. Traditional Loan Pricing 25 2.1.2 What is Machine Learning? 08 3.1.1 Loan Pricing 25 2.2 Cloud Computing 09 3.1.2 Traditional Pricing Methods 25 2.2.1 How the Cloud Works 09 3.1.3 Problems with Traditional Risk Measures 26 2.2.2 Cloud Computing Overview 10 3.2 Machine Learning Approach 28 2.2.3 Cloud Benefits for Credit Unions 3.2.1 What Big Banks are Doing with ML 28 2.3 Modern Data Architecture and Why it Matters for Credit Unions 13 3.2.2 Why Lending is a Perfect Problem for Machine Learning 29 2.3.1 Framework 13 3.2.3 Advantages and Disadvantages of ML 29 2.3.2 Data Architecture Matters for BI 13 3.2.4 How ML Improves Lending 30 2.3.3 Legacy Architecture: An Evolved System 14 3.2.5 Thinking like a Data Scientist 32 2.3.4 What the Heck is a Data Lake? 16 3.2.6 Machine Learning Models for Lending 33 2.3.5 Why Do Data Lakes Matter? 17 2.3.6 The Modern Data Pipeline 18 2 3
3.3 Lending Club Case Study 36 3.3.1 Defining an Objective 36 3.3.2 Exploring and Cleaning Data 37 3.3.3 Modeling 38 3.3.4 How to Evaluate an ML Model 38 3.3.5 Results of High/Low Risks Loans 39 3.3.6 FICO Model 41 3.3.7 Discussion 43 Section 4: Bringing it Home 45 4.1 Determining if Your CU is Ready for Predictive Analytics 45 4.1.1 Is Predictive Analytics a Good Fit for Your Business? 45 4.1.2 Do You Have Good Data Infrastructure? 46 4.1.3 Do You Have the Right Team? 46 4.2 Leading Change 48 01 Section 5: Conclusion 51 Introduction 4 05
1 Introduction Where We’re Going The purpose of this whitepaper is to share how modern data can work for credit unions. Not every community credit union is going to have the data access or institutional buy-in to Summary implement a machine learning derived approach to lending. But understanding the opportunities and implications of modern data is accessible to everyone. A number of Over the last 10 years, we’ve seen businesses fundamentally change how they make trends are converging that will categorically transform how the credit union industry decisions. Business leaders are able to uncover new patterns in complexity. Find signals in a operates. Loan pricing, while powerful for driving profitability, is just a single example of world with more noise. This ultimately improves their ability to understand, then influence these predictive applications. desired outcomes. The paper is broken down into 3 main sections: This dramatic shift in how organizations operate can be attributed to a simple thing: data. 1. A tour of modern data to provide the necessary background of the historical changes But changing into a data-driven organization is no small feat. Today “data is the new oil” underway. powering improved business insight with an incredible combination of vast new data 2. The details of data-driven loan pricing. sources, cheap computing, and ever-stronger algorithms to dissect them. Like industrial 3. Implementing predictive analytics at your institution. transformation a century ago, adjusting to this world requires new tools, infrastructure, and mindsets. With all this opportunity, it’s an exciting time to look at the future of credit unions. The industry is being transformed by data analytics and cloud computing. The advent of predictive analytics has made data-driven problem solving accessible to all CUs. The winners in this new competitive landscape will be the ones who rapidly embrace advanced prediction. They will invest heavily in machine learning and the modern data architecture to support. Improving loan pricing and profitability is an excellent use case for these tools. By unifying internal and external data sets and applying readily available predictive tools (off the shelf models, 3rd party business intelligence software, internal upskilling), credit unions can increase loan profitability and reduce risk. But even if your credit union isn’t ready to immediately invest in a data lake and machine learning, there are specific steps you can take to both instill data-driven decision making in your department and more profitably price loans. 06 07
Section 2: Tour de Data 2.1 Distilling the Buzzwords 2.1.1 AI Lowers the Cost of Prediction “AI serves a single, but transformative, economic purpose: it significantly lowers the cost of prediction” --Ajay Agrawal AI lowers the cost of prediction. From here on out, when you hear AI or Machine Learning, think cheap prediction. Once you start to see challenges in this light and ask what data do we have or can we acquire to reframe a business question as a predictive one, you have already taken the most important conceptual step to applying data-driven decision making at your institution. Today advances in machine learning and cloud computing have democratized access to advanced predictive problem-solving. Take member lifetime value. Many credit unions don’t even know the economic value of a specific member. How likely are they to purchase a new product, open another account, respond to an email offer, or leave for another institution altogether? But understanding this number is the key to making informed decisions about how to allocate internal resources and prioritize members. It can also serve as a driver of strategic consensus, cross- departmental collaboration, and better member experience. Accurate prediction and 02 member segmentation underpin a data-driven flywheel. 2.1.2 What is Machine Learning? Machine Learning (ML) is a type of computer algorithm (or multiple algorithms) that can repeatedly apply statistical models and analyze the results to find patterns in data. What makes machine learning algorithms valuable are the feedback loops--a well-designed Tour de Data algorithm continues to learn from new input data to increase the accuracy of the solution it comes up with. ML discovers correlations in the data to make these predictions. 08 09
2.2.2 Cloud Computing Overview Service Models Take Fraud detection. Humans cannot efficiently review and classify the volume and velocity There are 3 levels of cloud computing services that correspond to the level of user of today’s transactional data. There are two major problems: effectively scanning billions of management required. transactions to determine whether they are fraudulent and detecting new forms of fraud. • IaaS (Infrastructure as a Service) Machine learning algorithms are trained against large amounts of existing data to discover IaaS are the building blocks of cloud IT. These are virtual resources that resemble the patterns that correlate to fraudulent behavior. Then the models are applied against current ones in physical data centers and allow users to create custom computing environments. transaction data to detect fraud in real-time and identify new forms of fraud far more Servers, storage drives, networking tools (virtual routers & firewalls) are examples of low- effectively than deploying a team of humans on the same task. level resources. • PaaS (Platform as a Service) 2.2 Cloud Computing PaaS involves pre-configured platforms where the user is not responsible for the underlying infrastructure. This allows users to deploy and maintain their own applications and custom development environments without worrying about provisioning servers or security maintenance. A managed database is a good example. 2.2.1 How the Cloud Works • SaaS (Software as a Service) Cloud computing is the on-demand availability SaaS software tools and applications are hosted in the cloud, giving users full access to software via a web browser or mobile app. Salesforce is a dominant example of a CRM of computer resources as a managed service software service delivered over the web, where some customization may be required but over the internet. Rather than devoting the user doesn’t think about any of the underlying implementation details. expertise to running on-premises data infrastructure, credit unions can focus on what Deployment Models they do best--serving members--while Because many organizations transition to the cloud with existing on-premise software companies like Amazon and Microsoft maintain resources, there are 3 different models for cloud deployment. the underlying infrastructure. But the advantages of the cloud go far beyond Cloud Hybrid On-Prem specialization. The ability to immediately spin up Full cloud deployment is done A hybrid deployment is done Maintaining in-house cloud services that most credit unions could not when all parts of a solution (both when a solution integrates both infrastructure (also known as a develop internally, scale on-demand, and pay high-level services and low-level cloud and on-premise resources. private cloud) is a deployment only for what you use with world-class security infrastructure components) A common scenario would be model for optimizing legacy run in the cloud. An example when the data must be stored in IT infrastructure with cloud creates incredible strategic and tactical of a fully deployed AWS-based a self-maintained, private data tools. While still requiring flexibility, and in the process trades Capex solution is Starling Bank, which center, while data processing expert knowledge for system guesses for fully utilized operational expenses. A took a cloud-first approach to applications run in the cloud. maintenance, it provides new server or service is literally a click away. develop its infrastructure. Cloud The former provides operational full control of resources and services and applications that stability and compliance, while optimization strategies. Union integrate seamlessly provide the latter offers the agility and Bank has integrated a private great flexibility and scaling. The scalability needed for customer- cloud solution to improve added bonus of off-loading on- facing applications. This application delivery times and premise maintenance makes this deployment model is a common automate already established IT deployment model very popular. step in most cloud adoption processes. strategies since it allows for an incremental, step-by-step migration. 10 11
• Lower Capital Expenses for Computing Infrastructure “With the new private cloud infrastructure in place, the bank has already Cloud Computing enables the pay-as-you-go model, where the fixed costs of cutting edge started seeing significant benefits. Deploying new applications, which infrastructure are effectively amortized over millions of purchasers. Resources that previously required 8-12 weeks, can now be executed in just a few hours. previously required the acquisition and maintenance of hardware, equipment, and This allows IT to speed time to market for new applications, enhancing the physical infrastructure can instead be invested directly in provisioning computing bank’s ability to serve its internal and external customers.” resources, when and if they are needed. This allows organizations to focus on projects that differentiate their business and lets the cloud vendor handle the underlying infrastructure. 2.2.3 Cloud Benefits for Credit Unions • Cloud Solutions as Incremental Systems “Moving to the cloud” is not a process that happens overnight. The full transformation The major cloud vendors (Amazon Web Services, Microsoft Azure, Google Cloud Platform) are continually increasing their service offerings to cover business and technical needs requires a carefully planned, multi-stage implementation strategy. Credit unions can across industries. This has enabled institutions to focus on their differentiated value rather start out with an in-house experimental phase, use expert knowledge for the assessment than struggle with technical maintenance issues (often at a lower cost as well). These of cloud opportunities, and perform a thorough comparison of available cloud services in benefits play out in multiple ways: an easily reversible way. This empirical knowledge lays the foundation for further steps in the cloud adoption strategy. • Reliability & Data Security Cloud Vendors use state-of-the-art tools and knowledge. They invest heavily in research • Scalability & Flexibility and development of security protocols, encryption standards, and novel data access The ability to scale on-demand makes cloud-based solutions extremely versatile. schemes, with the additional ability to fully isolate your data in a private cloud. This Software and hardware resources can be scaled up and down depending on current allows them to provide very high standards of member data safety, even for industries business needs. For large enterprises, this helps optimize operational costs. For small that require high levels of data security (e.g. military, trading, insurance). The fact that startups (or startup-style projects within credit unions), it allows for rapid testing and data is stored redundantly in the cloud also ensures the protection of member data enables accommodating a large number of users on short notice. Such agility is a against natural disasters and an easier data recovery process. fundamental business need in today’s competitive market. • Unified Data Cloud-based solutions make it far easier to integrate multiple data types and managed services, which gives credit unions greater ownership and control of their data. Regardless of file formats or database schemas, data that was once stored in isolated silos can now be kept in a single location. The simple decision of unifying the data in the cloud (while keeping the same variety of formats) enables more effective data analytics. The unified approach gives credit unions that bring their disparate data to the cloud more insight and access to their data sources. 12 13
2.3 Modern Data Architecture and Why it Matters for CUs 2.3.1 Framework 2.3.3 Legacy Architecture: An Evolved System Data architecture is admittedly an abstract topic. But it matters because without having a flexible, performant architecture any attempts at integrated analytics will be immediately This is your brain...on a legacy data pipeline. stopped in their tracks. Legacy architectures, rarely update spreadsheets, unintuitive UIs, and siloed data storage leave employees, members, and leadership in a reactive state. Many vendors offer single solution dashboards to address specific needs, but often this further compounds the foundational problem: disparate data systems only provide a partial picture. However, the new cloud paradigm and tools offer a compelling solution. Credit unions can own their data, directly answer business questions, easily combine internal and external data sources, and have the flexibility to ask new questions and bring on new vendors without being constrained by past architectural decisions. Beyond a commitment to a major cloud provider, you are not locked into traditional credit union data vendors with their siloed business models. 2.3.2 Data Architecture Matters for BI Business intelligence is best defined by its end goal: Making better business decisions. In Wikipedia’s parlance, BI “enables more effective strategic, tactical, and operational insights and decision-making” [3]. To do this, we need two things: available data and a way to understand it. Making good, well-informed decisions is the goal. Performing queries, drawing graphs, training models and making predictions are some of the ways we can move towards this goal. But without accurate, accessible, and relevant data there is nothing to apply analytics on. For these tools to be truly meaningful and effective, your whole organization needs to deploy a modern data architecture. Figure 1: Legacy Dataflow (Blue Orange Sales team design) Traditional BI has required businesses to invest in sophisticated infrastructure and Figure 1 shows how a typical legacy data pipeline would look. This is not a planned system, expensive development projects. This approach incurs continual support costs and make it but the result of evolving business needs. As more and more external services are added, difficult to be responsive to changing business needs. Think of having to request a monthly and more independent internal teams make use of the data, the architecture evolves into a sales report from IT or internally manage a server network, which all too often devolves into complex tangled spider web of data connections. Each team has to figure out what data is this. available, where to find it, what each column and field actually means (the data semantics), Major advances in tooling, along with sophisticated data processing, have democratized the and do its best (using custom code and legacy ETL software) to pull it all together so they job of analytics, moving it from IT departments to business stakeholders. Let’s start by can perform their computations. learning how the traditional approach to BI data processing evolved. 14 15
This works initially. However, as the company grows it becomes increasingly hard to 2.3.4 What the Heck is a Data Lake? manage. The process is very error-prone and requires significant, costly coordination and effort from each team. The end result is fragile and inaccurate. If any service, whether In order to avoid the issues detailed above, many credit unions have considered or (often internal or external changes at all, the code either breaks or, worse yet, produces inaccurate tragically) tried to implement a data warehouse. Data warehouses pair well with ETL, an results. extraction technique that formats data before putting it into storage. Much like a giant database where data is stored in a standard format, warehouses conform data to On top of the architectural and design challenges, much more data is produced today than constraints to be queried more easily. However, over time, these constraints create usability the systems were initially designed for, further stressing the legacy approach. Traffic is challenges that replicate many of the legacy issues for organizations that want to increasing, analytics are becoming more prevalent and information-rich, and media is now implement advanced predictive analytics. ubiquitous (e.g. screenshots, recording mouse movements, images, and video). If that isn’t So how is a data lake different? An analogy might help here. There are two enough, large third party data sources, both free and premium, are also needed to enrich Order Matters: ETL vs ELT prototypical drawers in most kitchens: a internal data sets. Every addition requires more code, more ETL operations, and ultimately ETL stands for Extract, Transform, and Load and utensil drawer where each type of utensil much more processing. represents any technique used to get the data you has its own special place and that catch- Processing all this data cost-effectively is the next challenge. Employing advanced need in the format and shape you need to perform all-drawer where you throw everything your task — before loading it into storage. First you else you need. The former is a data processing on this growing mass data makes server loads difficult to balance. Similar to a extract the data, for example by reading data from warehouse; the latter, a data lake. power grid, designing computing power to meet peak load requirements means that full disk, performing an HTTP call or an API call. Next you utilization only occurs for part of the day or week. The result is expensive server allocations transform it through one or more of the following: A data lake can be thought of like a giant that often sit idle. filtering the data, selecting fields of interest, joining directory or file system. Data lakes pair several disparate data sources together, enriching well with ELT, an extraction technique Finally, since the data is stored and only made available through proprietary services the data by performing a calculation to produce new fields, cleaning it, filling in missing values, or that doesn’t alter the data before storing (designed by and for engineers), leveraging this data relies on direct support from it. This simple-sounding difference identifying errors and correcting them. The load is engineering and IT. Any special information, query, or graph required by a decision-maker the final phase by which you load the data into your has big implications. Anybody in the will need coordination and assistance from an engineer or data scientist. Even data discovery target storage, be it a data lake, data warehouse or organization can store their data there, in becomes a major hurdle. Management, and even the engineers who run the system, whatever tickles your fancy. The key point is that any format, schema, or order. your data will be immediately ready for use. struggle to know what data is available, what data would prove useful, what the data means, and how to access it. The result is simplistic, slow decision making and a slow-moving But now you might ask, “Isn’t it easier to ELT (Extract, Load, Transform) is a variation of this organization. find a fork if all the forks are organized idea. Rather than transforming the data prior to the load phase, and effectively storing the result of together?” Well, yes, it is easier to find the process described above, you perform these a fork, but finding and counting forks is LEGACY CHALLENGES manipulations when you need the data. just scratching the surface of modern To summarize the challenges facing a legacy data pipeline: analytics. • Hard to scale - difficult data discovery, data access, etc. With today’s cheap and powerful computing, grouping things in whatever way you want is • Error-prone nearly instant. The data doesn’t need to be stored in a specific format, which opens the door • Exclusive data access to specialists to endless types of data analytics. Modern tools such as Spark/Hadoop can process any • Expensive to maintain organizational data out of a unified environment and analyze it in any given way. Like a knife • Redundancy of data and code to fork or a fork to fork-from-3-years-ago. Ok...we may have worn out this analogy. • Wasteful and costly computing allocation 16 17
2.3.5 Why Do Data Lakes Matter? A credit union with modern data architecture has reliable, accurate data pipes running all the way from the user fumbling through your web site, to adding third party information The data lake pattern is the preferred architectural solution for 2020 because it provides and services, conducting analysis, joining disparate data sources, performing a low cost, flexible, and scalable way to store large amounts of diverse data. All your enrichments, distributing that data to internal servers, training and deploying machine credit union data that is now sitting in dozens of siloed, vendor-specific systems is stored in learning models, and finally pulling all this data together again to allow the decision- a central location that you own, so it’s readily available to be categorized, processed, maker to interact, visualize, query and explore it in order to make the best choice. Specific analyzed, enriched, and consumed by downstream analysts and operational teams. By data is collected, modeled, transformed, and visualized to be put in an actionable using ‘metadata’ to tag data rather than coerce it into a predetermined warehouse format, business context. broader types of data can be used effectively for business intelligence. When a business question arises, the data lake can be queried for relevant data in any format, and this Here is a map of modern AWS data architecture for credit unions. It looks complicated, smaller set of data is rapidly available to be analyzed and help answer that specific question. but the beauty is that these are all high-end data services that you don’t have to build. AWS has created impressive tools at every stage of the data lifecycle that can be pulled From an executive perspective, the main advantage of a data lake is that it delivers strategic together, revolutionizing how credit unions think about and use their data. flexibility. Owning your data in a cloud data lake provides high levels of security, portability, and access to better tools. It doesn’t constrain future changes in business strategy because the raw data is still there. If you decide on a new BI or reporting tool, just write a job to structure the data in that format. If next year’s annual plan calls for a completely new initiative that requires answers to questions you didn’t know you needed to ask when the data lake was stood up, fine. If a new source of member data starts growing exponentially, you can handle it. The data lake approach provides a solid architectural foundation to build on for the future, even if you don’t know exactly what that future entails. Strategic Flexibility Cheaper Takeaway: Data Flexibility Single Source of Truth Future Proofed Better BI Integration 2.3.6 The Modern Data Pipeline Your data is stored in its raw, persistent form in an AWS data lake built on Lake Formation. The data you need for a specific job is pulled out, transformed, normalized, and sent to The deeper purpose of deploying a data lake is to simplify decision making across the an analytics engine in real-time. Let’s say you have already purchased a loan analyzing institution. It provides the foundation for advanced analytics and the control to use the software that you love. We write a ‘job’ that transforms a representation of your loan data vendors you value without being locked-in. After building dozens of data lakes for Fortune into a schema the software can use. But the underlying data in the single source of truth 500 companies to startups, we at Blue Orange have seen a fundamental shift in the past doesn’t change. The next level is actually applying modern data science tools like EMR and year that democratizes access to cutting edge data tools. In terms of cloud computing, we SageMaker to implement machine learning models and predictive analytics. Once the data are “standing on the shoulders of giants,” and credit unions of any size can be the pipeline is built, you just rent the processing time. beneficiaries. 18 19
2.4 Predictive Analytics: Transforming Decision Making 2.4.1 Prediction is Now Cheap 2.4.2 Advanced Analytics Examples Making accurate predictions about the future is a fundamental business need because it helps leaders make the right decisions for their organization. However, traditional BI tools Let’s take a closer look at a few examples of advanced analytics in finance to understand and subjective interpretations are no longer enough to gain actionable insights from the how the industry is learning to capitalize on member and product data. Financial vast amounts of data that are deluging today’s credit unions. In a data-centric world, institutions that have embraced digital transformation are successfully employing modern powerful machine learning technologies are already driving enterprise performance. machine learning tools for a variety of tasks. Companies large and small are rapidly embracing the transformative economic purpose of ML: It significantly lowers the cost of prediction. Member Acquisition The technological advances of the past decades and the increasing focus on data-driven solutions played a huge role in lowering these prediction costs: Predictive models improve the efficiency of the sales process and enable more accurate lead qualification. Hidden data features can be isolated and help identify ideal member • Data collection (from humans, machines, and software) is now a common practice profiles or implement probabilistic lead scoring models. This allows sales teams to focus and across industries and the volume allows for more powerful and accurate models. their attention on the most valuable prospects. • Predictive software tools are becoming increasingly specialized and provide rapid Personalized Marketing solutions for previously expensive tasks. Predictive analytics enable personalized interactions with prospective members. By • Academia and the industry are continually researching and developing new AI tools revealing novel data patterns, response modeling empowers marketing teams to reach that have created a flourishing ecosystem. specific groups with personalized content and more effective cross-selling offers. • Commodity hardware is also becoming cheaper and modern computational units Member Segmentation needed for Machine Learning (such as GPUs and CPUs) lower the processing times of previously intensive tasks. Data insights allow better segmentation of member groups. Modern clustering techniques can make sense of online & offline member behavioral patterns. Predicting member • The multitude of data-oriented cloud tools makes it possible for any organization to lifetime value and churn allows credit unions to effectively tailor member service and offers tackle cutting edge machine learning challenges. From big data acquisition to building for subsets of their members with an understanding of potential impacts on the bottom complex processing pipelines and training deep learning models, there are countless line. opportunities at a low cost, pay on-demand manner. Revenue Forecasting These advances make possible what was previously unimaginable. Heterogeneous data sources can be integrated and modern learning architectures deployed with the click of a Machine learning tools are modern alternatives for cash flow management. Integrating button. This allows companies to regularly leverage machine technologies for data analysis multiple data sources (marketing, sales, operation & member data) has been shown to and in turn, the insights obtained from these data approaches are transforming industries. provide more accurate demand predictions for lending and financial products. Advanced predictive analytics has never been as affordable as it is right now and the costs (both in money and time) are still going down. 20 21
These are just a few applications in the financial sector where predictive ML has been Here’s a fun thought experiment. How accurate would Amazon’s product recommendation successfully used to leverage the power of data. There are hundreds of additional engine need to be before they just started shipping your boxes of items they think you’ll opportunities for predictive analytics in every line of business in a credit union. Wherever want rather than waiting for you to order? 60%...70%? At some point, the entire business a data-based process exists, that data can be leveraged to optimize it. Better prediction model flips and it would be more profitable for Amazon to shift its business model to improves the credit union member experience, enhances revenue, and reduces ship-then-shop. People would be happy that a spatula showed up to replace their old one operational costs. and if they didn’t want to keep it a courier would take it away for free the next day. With a sufficiently accurate prediction of what shoppers will buy, the risk of shipping an unwanted product is smaller than the chance the shopper decides to keep the product. Once Amazon’s strategy proved successful and they secured the first-mover advantage, it is expected that competitors would try to follow. Relying on predictive technology for strategic business decisions becomes the only way to stay ahead. This example shows how advanced prediction can go beyond the strategy of a single organization to impact the economics of an entire industry. 2.4.4 What Does this Mean for Credit Unions? There are multiple ways that financial institutions can pursue predictive analytics as part of their data-driven transformation strategy. Here are some examples: Automation 2.4.3 Prediction Can Change Business Models Predictive analytics enables financial institutions to increase efficiency and achieve The trend is similar across industries: predictive analytics is a major asset in the hands of operational agility by automating recurring processes. Traditional back-office processes are any organization that wishes to transform itself. Moreover, increased predictive accuracy tedious, labor-intensive and error-prone. Accounts payable, member service tasks, and may even change the business model itself. even loan approval are examples of rule-based processes where AI solutions could improve manual worker performance. Automating these manual tasks and reducing human In retail, a powerful example is related to Amazon’s advanced analytics strategy and the involvement has a direct impact on operational performance, staffing, and expenses. resulting superior prediction. With a recommendation engine that is based on increasing amounts of data and heterogeneous sources (online behavior, social metadata, offline Unbiased Indicators shopping history), Amazon is steadily improving how accurately they can predict what shoppers will buy. Since accurate indicators can provide a more complete picture of an organization’s health, it is critical to interpret performance values by means of well-established, reliable statistical tools. Advanced analytics provides credit unions with the opportunity to accurately aggregate KPIs and discover hidden trends that would otherwise remain unknown. 22 23
By replacing man-made features with automated ML-enhanced predictions, the relevant features are learned from the data itself and the resulting insights are less biased. This translates to higher model accuracy and better-informed decision making. Data-Driven Decisions The main purpose of predictive analytics is to make sense of raw data and turn it into meaningful insights that enable smarter decisions. The biggest advantage of modern predictive analytics is the possibility of handling heterogeneous data sources. For example, a marketing campaign may not achieve the expected outcome when based on an incomplete member profile. Pulling in third-party data (such as social media information, shopping behavior, online transaction history) can help reveal and target ideal member segments. Successful strategies and informed decisions are easier to find when aided by data analytics. Real-time Analytics Predictive analytics should not only be performed on historical data. There are many situations in which real-time analytics can solve an important business use case. A great financial example is fraud prevention, which detects cybercrime as (or even before) it even occurs. By modeling user behavior and performing anomaly detection, fraud monitoring tools can employ machine learning continuously. Other related examples of real-time predictive analytics also include money laundering detection and stock market surveillance software. 03 Data-Driven Loan Pricing 24 25
Section 3: Data-Driven Loan Pricing Price Leadership In a competitive world, the credit union in our example above may not be able to charge 8% if another institution is offering the borrower the same loan at 6.5%. With deregulation, 3.1. Traditional Loan Pricing alternative lending sources, and fully online banks, it is difficult to maintain traditional cost plus profit margins. The alternative is to use a price leadership model to establish a credit 3.1.1 Loan Pricing union's cost of credit. A prime lending rate is set for the credit union's most creditworthy borrowers, which serves as the benchmark rate for other loans. One of the challenges for One of the greatest challenges for lenders is figuring out the interplay between credit credit unions who assume this model is minimizing funding and operating costs while score, term length, and collateral to determine the risk premium. Fundamentally, a loan is a keeping the risk premium competitive. A low prime rate may result in additional lending contract that allows people to trade money across time. The interest rate that the borrower but accumulate hidden operational and credit risks. pays should account for both the opportunity cost of money to the lender and the risk of default. In this section we will focus on risk scoring, and, in particular, using machine Credit Risk Pricing learning to create a more nuanced determination of credit risk. A credit score or risk-based pricing method focuses on the differential characteristics of a borrower to assign the risk or default premium to the loan. Under the credit risk model, 3.1.2 Traditional Pricing Methods the decision of lenders to issue credit is heavily influenced by the credit score of the borrower and the credit report provided by the largest US credit offices--Equifax, Experian, There are three main pricing models lenders use today: cost plus, price leadership, and and TransUnion. FICO (Fair Isaac Corporation) is the most popular credit scoring program, credit scoring/risk-based pricing. ranging from 300 to 850. The rating algorithms of FICO are designed to forecast consumer behavior with a statistically determined default rate. The model assumes that a borrower Cost Plus with a subprime credit score of 620 may have trouble repaying on loan based on their past borrowing and payment history. Applying a credit scoring system allows credit unions to Cost plus pricing is the simplest model as it assumes there is no market pressure from other assign a default premium and use cutoff scores to compete for certain borrower groups lenders. The interest rate is determined by four components: and reject or offer higher prices to potentially riskier borrowers. On the flip side, • Funding Cost: How much it costs the credit union to raise the loaned funds prospective borrowers with a high credit score will typically be rewarded by credit unions for their responsible behavior with a lower-priced loan. • Operating Costs: Credit union overhead and servicing costs • Risk Premium: Compensation for default risk of the loan 3.1.3 Problems with Traditional Risk Measures • Profit margin: Assures credit union receives an adequate return While traditional loan pricing methods have been functioning for decades, there are 4 key For example, someone comes into the credit union asking for a small $1000 loan. If the problems that emerge from the way they measure loan risk: cost of funds is 3%, operating costs are 1%, the default risk premium is 2% and the credit union needs a 2% profit margin, then the credit union sums it up and offers to lend at 8%. Done and done. 26 27
1) Limited features Assessments of creditworthiness are based on a limited number of features defined by the model. For example, FICO uses payment history, credit utilization, length of credit history, new credit, and credit mix as its primary categories. This approach leads to more general decisions not tailored to the needs of the specific borrower. There are clearly other features that would influence loan pricing if they were able to be considered. 2) Disqualification of qualified borrowers Disqualification of potentially creditworthy persons because they are unable to meet certain criteria is also problematic. A person’s credit history, for example, has a significant impact on the final score. This often excludes young people or people from other countries who would have the financial ability to assume the debt. Another example would be unduly penalizing someone who has one or more late payments in the past due to explicable circumstances that are no longer present. Judging someone’s creditworthiness based 3.2 Machine Learning Approach primarily on outdated or unexamined factors can lead to false lending assumptions, which is a loss for both the lender and borrower. 3.2.1 What Big Banks are Doing with ML 3) FICO primacy The four biggest banks in the US are actively applying machine learning today: • JP Morgan Chase: Contract Intelligence - COIN In the credit scoring industry, the FICO scoring system has little competition. Since most lenders receive the FICO score of the borrower from the same 3 major credit offices, if a • Bank of America: Chatbot, fraud detection, trading certain lender refuses a potentially creditworthy borrower for a loan, they are likely to be • Wells Fargo: Blockchain, improved analytics, NLP, chatbots rejected elsewhere. This overemphasis on FICO results in the inability to meet the demands of the member (as that is what the borrower ultimately is for credit unions). • Citigroup: Document digitization using CV and NLP. Chatbots for customer service and anomaly fraud detection on transactions JP Morgan’s project is the most revealing because it involves the full-scale automation of 4) Increasing member demands white-collar jobs. The company created a contract review software using Natural Language Processing algorithms that replaces 360,000 hours of annual legal contract review. They The demands of consumers for speed, accuracy, and ease of use have increased have been able to reduce the time spent on this job to a matter of seconds using machine exponentially due to the technological revolution witnessed over the past 10 years. Existing learning. credit issuers and credit offices need to meet the needs of today’s consumers or will be replaced by newly formed companies paying better attention to the experience of their members. This demand for customized, personal approaches is both an opportunity for credit unions to differentiate now and will become table stakes in the years ahead. 28 29
3.2.2 Why Lending is a Perfect Problem for Machine Learning Disadvantages • Exclusion Risk: Lending dependent on ML bears financial exclusion risks. ML models Machine learning algorithms are designed to find patterns in data. Simply put, there are far are trained using available data that may not necessarily be representative of all more data sources available to provide an edge in lending than are currently being used by classes of borrowers the lender is considering. If the training set is missing critical credit unions. For example, an individual’s primary financial institution has years of member variables for a specific borrow, the model may not perform well. data that hold statistical correlations which would bear on future loans if the additional information could be analyzed. Credit unions already incorporate “non-standard”, i.e. local • Privacy Issues: Credit rating based on ML could cause problems with consumer or relational, data to make lending decisions. Internal and external data sources can be protection, ethical issues, and privacy by incorporating untoward alternative data systematically applied to underwriting decisions, improving both profitability and lending sources. assessment accuracy. • Shifting Macro Environment: A major change in the environment could make historical data irrelevant for assessing new members. Historical sample analysis may 3.2.3 Advantages and Disadvantages of ML not generalize for new applicants due to significant structural changes (such as changes in financial development, macroeconomic policies, and industries). Before diving in, it makes sense to understand the relative strengths and weaknesses of machine learning models for credit union credit risk analysis. • Applicant Gaming: If a factor becomes generally known as a credit indicator, borrowers will often try to artificially improve it. While this happens today with FICO Advantages scores (applicants focus on inflating their score when seeking a mortgage), ML-based • Makes smaller loans viable. ML makes it feasible and affordable for small lenders to models would not be immune. For example, if borrowers discovered that the number determine credit risk. Credit unions will often refrain from investigating a small of social media connections was valued by the algorithm, they could expand the size borrower’s credit because of the size of the loan. Reducing the cost of conducting a of their network to achieve a better credit rating, thus gaming the system. credit review is a major financial consideration. • Increases the value of “soft’ information. ML allows lenders to go beyond debt, 3.2.4 How ML Improves Lending income, and repayment history. While those variables are critical they are not the whole story on a borrower’s debt-servicing capacity. Think of machine learning as a power tool. If there is enough data (electricity), ML will significantly outperform manual tools (historical statistics). It’s not like current lending • Identi ies nonlinear relationships. By searching for small sample partitioning approaches are bereft of data and analytics. Even the most basic lending model uses relationships, ML models can capture local relationships between risk indicators and historical data to determine a projected default rate in order to calculate the risk premium. credit risk outcomes that are not identified by traditional models or human agents. The difference is that ML approaches generate far more predictive accuracy, which can be directed towards multiple aspects of a lending problem. This becomes a difference in kind • Reduces information asymmetry: ML can reduce the asymmetry of information when applied in the right way. between the lender and the borrower. Moral hazard and adverse selection problems are less likely to arise when the lender incorporates more reliable information. 30 31
Opening New Markets Increased Control 3.2.5 Thinking Like a Data Scientist With increased competition in the lending Another way ML helps credit unions is by So you’re convinced that machine learning offers significant advantages compared to space, credit unions are looking to create a increasing control over formerly traditional loan pricing and risk methods. How do you get started? differentiated offering without increasing independent aspects of the lending risk or sacrificing profitability. Data-driven problem. Take the 4 criteria of the cost- The first step to applying ML is beginning to think like a data scientist. The diagram below is ML approaches can identify good borrowers plus model: Funding cost, operating cost, the bread and butter of how Blue Orange data scientists approach problems. Exploring the among populations that were previously risk premium, and profit margin. They are amount and type of the data you have allows you to identify patterns where data is valuable dismissed as high-risk. Increased predictive actually interdependent and credit unions and determine the best method for analyzing it. Most business questions involve regression accuracy permits a nuanced understanding can allocate the increased predictive or classification, which we will explore in the next section. of credit risk that opens up these new accuracy of machine learning to optimize markets. along any dimension. ML-driven loan approval would lower operating costs, A good analogy is junk bonds when they better credit risk prediction allows for were first popularized in the 1980s. This was more accurate and competitive risk a specific submarket where the risk/return premiums, and selective lending for ratio for lenders was out of whack. Due to profitability would increase ROE. How and historical norms and bondholders failing to where to make the tradeoff between those incorporate new information, there were factors is a business decision. This decision essentially free risk-adjusted returns itself (and the consequences of favoring available to data-savvy lenders. Obviously, different criteria) and be optimized with this thinking can be taken too far when data. speculative dynamics come into play, but early movers who are willing to go beyond the consensus with data often reap outsized rewards. At a higher level, there are 4 stages to implement an effective machine learning solution. 32 33
Logistic Regression 1) Discover: We start by understanding a company’s business objective and exploring the Regression involves predicting a quantity by figuring out which features are important for initial data. We need to figure out what data is available to determine if machine learning is the outcome. Logistic regression is a method for estimating the probability of an event a viable solution. Robust, production level ML solutions typically require a range of tools and occurring with binary data. This works great for predicting loans because the outcome data a deep assessment of internal capabilities. (default/repay or approve/deny) is binary. 2) Structure: Data gathering, cleaning, formatting, imputing, over/subsampling. ML demands In the example below, specific borrower features are given different weights depending significant computing power and access to highly available data, which is why building a on how well they predict the loan being approved. The weighted sum correlates to the modern data pipeline is a prerequisite for successful machine learning projects. probability of approval and places the borrower at a point on the S-shaped curve (0 to 1). 3) Model: This is where the magic happens. Our Ph.D. data scientists experiment with Once the probability crosses a certain threshold determined by the model (say .6), the loan models, tune hyperparameters, and train specific features in order to increase accuracy is considered likely to be approved and categorized as such. beyond traditional statistical and business intelligence tools. ML models are used to capture new insights and unlock the value of client data. 4) Integrate: The greatest ML model is useless unless it is implemented well. Technically, we deploy the model to production and monitor it for accuracy, functionality, and improvements. Strategically, we actively review the results and additional use cases with the client team for feedback. The full technical stack is focused on ensuring business outcomes. By approaching business questions with an understanding of the data science development path, credit union leaders are able to vastly improve the technical capabilities and implementation outcomes of their teams. Decision Trees 3.2.6 Machine Learning Models for Lending A decision tree is a model that predicts the value of a target variable by learning simple At the highest level, ML models use inputs (like income, age, and loan duration) to estimate decision rules inferred from the data features. It adopts a hierarchical structure like a if a person should get a loan or not. Each of these variables has a different weight in the flowchart that starts from a root node, progresses to lower nodes through possible states final decision. The models become more accurate over time by improving the distribution of or decisions (represented as a branch), and ends at the terminal node that shows the input weights to better predict outcomes. consequence of the entire branch. Decision trees have the advantage of being used for both regression and classification models. Let’s dive in and explore 3 types of machine learning techniques that apply to lending: Logistic Regression, Decision Trees, and Ensemble Learning. One of the advantages of decision trees is how well they can be understood by non-experts. Consider the not granted decision on the left. If the borrower is under 40, has an income below $166500, and saves less than $657 per month, then they don’t get approved for a loan. This can seem similar to a heuristic or scorecard method, but the prioritization of features is determined by the model itself for the greatest predictive accuracy. 34 35
3.3 Lending Club Case Study The Lending Club is a private company founded in 2006 that loans money peer-to-peer. The amount could be between 1K to 40K USD and the standard loan period is 36 months. The platform has grown from $33 million in loan originations in 2009 to $10.9 billion in 2018. The average LendingClub borrower has a FICO score of 699, 17.7% debt-to-income ratio (excluding mortgage), 16.2 years of credit history, $73,945 of personal income and takes out Ensemble Learning an average loan of $14,553 that they typical use for debt consolidation or paying off credit Ensemble learning is a method that combines the results of multiple machine learning cards. models into one prediction. Imagine you have to answer a challenging question (like should Lending Club has released a public dataset which has allowed us to compare the predictive you approve a loan). Would you rather ask a single expert opinion or get the aggregate accuracy of a machine learning model to one based on FICO. This approach has general answer from a group of experts? Ensemble learning gives you the latter. applicability to credit union lending and allows us to take you through the process of This is a more robust solution and provides more accurate results. In the example below building a real-world machine learning model step by step. 4 different ML models are combined to find a better solution. The downsides are that the resulting model is less interpretable, the computation and design time is high, and model 3.3.1 Defining an Objective selection is challenging. Nevertheless, if you have access to a data scientist, we have found that Ensemble Learning is a superior method for lending accuracy. The first step of an ML project is to define the objective. With Lending Club, borrowers with higher FICO credit scores (more stable and less risky) get lower interest rates on their loans, whereas borrowers with lower credit scores (less reliable and riskier) get higher rates. These loans are grouped in tranches from A (low risk) to E (high risk). Loans with higher interest rates are more enticing from the investor’s point of view because they provide a higher investment return. On the other hand, they pose a higher risk of default. Therefore, a machine learning model that could predict which of the high-interest loans are more likely to be repaid would bring added investment value while minimizing the associated repayment risks. Our objective is to build a machine learning scoring model that outperforms FICO with a focus on high risk/high profit loans. Now let’s see machine learning in action. 36 37
3.3.2 Exploring and Cleaning Data 5. Search for and remove outliers in the entire dataset. 6. Apply techniques to handle missing values. Options include setting a removal threshold or imputing the value using the min/median flag. Exploration 7. Search for multicollinearity and correlations among columns. That means finding the We start by segregating the loans based on payment status. A Lending Club loan can have variables which are highly linearly related. The best practice is to remove them or set one of six statuses: Current, Fully Paid, Grace Period, Late, Default, Charged Off. We group a threshold and keep only those that fulfill the condition. these into two sets of loans: good or bad. After completing these steps, our data gets filtered to only 20K rows by 85 columns. Good loans 3.3.3 Modeling • Current: All pending payments are up to date. • Fully paid: The loan was fully reimbursed either at the expiration date or as a result of As discussed in the ML models section (3.2.6), we chose an ensemble learning method an advance payment. (which combines several machine learning techniques into one predictive model) because • Grace Period: Loan payment is due but within the 15 day grace period. it out performed any single model. The Gradient Boost Model (GBM) is a series of weak predictive models based on decision trees used for regression and classification problems Bad loans like lending. • Late (16-30)/(31-120): Loan has not been current for either 16-30 or 31-120 days. This model is very efficient for handling features with many categories. GBM uses a mean • Default: Loan has not been paid for an extended period (more than 120 days). encoding that replaces each categorical feature with only one numerical feature. Traditional • Charged Off: There is no longer a reasonable expectation of further payments. one-hot encoding generates a lot of features, which leads to shallow decisions trees and makes the classification complicated. In Boosting, each new tree is a fit on a modified version of the original data set. Preparation To make sure that our model can generalize to new data, we split the dataset into three The next step is preparing the data for analysis. The dataset has nearly 500,000 parts: train, validation, and test. The first two are used for training and to validate that our observations and contains 150 measurable pieces of data that can be used for analysis. That model is doing well. The test set is a portion of the original data that was not used to build means we start with a dataset of 0.5M rows by 150 columns. It’s necessary to reduce the the model. Running the model against it lets us see how the model will perform in the feature dimensions and remove duplicates and errors to effectively analyze it. presence of completely unknown data. Preparation Steps: 3.3.4 How to Evaluate an ML Model 1. Remove syntax errors such as capital letters, underscores, whitespaces. 2. Verify the name of columns and check for uniqueness and semantic similarity. So how do we actually measure if our model is effective? By using a Confusion Matrix. A 3. Decide if the date is going to be considered with or without time. confusion matrix is a method that allows us to visualize the output of an algorithm. The 4. Standardize data types across all the columns such as decimals, integers, strings, etc. columns represent the actual values and the rows represent the predicted values. Looking at the boxes lets us see how many predictions our model got right and wrong (and in what way). 38 39
You can also read