Last year while attending Febraban Tech 2024, a financial services industry event in LATAM, we were intrigued to learn that 96% of surveyed banks have artificial intelligence (AI) initiatives. We wondered–how many banks are actually taking advantage of AI initiatives? And, how can FSIs take advantage of what small language models have to offer? And, can these advantages apply beyond the world of financial services?
What are small language models?
A key aspect of an AI model is the quantity of parameters that are used to train a given model. When exploring models, you’ll find this number to be in the billions for each model. The greater the value, the richer the learning and more advanced its ability to handle more complex language tasks. The bigger, the better? Well, that depends. If you need to customize a hundreds of billions of parameters model, it will take significant (think of days or weeks) computational and GPU capacity to carry a single round of training of that model.
A small language model (SLM) is a model that you can customize, or fine-tune, with your own data within a reasonable time with currently available hardware (assuming you have accelerated compute resources in place or access to an appropriate cloud instance).
If you do your own research into AI models, you‘ll find that some offer parameter variants of the base model. For example, the Granite 3.1 model family offers from sub-billion to 34 billion parameter variants. Consider starting with lower sized variants to find out the capabilities of a model and scaling up in case of need—and to discover the improvements, if any, a larger parameter model can provide. A smaller model will enable you to start more quickly and tune in a reasonable time frame.
SLMs in financial services
While looking into this topic, we came across an interesting article by Drew Breunig. In his article, he breaks down AI into 3 main use cases: Gods, Interns and Cogs. The “God” case is the full human replacement AI or artificial general intelligence (AGI), where lots of funding and research is currently going; “Interns” produce very good content, but must have their content reviewed to confirm that it is accurate and appropriate; and finally, the “cogs”—where I see SLMs existing—are purpose-built models with a low error rate, allowing them to run fairly unsupervised and a domain where users can achieve significant cost savings.
Tests, code development, stress tests and forecasts are strong candidates for “intern”-type AI. Specialist and low error margins in the financial services industry, such as customer service, hyper-personalization, fraud detection and document treatment are natural candidates for purpose-built and specialist “cog” models. These tasks aren’t unique to FSI in principle, so it’s certainly feasible that other industries can take advantage of SLM-based cogs in a similar fashion.
Benefits of SLMs
Is there interest for larger and more complex AI models? Of course. They can be useful for oversight mechanisms, judge/teach base models or where general broad knowledge is desired, but those can come with their associated costs, complexity and requirements.
Instead of taking a monolithic approach to AI and going all-in on a single LLM, organizations should consider a compound solution of purpose-built models, potentially even a mix of SLMs and/or LLMs. This becomes especially interesting if you’re able to pair these models with existing cloud-native or traditional IT applications, creating an interconnected workflow across the hybrid cloud.
A collection of purpose-built SLMs can help:
- Improve the security posture of your content by controlling all the aspects of your model.
- Accelerate innovation through fine-tuning.
- Segregate access by implementing per model access to back-end systems.
- Improve quality control by versioning each of the models.
- Manage costs and enable efficiency by decreasing a model’s training time.
The good news is that open source tools for experimenting with and building your own SLMs are readily available. I recommend trying InstructLab, a groundbreaking project developed by IBM and Red Hat, that provides model alignment tooling to help organizations more efficiently contribute skills and knowledge to their gen AI models in order to address the needs of their AI-enabled applications and business.
There’s a lot about SLMs to love, not just for the FSI world but for the broader business community at large. Explore your options, and with open source tools and open source-licensed models, you can customize and tune an AI model that works for you. So try it out, and you may realize how using your own SLMs can help you solve your company’s problems.
blog
How to get started with InstructLab today
About the authors
Rodrigo is a tenured professional with a distinguished track record of success and experience in several industries, especially high performance and mission critical environments in FSI. A negotiator at his heart, throughout his 20+ year career, he has leveraged his deep technical background and strong soft skills to deliver exceptional results for his clients and organizations - often ending in long-standing relationships as a trusted advisor. Currently, Rodrigo is deep diving on AI technology.
Thiago Araki is Red Hat's Senior Director of Tech Sales in Latin America. He is responsible for defining the portfolio strategy, simplifying the adoption of emerging technologies by the market, and supporting organizations in their digital transformation process.
He joined Red Hat in 2013 as a Solutions Architect. Previously, he worked as a Systems Designer at Digitel. He was also a consultant at Accenture where he carried out business transformation projects for large Brazilian companies and designed the financial and management information systems at major Brazilian banks.
He has extensive experience in the IT industry as a Product Marketing Manager, Consultant and Solutions Architect, backed by more than 20 years working in leading companies and participating in major transformation projects.
More like this
Browse by channel
Automation
The latest on IT automation for tech, teams, and environments
Artificial intelligence
Updates on the platforms that free customers to run AI workloads anywhere
Open hybrid cloud
Explore how we build a more flexible future with hybrid cloud
Security
The latest on how we reduce risks across environments and technologies
Edge computing
Updates on the platforms that simplify operations at the edge
Infrastructure
The latest on the world’s leading enterprise Linux platform
Applications
Inside our solutions to the toughest application challenges
Original shows
Entertaining stories from the makers and leaders in enterprise tech
Products
- Red Hat Enterprise Linux
- Red Hat OpenShift
- Red Hat Ansible Automation Platform
- Cloud services
- See all products
Tools
- Training and certification
- My account
- Customer support
- Developer resources
- Find a partner
- Red Hat Ecosystem Catalog
- Red Hat value calculator
- Documentation
Try, buy, & sell
Communicate
About Red Hat
We’re the world’s leading provider of enterprise open source solutions—including Linux, cloud, container, and Kubernetes. We deliver hardened solutions that make it easier for enterprises to work across platforms and environments, from the core datacenter to the network edge.
Select a language
Red Hat legal and privacy links
- About Red Hat
- Jobs
- Events
- Locations
- Contact Red Hat
- Red Hat Blog
- Diversity, equity, and inclusion
- Cool Stuff Store
- Red Hat Summit