Federated Learning for Finance: Saves 200+ BDRs 2-4 hrs/wk, Cuts GTM 50%
Industry: Financial services
Client
Financial Services/Swift
Goal
Help business development team of 200+ BDRs identify suitable upselling capabilities within existing 10k+ client organisations based on usage patterns.
Challenges
- Some data sources were extremely confidential and could not be moved
- CRM data was inaccurate and could not be used as training data
Solution
The team implemented a federated learning approach to ensure confidential data does not leave the network. The application was deployed in an air-gapped environment. Strict rule-based access controls and comprehensive logging of user actions were implemented.
The HoAI’s team reasoned from first principles to identify the key indicators that make a client suitable and trained the model on target customers
Impact:
Saved a global team of 200+ BDRs 2-4 hours per week on data analysis
Saved a GTM team of four approximately 50% of their time on analyzing data and responding to data analysis requests
Context
A global financial services organization supporting real-time messaging and clearing (Swift-style) sought to improve upsell effectiveness across more than 10,000 client organizations. The objective was to help a business development team of 200+ business development representatives (BDRs) identify suitable upselling opportunities inside existing accounts based on product usage patterns and operational telemetry. The program was positioned as a go-to-market enablement initiative to convert product usage signals into prioritized outreach targets while preserving the extreme confidentiality required across the finance sector.
Challenges
Several critical constraints shaped the project. Key data sources were extremely confidential and legally or contractually bound to remain on-premises; these data could not be moved to a central cloud for model training. The production environment included air-gapped segments and strict regulatory controls. In addition, CRM records were known to be incomplete and inaccurate and therefore could not be used as reliable training data. The combination of immovable, sensitive data and poor CRM quality required a solution that could learn from distributed signals without compromising confidentiality or relying on flawed centralized customer records.
Implementation
The team implemented a federated learning architecture that allowed local model training at each data-holding site so that raw transaction and usage records never left the originating network. Only model updates and aggregated gradients were shared to a coordinating server, preventing exfiltration of sensitive information. To satisfy the most stringent security requirements, the solution was deployed into air-gapped environments at regulated sites and integrated with existing on-premises infrastructure.
Access was governed by strict rule-based controls and comprehensive auditing: role-based permissions defined which users and services could trigger training cycles, request model inferences, or read aggregated outputs, while tamper-evident logs recorded every user action for compliance and forensic review. The implementation team reasoned from first principles to define the most predictive indicators of upsell suitability—metrics such as feature adoption velocity, transaction volume growth, peak utilization patterns, error/retry rates, and peripheral product adjacency—then designed local feature extraction pipelines to compute those indicators where the data resided. Given unreliable CRM entries, deterministic matching and local canonicalization were used to align account contexts across sites before participating in federated aggregation.
Model training used a mix of supervised and semi-supervised techniques: target customers were identified from high-confidence segments (e.g., recent product adopters with measurable usage growth) and used to bootstrap a classifier that generalized across clients while respecting client-specific features. Privacy-preserving aggregation and differential update thresholds further minimized leakage risks. The resulting application delivered ranked upsell signals and contextual evidence to the global BDR workflow via secured APIs and a read-only analytics interface that exposed only allowed summaries, not raw records.
Results
The federated learning deployment produced tangible productivity gains for both field and go-to-market teams while preserving regulatory and contractual data boundaries. The global team of 200+ BDRs saved between 2–4 hours per person per week previously spent on manual data collection and analysis, allowing them to focus on outreach and relationship building. The central go-to-market analytics team of four reclaimed roughly 50% of their time previously consumed by ad hoc reporting and fulfilment of data analysis requests. By surfacing prioritized upsell candidates based on observed product usage signals rather than unreliable CRM notes, the organization improved the speed and relevance of account conversations without moving confidential data offsite. Auditing and strict access controls delivered the required compliance posture for finance, and the air-gapped deployment satisfied the most restrictive operational constraints—making the approach a practical pattern for other regulated environments seeking data-driven sales enablement without compromising data sovereignty.
*Case studies reflect work undertaken by our Heads of AI either during their tenure with Head of AI or in prior roles before they were part of the Head of AI network; they are provided for illustrative purposes only and are based on conversations with our Heads of AI.