Blog

Covid-19 and compliance challenges for payment companies

Site Logo
Tookitaki
10 May 2021
read
5 min

The COVID-19 pandemic has brought in fundamental changes to many industries, and the payments sector is no exception. Consumers across the globe now have a different approach as they consider commerce and payment options. They tend to flock to those options that are safer and faster. According to the Visa Back to Business Study, 78% of global consumers have adjusted the way they pay for items in the wake of intensified safety concerns. “This shift to digital-first commerce and technologies like contactless payments has ushered in a new generation of consumer tendencies that will have a ripple effect on the global economy for years to come,” says the study.

Businesses also prefer non-cash and contactless payment methods amid lockdowns and changing consumer preferences. Expectations abound for global cash in circulation to reduce in the next few years on the back of digitization initiatives by governments led by Finland and Sweden. Increasing digitalization in payments gave rise to a new number of payment companies, while incumbent payment processors are adapting their strategies to match the new-age payment technologies. However, digitalization may not address the age-old problems related to transactions -- money laundering, terrorist financing and other financial crimes. It would rather help criminals adapt their strategies and remain undetected.

Transaction laundering and its possibilities

The proliferation of e-commerce and digital payments have already given new avenues for criminals to run their illegal businesses and then launder crime proceeds. Criminals resorting to transaction laundering, also known as electronic money laundering or cyberlaundering, have a favourable situation now as people shop online more. Criminals can set up apparently legitimate websites to sell illegal goods and divert the payments to merchant payment accounts. They may either open by their own accounts or manage to get access to accounts set up by third parties. By routing payments to these accounts, they would be able to launder criminal proceeds through Payment Service Providers. Given below are some factors that help transaction laundering:

  • The process of creating an online store with a checkout page and credit card info storing feature is very easy now. This helps criminals conduct the three steps of money laundering -- placement, layering and integration -- digitally.
  • There are a large number of e-merchants and payment facilitators that provide a great degree of anonymity, helping criminals.
  • Banks are outsourcing their merchant acquisition operations to third-party payment service providers and facilitators who don’t have rigorous onboarding systems.
  • Alternate payment methods such as digital wallets, mobile wallets and payment gateways have seen significant growth in userbase but they lack proper monitoring systems.
  • Payment companies either do not come under the purview of AML/CFT regulations or are not governed by the same level as that of banks.

Transaction Laundering puts Payment Service Providers -- a group of banks, acquirers and payment processors -- into a difficult situation where they unknowingly become facilitators of money laundering. The results are legal actions, hefty regulatory penalties and ultimately severe damage to the reputation.

Transaction Laundering Types

There are three forms of transaction laundering:

  • Front Companies: Appearing as legitimate online businesses selling goods or services, these are set up criminals to launder criminal funds by means of over/under-invoicing, inflating transactions or misreporting earnings. For authorities, it is very hard to detect these criminals as the value of the services involved is often subjective.
  • Pass-through Companies: These are not set up by criminals but they allow criminals to launder money through their accounts, in many cases for a hefty commission.
  • Funnel Accounts: They involve payment processors for multiple companies who knowingly or unknowingly process illegitimate transactions along with legitimate ones. These illegal transactions are hard to detect as they are mixed in with legal ones.

Possibilities of Micro Laundering

Micro laundering refers to the money laundering technique where criminals break large sums of ill-gotten money into tiny portions and transact them without being detected by AML systems. At present, criminals can operate the scheme from their mobile phones making use of digital payment platforms. While a transaction worth US$10,000 would form a red flag, 100 transactions worth US$100 each would easily surpass system checks. A 2018 study by HP Threat Research found that an estimated 10% of cybercriminals are using PayPal to launder money. A further 35% use other digital payment systems, including Skrill, Dwoll, Zoom, and mobile payment systems like M-Pesa.

How to Prevent Money Laundering via Payment Companies

Payment firms must reinforce their AML checks and monitoring processes in compliance with applicable regulations in order to prevent the misuse of their platforms for criminal purposes. Given below are some of the areas where they should focus on:

Know Your Customer (KYC) and Customer Due Diligence

Strengthened Know Your Customer (KYC) and customer due diligence (CDD) measures can prevent transaction laundering to a great extent. Payment processing companies must ensure they do not onboard malicious merchants by mistake. Firms should gather necessary information on merchants and their beneficial owners, if applicable. They should analyse each piece of information and assess the potential risks of doing business with each client. Proper KYC measures also ensure that honest merchant customers are not troubled at all.

Transaction Monitoring

Payment firms need to properly monitor the transactional patterns of their clients. In the case of online merchants, a sudden increase in sales volumes or amounts compared to historical levels or projections should be thoroughly investigated. Unrealistic sales projections should also form a red flag. International payment providers need to develop deep understandings of the markets and merchant sectors they operate in to be able to pinpoint high-risk interactions.

Proper customer risk assessment while onboarding and on an ongoing basis has become difficult. With systems dependent on a few parameters prescribed by regulations, payment firms might not be able to gauge risk accurately. They need a system that can take into account a large number of parameters and dynamically adjust risk rating. In the case of transaction monitoring, human investigators might find it cumbersome to examine each and every one of hundreds of alerts generated on an hourly basis. The current need is for a mechanism that accurately sifts out ‘false positives’ so that compliance staff can focus on true suspicious cases.

And no doubt, modern technologies such as artificial intelligence (AI) and machine learning can take up the task. For example, machine learning can analyze thousands of transactions in a faster manner than human investigators, helping reduce the number of false positives drastically. In addition, it can detect true cases better than myopic legacy systems.

How We Can Help

Tookitaki’s end-to-end AML operating system, the Anti-Money Laundering Suite (AMLS), powered by AML Federated Knowledge Base is intended to identify hard-to-detect money laundering techniques. AMLS is available as a modular service across the three pillars of AML activity – Transaction Monitoring, AML Screening for names and transactions and Customer Risk Scoring. The AI-powered solution has the following features to aid payment companies with their AML/CFT compliance.

  • AI-powered detection of interactions and network relationships between customers or interested parties to flag suspicious activity
  • World’s biggest repository of AML typologies providing real-world AML red flags to keep our underlying machine learning detection model updated with the latest money laundering techniques across the globe.
  • Smart alert management to identify alerts that matter and that are non-productive
  • Advanced data analytics and dynamic segmentation to detect unusual patterns in transactions
  • Risk scoring based on matching with watchlist databases or adverse media
  • Visibility on customer linkages and related scores to provide a 360-degree network overview
  • Constantly updating risk scoring which learns from incremental data changes

AMLS was designed keeping in mind the ability of AML/CTF compliance systems to integrate with disparate data sources and platforms. Users may it as a standalone system or on top of legacy systems to augment their efficiency.

Our solution has been proven to be highly accurate in identifying high-risk customers and transactions. For more details of our AMLS solution and its ability to identify the latest money laundering techniques, please contact us.

Ready to Streamline Your Anti-Financial Crime Compliance?

Our Thought Leadership Guides

Blogs
27 Oct 2025
6 min
read

Eliminating AI Hallucinations in Financial Crime Detection: A Governance-First Approach

Introduction: When AI Makes It Up — The High Stakes of “Hallucinations” in AML

This is the third instalment in our series, Governance-First AI Strategy: The Future of Financial Crime Detection.

  • In Part 1, we explored the governance crisis created by compliance-heavy frameworks.

  • In Part 2, we highlighted how Singapore’s AI Verify program is pioneering independent validation as the new standard.

In this post, we turn to one of the most urgent challenges in AI-driven compliance: AI hallucinations.

Imagine an AML analyst starting their day, greeted by a queue of urgent alerts. One, flagged as “high risk,” is generated by the newest AI tool. But as the analyst investigates, it becomes clear that some transactions cited by the AI never actually happened. The explanation, while plausible, is fabricated: a textbook case of AI hallucination.

Time is wasted. Trust in the AI system is shaken. And worse, while chasing a phantom, a genuine criminal scheme may slip through.

As artificial intelligence becomes the core engine for financial crime detection, the problem of hallucinations, outputs not grounded in real data or facts, poses a serious threat to compliance, regulatory trust, and operational efficiency.

What Are AI Hallucinations and Why Are They So Risky in Finance?

AI hallucinations occur when a model produces statements or explanations that sound correct but are not grounded in real data.

In financial crime compliance, this can lead to:

  • Wild goose chases: Analysts waste valuable time chasing non-existent threats.

  • Regulatory risk: Fabricated outputs increase the chance of audit failures or penalties.

  • Customer harm: Legitimate clients may be incorrectly flagged, damaging trust and relationships.

Generative AI systems are especially vulnerable. Designed to create coherent responses, they can unintentionally invent entire scenarios. In finance, where every “fact” matters to reputations, livelihoods, and regulatory standing, there is no room for guesswork.

ChatGPT Image Oct 27, 2025, 01_15_25 PM

Why Do AI Hallucinations Happen?

The drivers are well understood:

  1. Gaps or bias in training data: Incomplete or outdated records force models to “fill in the blanks” with speculation.

  2. Overly creative design: Generative models excel at narrative-building but can fabricate plausible-sounding explanations without constraints.

  3. Ambiguous prompts or unchecked logic: Vague inputs encourage speculation, diverting the model from factual data.

Real-World Misfire: A Costly False Alarm

At a large bank, an AI-powered monitoring tool flagged accounts for “suspicious round-dollar transactions,” producing a detailed narrative about potential laundering.

The problem? Those transactions never occurred.

The AI had hallucinated the explanation, stitching together fragments of unrelated historical data. The result: a week-long audit, wasted resources, and an urgent reminder of the need for stronger governance over AI outputs.

A Governance-First Playbook to Stop Hallucinations

Forward-looking compliance teams are embedding anti-hallucination measures into their AI governance frameworks. Key practices include:

1. Rigorous, Real-World Model Training
AI models must be trained on thousands of verified AML cases, including edge cases and emerging typologies. Exposure to operational complexity reduces speculative outputs.At Tookitaki, scenario-driven drills such as deepfake scam simulations and laundering typologies continuously stress-test the system to identify risks before they reach investigators or regulators.

2. Evidence-Based Outputs, Not Vague Alerts
Traditional systems often produce alerts like: “Possible layering activity detected in account X.” Analysts are left to guess at the reasoning.Governance-first systems enforce data-anchored outputs:“Layering risk detected: five transactions on 20/06/25 match FATF typology #3. See attached evidence.”
This creates traceable, auditable insights, building efficiency and trust.

3. Human-in-the-Loop (HITL) Validation
Even advanced models require human oversight. High-stakes outputs, such as risk narratives or new typology detections, must pass through expert validation.At Tookitaki, HITL ensures:

  • Analytical transparency
  • Reduced false positives
  • No unexplained “black box” reasoning

4. Prompt Engineering and Retrieval-Augmented Generation (RAG)Ambiguity invites hallucinations. Precision prompts, combined with RAG techniques, ensure outputs are tied to verified databases and transaction logs, making fabrication nearly impossible.

Spotlight: Tookitaki’s Precision-First AI Philosophy

Tookitaki’s compliance platform is built on a governance-first architecture that treats hallucination prevention as a measurable objective.

  • Scenario-Driven Simulations: Rare typologies and evolving crime patterns are continuously tested to surface potential weaknesses before deployment.

  • Community-Powered Validation: Detection logic is refined in real time through feedback from a global network of financial crime experts.

  • Mandatory Fact Citations: Every AI-generated narrative is backed by case data and audit references, accelerating compliance reviews and strengthening regulatory confidence.

At Tookitaki, we recognise that no AI system can be infallible. As leading research highlights, some real-world questions are inherently unanswerable. That is why our goal is not absolute perfection, but precision-driven AI that makes hallucinations statistically negligible and fully traceable — delivering factual integrity at scale.

Talk to an Expert

Conclusion: Factual Integrity Is the Foundation of Trust

Eliminating hallucinations is not just a technical safeguard. It is a governance imperative. Compliance teams that embed evidence-based outputs, rigorous training, human-in-the-loop validation, and retrieval-anchored design will not only reduce wasted effort but also strengthen regulatory confidence and market reputation.

Key Takeaways from Part 3:

  1. AI hallucinations erode trust, waste resources, and expose firms to regulatory risk.

  2. Governance-first frameworks prevent hallucinations by enforcing evidence-backed, auditable outputs.

  3. Zero-hallucination AI is not optional. It is the foundation of responsible financial crime detection.

Are you asking your AI to show its data?
If not, you may be chasing ghosts.

In the next blog, we will explore how building an integrated, agentic AI strategy, linking model creation to real-time risk detection, can shift compliance from reactive to resilient.

Eliminating AI Hallucinations in Financial Crime Detection: A Governance-First Approach
Blogs
13 Oct 2025
6 min
read

When MAS Calls and It’s Not MAS: Inside Singapore’s Latest Impersonation Scam

A phone rings in Singapore.
The caller ID flashes the name of a trusted brand, M1 Limited.
A stern voice claims to be from the Monetary Authority of Singapore (MAS).

“There’s been suspicious activity linked to your identity. To protect your money, we’ll need you to transfer your funds to a safe account immediately.”

For at least 13 Singaporeans since September 2025, this chilling scenario wasn’t fiction. It was the start of an impersonation scam that cost victims more than S$360,000 in a matter of weeks.

Fraudsters had merged two of Singapore’s most trusted institutions, M1 and MAS, into one seamless illusion. And it worked.

The episode underscores a deeper truth: as digital trust grows, it also becomes a weapon. Scammers no longer just mimic banks or brands. They now borrow institutional credibility itself.

Talk to an Expert

The Anatomy of the Scam

According to police advisories, this new impersonation fraud unfolds in a deceptively simple series of steps:

  1. The Setup – A Trusted Name on Caller ID
    Victims receive calls from numbers spoofed to appear as M1’s customer service line. The scammers claim that the victim’s account or personal data has been compromised and is being used for illegal activity.
  2. The Transfer – The MAS Connection
    Mid-call, the victim is redirected to another “officer” who introduces themselves as an investigator from the Monetary Authority of Singapore. The tone shifts to urgency and authority.
  3. The Hook – The ‘Safe Account’ Illusion
    The supposed MAS officer instructs the victim to move money into a “temporary safety account” for protection while an “investigation” is ongoing. Every interaction sounds professional, from background call-centre noise to scripted verification questions.
  4. The Extraction – Clean Sweep
    Once the transfer is made, communication stops. Victims soon realise that their funds, sometimes their life savings, have been drained into mule accounts and dispersed across digital payment channels.

The brilliance of this scam lies in its institutional layering. By impersonating both a telecom company and the national regulator, the fraudsters created a perfect loop of credibility. Each brand reinforced the other, leaving victims little reason to doubt.

Why Victims Fell for It: The Psychology of Authority

Fraudsters have long understood that fear and trust are two sides of the same coin. This scam exploited both with precision.

1. Authority Bias
When a call appears to come from MAS, Singapore’s financial regulator, victims instinctively comply. MAS is synonymous with legitimacy. Questioning its authority feels almost unthinkable.

2. Urgency and Fear
The narrative of “criminal misuse of your identity” triggers panic. Victims are told their accounts are under investigation, pushing them to act immediately before they “lose everything.”

3. Technical Authenticity
Spoofed numbers, legitimate-sounding scripts, and even hold music similar to M1’s call centre lend realism. The environment feels procedural, not predatory.

4. Empathy and Rapport
Scammers often sound calm and helpful. They “guide” victims through the process, framing transfers as protective, not suspicious.

These psychological levers bypass logic. Even well-educated professionals have fallen victim, proving that awareness alone is not enough when deception feels official.

The Laundering Playbook Behind the Scam

Once the funds leave the victim’s account, they enter a machinery that’s disturbingly efficient: the mule network.

1. Placement
Funds first land in personal accounts controlled by local money mules, individuals who allow access to their bank accounts in exchange for commissions. Many are recruited via Telegram or social media ads promising “easy income.”

2. Layering
Within hours, funds are split and moved:

  • To multiple domestic mule accounts under different names.
  • Through remittance platforms and e-wallets to obscure trails.
  • Occasionally into crypto exchanges for rapid conversion and cross-border transfer.

3. Integration
Once the money has been sufficiently layered, it’s reintroduced into the economy through:

  • Purchases of high-value goods such as luxury items or watches.
  • Peer-to-peer transfers masked as legitimate business payments.
  • Real-estate or vehicle purchases under third-party names.

Each stage widens the distance between the victim’s account and the fraudster’s wallet, making recovery almost impossible.

What begins as a phone scam ends as money laundering in motion, linking consumer fraud directly to compliance risk.

A Surge in Sophisticated Scams

This impersonation scheme is part of a larger wave reshaping Singapore’s fraud landscape:

  • Government Agency Impersonations:
    Earlier in 2025, scammers posed as the Ministry of Health and SingPost, tricking victims into paying fake fees for “medical” or “parcel-related” issues.
  • Deepfake CEO and Romance Scams:
    In March 2025, a Singapore finance director nearly lost US$499,000 after a deepfake video impersonated her CEO during a virtual meeting.
  • Job and Mule Recruitment Scams:
    Thousands of locals have been drawn into acting as unwitting money mules through fake job ads offering “commission-based transfers.”

The lines between fraud, identity theft, and laundering are blurring, powered by social engineering and emerging AI tools.

Singapore’s Response: Technology Meets Policy

In an unprecedented move, Singapore’s banks are introducing a new anti-scam safeguard beginning 15 October 2025.

Accounts with balances above S$50,000 will face a 24-hour hold or review when withdrawals exceed 50% of their total funds in a single day.

The goal is to give banks and customers time to verify large or unusual transfers, especially those made under pressure.

This measure complements other initiatives:

  • Anti-Scam Command (ASC): A joint force between the Singapore Police Force, MAS, and IMDA that coordinates intelligence across sectors.
  • Digital Platform Code of Practice: Requiring telcos and platforms to share threat information faster.
  • Money Mule Crackdowns: Banks and police continue to identify and freeze mule accounts, often through real-time data exchange.

It’s an ecosystem-wide effort that recognises what scammers already exploit: financial crime doesn’t operate in silos.

ChatGPT Image Oct 13, 2025, 01_55_40 PM

Red Flags for Banks and Fintechs

To prevent similar losses, financial institutions must detect the digital fingerprints of impersonation scams long before victims report them.

1. Transaction-Level Indicators

  • Sudden high-value transfers from retail accounts to new or unrelated beneficiaries.
  • Full-balance withdrawals or transfers shortly after a suspicious inbound call pattern (if linked data exists).
  • Transfers labelled “safe account,” “temporary holding,” or other unusual memo descriptors.
  • Rapid pass-through transactions to accounts showing no consistent economic activity.

2. KYC/CDD Risk Indicators

  • Accounts receiving multiple inbound transfers from unrelated individuals, indicating mule behaviour.
  • Beneficiaries with no professional link to the victim or stated purpose.
  • Customers with recently opened accounts showing immediate high-velocity fund movements.
  • Repeated links to shared devices, IPs, or contact numbers across “unrelated” customers.

3. Behavioural Red Flags

  • Elderly or mid-income customers attempting large same-day transfers after phone interactions.
  • Requests from customers to “verify” MAS or bank staff, a potential sign of ongoing social engineering.
  • Multiple failed transfer attempts followed by a successful large payment to a new payee.

For compliance and fraud teams, these clues form the basis of scenario-driven detection, revealing intent even before loss occurs.

Why Fragmented Defences Keep Failing

Even with advanced fraud controls, isolated detection still struggles against networked crime.

Each bank sees only what happens within its own perimeter.
Each fintech monitors its own platform.
But scammers move across them all, exploiting the blind spots in between.

That’s the paradox: stronger individual controls, yet weaker collaborative defence.

To close this gap, financial institutions need collaborative intelligence, a way to connect insights across banks, payment platforms, and regulators without breaching data privacy.

How Collaborative Intelligence Changes the Game

That’s precisely where Tookitaki’s AFC Ecosystem comes in.

1. Shared Scenarios, Shared Defence

The AFC Ecosystem brings together compliance experts from across ASEAN and ANZ to contribute and analyse real-world scenarios, including impersonation scams, mule networks, and AI-enabled frauds.
When one member flags a new scam pattern, others gain immediate visibility, turning isolated awareness into collaborative defence.

2. FinCense: Scenario-Driven Detection

Tookitaki’s FinCense platform converts these typologies into actionable detection models.
If a bank in Singapore identifies a “safe account” transfer typology, that logic can instantly be adapted to other institutions through federated learning, without sharing customer data.
It’s collaboration powered by AI, built for privacy.

3. AI Agents for Faster Investigations

FinMate, Tookitaki’s AI copilot, assists investigators by summarising cases, linking entities, and surfacing relationships between mule accounts.
Meanwhile, Smart Disposition automatically narrates alerts, helping analysts focus on risk rather than paperwork.

Together, they accelerate how financial institutions identify, understand, and stop impersonation scams before they scale.

Conclusion: Trust as the New Battleground

Singapore’s latest impersonation scam proves that fraud has evolved. It no longer just exploits systems but the very trust those systems represent.

When fraudsters can sound like regulators and mimic entire call-centre environments, detection must move beyond static rules. It must anticipate scenarios, adapt dynamically, and learn collaboratively.

For banks, fintechs, and regulators, the mission is not just to block transactions. It is to protect trust itself.
Because in the digital economy, trust is the currency everything else depends on.

With collaborative intelligence, real-time detection, and the right technology backbone, that trust can be defended, not just restored after losses but safeguarded before they occur.

When MAS Calls and It’s Not MAS: Inside Singapore’s Latest Impersonation Scam
Blogs
13 Oct 2025
6 min
read

How Collective Intelligence Can Transform AML Collaboration Across ASEAN

Financial crime in ASEAN doesn’t recognise borders — yet many of the region’s financial institutions still defend against it as if it does.

Across Southeast Asia, a wave of interconnected fraud, mule, and laundering operations is exploiting the cracks between countries, institutions, and regulatory systems. These crimes are increasingly digital, fast-moving, and transnational, moving illicit funds through a web of banks, payment apps, and remittance providers.

No single institution can see the full picture anymore. But what if they could — collectively?

That’s the promise of collective intelligence: a new model of anti-financial crime collaboration that helps banks and fintechs move from isolated detection to shared insight, from reactive controls to proactive defence.

Talk to an Expert

The Fragmented Fight Against Financial Crime

For decades, financial institutions in ASEAN have built compliance systems in silos — each operating within its own data, its own alerts, and its own definitions of risk.
Yet today’s criminals don’t operate that way.

They leverage networks. They use the same mule accounts to move money across different platforms. They exploit delays in cross-border data visibility. And they design schemes that appear harmless when viewed within one institution’s walls — but reveal clear criminal intent when seen across the ecosystem.

The result is an uneven playing field:

  • Fragmented visibility: Each bank sees only part of the customer journey.
  • Duplicated effort: Hundreds of institutions investigate similar alerts separately.
  • Delayed response: Without early warning signals from peers, detection lags behind crime.

Even with strong internal controls, compliance teams are chasing symptoms, not patterns. The fight is asymmetric — and criminals know it.

Scenario 1: The Cross-Border Money Mule Network

In 2024, regulators in Malaysia, Singapore, and the Philippines jointly uncovered a sophisticated mule network linked to online job scams.
Victims were recruited through social media posts promising part-time work, asked to “process transactions,” and unknowingly became money mules.

Funds were deposited into personal accounts in the Philippines, layered through remittance corridors into Malaysia, and cashed out via ATMs in Singapore — all within 48 hours.

Each financial institution saw only a fragment:

  • A remittance provider noticed repeated small transfers.
  • A bank saw ATM withdrawals.
  • A payment platform flagged a sudden spike in deposits.

Individually, none of these signals triggered escalation.
But collectively, they painted a clear picture of laundering activity.

This is where collective intelligence could have made the difference — if these institutions shared typologies, device fingerprints, or transaction patterns, the scheme could have been detected far earlier.

Scenario 2: The Regional Scam Syndicate

In 2025, Thai authorities dismantled a syndicate that defrauded victims across ASEAN through fake investment platforms.
Funds collected in Thailand were sent to shell firms in Cambodia and the Philippines, then layered through e-wallets linked to unlicensed payment agents in Vietnam.

Despite multiple suspicious activity reports (SARs) being filed, no single institution could connect the dots fast enough.
Each SAR told a piece of the story, but without shared context — names, merchant IDs, or recurring payment routes — the underlying network remained invisible for months.

By the time the link was established, millions had already vanished.

This case reflects a growing truth: isolation is the weakest point in financial crime defence.

Why Traditional AML Systems Fall Short

Most AML and fraud systems across ASEAN were designed for a slower era — when payments were batch-processed, customer bases were domestic, and typologies evolved over years, not weeks.

Today, they struggle against the scale and speed of digital crime. The challenges echo what community banks face elsewhere:

  • Siloed tools: Transaction monitoring, screening, and onboarding often run on separate platforms.
  • Inconsistent entity view: Fraud and AML systems assess the same customer differently.
  • Fragmented data: No single source of truth for risk or identity.
  • Reactive detection: Alerts are investigated in isolation, without the benefit of peer insights.

The result? High false positives, slow investigations, and missed cross-institutional patterns.

Criminals exploit these blind spots — shifting tactics across borders and platforms faster than detection rules can adapt.

ChatGPT Image Oct 13, 2025, 12_54_11 PM

The Case for Collective Intelligence

Collective intelligence offers a new way forward.

It’s the idea that by pooling anonymised insights, institutions can collectively detect threats no single bank could uncover alone. Instead of sharing raw data, banks and fintechs share patterns, typologies, and red flags — learning from each other’s experiences without compromising confidentiality.

In practice, this looks like:

  • A payment institution sharing a new mule typology with regional peers.
  • A bank leveraging cross-institution risk indicators to validate an alert.
  • Multiple FIs aligning detection logic against a shared set of fraud scenarios.

This model turns what used to be isolated vigilance into a networked defence mechanism.
Each participant adds intelligence that strengthens the whole ecosystem.

How ASEAN Regulators Are Encouraging Collaboration

Collaboration isn’t just an innovation — it’s becoming a regulatory expectation.

  • Singapore: MAS has called for greater intelligence-sharing through public–private partnerships and cross-border AML/CFT collaboration.
  • Philippines: BSP has partnered with industry associations like Fintech Alliance PH to develop joint typology repositories and scenario-based reporting frameworks.
  • Malaysia: BNM’s National Risk Assessment and Financial Sector Blueprint both emphasise collective resilience and information exchange between institutions.

The direction is clear — regulators are recognising that fighting financial crime is a shared responsibility.

AFC Ecosystem: Turning Collaboration into Practice

The AFC Ecosystem brings this vision to life.

It is a community-driven platform where compliance professionals, regulators, and industry experts across ASEAN share real-world financial crime scenarios and red-flag indicators in a structured, secure way.

Each month, members contribute and analyse typologies — from mule recruitment through social media to layering through trade and crypto channels — and receive actionable insights they can operationalise in their own systems.

The result is a collective intelligence engine that grows with every contribution.
When one institution detects a new laundering technique, others gain the early warning before it spreads.

This isn’t about sharing customer data — it’s about sharing knowledge.

FinCense: Turning Shared Intelligence into Detection

While the AFC Ecosystem enables the sharing of typologies and patterns, Tookitaki’s FinCense makes those insights operational.

Through its federated learning model, FinCense can ingest new typologies contributed by the community, simulate them in sandbox environments, and automatically tune thresholds and detection models.

This ensures that once a new scenario is identified within the community, every participating institution can strengthen its defences almost instantly — without sharing sensitive data or compromising privacy.

It’s a practical manifestation of collective defence, where each institution benefits from the learnings of all.

Building the Trust Layer for ASEAN’s Financial System

Trust is the cornerstone of financial stability — and it’s under pressure.
Every scam, laundering scheme, or data breach erodes the confidence that customers, regulators, and institutions place in the system.

To rebuild and sustain that trust, ASEAN’s financial ecosystem needs a new foundation — a trust layer built on shared intelligence, advanced AI, and secure collaboration.

This is where Tookitaki’s approach stands out:

  • FinCense delivers real-time, AI-powered detection across AML and fraud.
  • The AFC Ecosystem unites institutions through shared typologies and collective learning.
  • Together, they form a network of defence that grows stronger with each participant.

The vision isn’t just to comply — it’s to outsmart.
To move from isolated controls to connected intelligence.
To make financial crime not just detectable, but preventable.

Conclusion: The Future of AML in ASEAN is Collective

Financial crime has evolved into a networked enterprise — agile, cross-border, and increasingly digital. The only effective response is a networked defence, built on shared knowledge, collaborative detection, and collective intelligence.

By combining the collaborative power of the AFC Ecosystem with the analytical strength of FinCense, Tookitaki is helping financial institutions across ASEAN stay one step ahead of criminals.

When banks, fintechs, and regulators work together — not just to report but to learn collectively — financial crime loses its greatest advantage: fragmentation.

How Collective Intelligence Can Transform AML Collaboration Across ASEAN