Incode is advancing the state of the art of AI models that identify and combat fraud. By leveraging foundational models trained on unique global fraud datasets and with the ability to continuously learn, Incode not only stops today’s fraud but also evolves at the speed of emerging gen-AI fraud threats.

In order to interpret complex identity, document, and behavioral signals, Incode develops Vision-Language Models, Large Language Models, and reasoning agents that work across modalities to evaluate fraud patterns and support adaptive detection as new attacks emerge.
Incode’s VLM is trained on global identity data, documents, templates, and synthetic fraud samples across 200+ regions. With few-shot learning, it adapts quickly to new attack types and unseen document formats. It analyzes visual and textual signals to detect tampering, synthetics, deepfakes, and altered documents with high precision.
.avif)
What it powers:
Tamper & Synthetic Detection: Identifies deepfakes, swaps, edits, and synthetic visual content.
Document Intelligence: Classifies document types, performs OCR, and extracts structured fields.
Visual–Text Consistency Checking: Cross-verifies that images, text, and metadata align and are authentic
Performance: Superior accuracy vs. traditional ML classifiers in production benchmarks and fraud simulations.
Key Performance improvements:
Incode’s Fraud LLM is trained on proprietary fraud datasets, identity metadata, behavioral sequences, device signals, and transactional flows. Through few-shot and transfer learning, it adapts rapidly to emerging fraud tactics and contextual manipulation attempts. It interprets complex, multi-source patterns in real time to uncover hidden anomalies and intent.
.avif)
What it powers:
Anomaly & Pattern Detection: Detects unusual sequences, behaviors, and fraud tactics.
Risk Signal Extraction: Derives structured insights from messy, multi-source data.
Fraud-Intent Classification: Distinguishes benign user mistakes from coordinated fraud attempts.
Performance: Early internal testing shows significant gains over traditional rule-based and ML classifiers.
Reasoning Agents are trained on hundreds of signals and millions of labeled identity and fraud outcomes. Using reinforcement learning and client-specific histories, they refine decision boundaries over time. They orchestrate outputs from VLMs, LLMs, device telemetry, and behavioral insights into a unified, context-aware risk assessment.
.avif)
What it powers:
Holistic Risk Decisions: Combine multi-modal signals to deliver real-time, context-aware outcomes.
Signal Coordination & Conflict Resolution: Weigh and balance signals when they disagree.
Adaptive Verification: Tailor decision logic to each client’s risk profile and tolerance.
Performance: Designed to minimize human intervention while improving error rates. Early Risk AI Agent results show reduced fraud and higher approval rates for genuine users.
Mexico:
USA:
Incode’s multimodal and agentic models rely on a strong data foundation covering global coverage, structured training pipelines, early exposure to high-fraud environments, and real-time network intelligence. The following sections outline how these data layers support accurate, secure, and adaptive AI model performance.

Incode trains resilient models with data from government databases, enterprise integrations, and billions of global verifications. This scale provides the diversity and depth required to train models with broad regional coverage and resilience against a wide range of fraud tactics.
Building on this global data access foundation, Incode uses structured labeling, synthetic data generation, and continuous stress-testing to turn raw signals into high-quality training data for fraud-resilient AI models.
Human and automated review across millions of records.
200+ human labelers creating training data and measuring AI performance for each customer.
We started in Latin America, one of the world’s highest-fraud regions, covering 66% of the adult population. This early exposure to complex, large-scale fraud hardened our models from the start.
Building on this foundation, Incode now serves 700+ enterprises worldwide, including 8 of the top 10 banks in North America and reaching about 65% of U.S. adults. The same models hardened in high-fraud markets are now deployed globally, supporting customers across industries and regions.
Because new fraud techniques often appear first in LATAM, our strong presence there gives us an early-warning advantage, helping models adapt faster and deliver stronger fraud prevention before threats spread to other regions.

Incode’s Trust Graph is a privacy-preserving technology that connects otherwise separate hundreds of millions data silos across governments and enterprises. By linking these signals safely, it uncovers hidden patterns, helping to identify serial fraudsters and organized crime that would otherwise go undetected.
This network effect not only improves fraud detection in real time but also increases the density and diversity of training data, making Incode’s models stronger and more adaptive over time.
This in-house technology powers Trust Graph by enabling instant search across hundreds of millions of identity embeddings generated by our recognition models. Optimized for speed, it delivers sub-20 ms response times with full recall while maintaining distributed reliability and in-memory performance.

Identity density expresses how confidently a user’s identity can be confirmed. Incode measures it by combining deterministic records with probabilistic AI signals, powered by our global data foundation, multimodal models, and Trust Graph intelligence.
We started in Latin America, one of the world’s highest-fraud regions, covering 66% of the adult population. This early exposure to complex, large-scale fraud hardened our models from the start.
Building on this foundation, Incode now serves 700+ enterprises worldwide, including 8 of the top 10 banks in North America and reaching about 65% of U.S. adults. The same models hardened in high-fraud markets are now deployed globally, supporting customers across industries and regions.
Because new fraud techniques often appear first in LATAM, our strong presence there gives us an early-warning advantage, helping models adapt faster and deliver stronger fraud prevention before threats spread to other regions.
Incode’s Network
400M+ identities confirmed by Incode, (e.g., 65% of USA adults).
Biometrics SOTs
15+ connections to biometrics government sources of truth.
Multimodal and agentic models VLMs, LLMs and Intelligent Agents extend coverage to new identities outside deterministic sources.
Together, deterministic and probabilistic sources help Incode create denser identity coverage by adding more datapoints, more signals, and greater certainty when verifying an identity.

End-to-end face perception that detects faces, creates robust embeddings, and matches identities at scale via a vector engine, continuously improving through calibration and hard‑case mining.

Detects and localizes faces in selfies and document images, serving as the foundation for downstream tasks such as recognition, liveness, and document validation. The model is trained to handle varied image conditions, including rotation, occlusions, and non-human distractors. Evaluated on datasets covering selfies, IDs, rotated samples, negatives, and non-human inputs.
Performs one-to-one biometric verification by comparing facial embeddings from a live selfie to embeddings extracted from a government document portrait or from an image of an identified person in the Incode system. The model is optimized to minimize both false accepts and false rejects under strict thresholds. Evaluated on a dataset of 5.8M+ selfie–document pairs
Performs one-to-many biometric search by embedding a live selfie into a high-dimensional feature space and comparing it against a gallery of enrolled identities. The model is designed for scalability and efficiency, supporting large databases while maintaining strict accuracy thresholds. It minimizes false accepts and false rejects through optimized indexing and similarity scoring.
A high-performance vector database for identity matching. It enables fast 1:1 authentication and 1:N identification with sub-20 ms responses on hundreds of millions of vectors. Built in C++ with HNSW indexing, FaceDB offers distributed reliability, flexible index options, and in-memory performance with manual scaling.
Multi-modal defenses that distinguish real users and physical documents from presentation attacks and deepfakes using spatial, temporal, and device-aware signals with continual hard‑negative training.
.avif)
Face Liveness: Detects whether a selfie comes from a live human rather than from a presentation attack(photo, screen replay, mask, or deepfake). Incode's default passive liveness has been evaluated on a dataset of 150,000+ spoof attempts, covering replays, paper copies, 2D masks, and 3D masks.
Protects against identity document presentation attacks, such as a printed copy or a screen replay. Uses passive, image-based analysis during capture to detect tell‑tale artifacts of reprints and displays while keeping the user experience low‑friction. Evaluated on diverse global documents and attack scenarios, it delivers speed and precision far beyond human capabilities.
Deepfake and Gen‑AI Defense: Multi‑modal models that detect and block AI‑generated fraud , deepfakes, face swaps, document injections, and synthetic identities.
.avif)
Detects whether a selfie has been synthetically generated, altered, or injected (e.g., face morphs, swaps, or AI-generated deepfakes). Evaluated on a dataset of 40,000+ digital spoof attempts
Protects against fake identity document images generated with AI or produced from ready-made templates sold on fake document marketplaces.
Policy-ready age estimation that provides calibrated predictions with uncertainty bounds and fairness constraints, routing edge cases to secondary verification.

Estimates a user’s age from a selfie to support age-based compliance with low friction. Designed and monitored for demographic fairness, with performance validated in external evaluations and internal bias analyses, trained on data from over 200,000 images.
Document understanding that classifies type, extracts and validates OCR, MRZ, and barcodes, and detects tampering, fusing signals into a document authenticity score that adapts with active learning.
.avif)
Identifies an identity document’s type and issuing authority by analyzing visual layouts and text cues. Proposes likely candidates from visual features, then refines using text signals to distinguish look-alike templates. Outputs the final type and issuer with confidence, and can return a ranked list of top candidates. Evaluated on a large dataset of global identity documents.
Protects against tampered identity documents — spotting portrait swaps, altered or covered text fields, and digital manipulations made with photo-editing tools.
Assesses whether a document’s text fields are readable for automated data extraction. Using a cropped document image and a text‑zone mask, the model classifies each sample as unreadable, no text fields of interest, or readable. Signals include text‑region contrast, character‑level structure cues, and artifact sensitivity tuned for OCR readiness.
Detects and crops identity documents from camera frames in real time on mobile and web, returning a standardized, perspective‑corrected document region. Supports multiple orientations, partial occlusions, and varied backgrounds; it produces a tight, consistent crop for downstream processing.
Ensures that barcode data extracted from identity documents is correct, complete, and compliant before use downstream.
A real-time orchestration layer that combines model outputs with fraud network intelligence and AI risk agents to score and route risk decisions, optimizing thresholds through continuous feedback and counterfactual evaluation.
.avif)
A high-performance fraud detection system that fuses 250+ signals across face, document, liveness, and behavioral telemetry into a single fraud probability, delivering enterprise-grade accuracy with built-in interpretability, adaptability and sel-improving capabilities.
A cross-ecosystem fraud intelligence defense that reveals repeat and coordinated fraud. It links entities such as faces, device fingerprints, sessions, and document identifiers, detects reuse and risky relationships in real time, and surfaces patterns that drive step-up or block decisions.
A model designed to block advanced fraud attempts that try to bypass liveness and face-recognition systems. It targets adversarial behaviors such as extreme expressions, partial or half-masks, occlusions, and other attempts to manipulate on-device capture.
Analyzes device environments and user interaction signals to strengthen fraud defenses. Blocks injected or emulated environments, flags abnormal interaction patterns, and uses device and cross-session intelligence to link related entities.
.avif)
Focuses on interaction patterns within supported modules to assess authenticity. Examples include gesture patterns and flow anomalies. Helps surface scripted or replayed activity and unusual usage indicative of automation or fraud
Evaluates the integrity of the device and capture environment. Inputs include hardware and OS traits, browser and app attributes, IP and network indicators, emulator and virtual camera flags, and other sensor‑level signals. Detects compromised, emulated, or suspicious environments to prevent unauthorized access and fraud.
Comprehensive governance framework covering data practices, security, model development, fairness, and compliance to ensure responsible AI

Incode joins the OpenAge Initiative alongside leading identity providers to accelerate interoperable, privacy-preserving age assurance with reusable AgeKey credentials that help platforms meet global standards and protect user privacy.
Orchestrate identity verification, compliance, and fraud prevention in one platform designed to grow with your business.