← Back to research

Research report · March 2026

AI Cheating in Hiring Assessments: The 2026 Statistics Report

A comprehensive, annually updated dataset covering AI-assisted cheating rates, fraud trends, tool ecosystems, regional breakdowns, behavioral detection signals, and the financial cost to employers — compiled from primary research and leading industry sources.

By Akshay Aggarwal and Divya BhanushaliEndorsed by Graham Hudson, Chairman, eAA Assessment Association

Last updated March 2026 · Data through January 2026 · Next revision: Q3 2026

Executive summary

The five sentences that summarize 2026.

Assessment fraud has crossed from edge case to structural crisis. In the year ending December 2025, AI-assisted cheating rates more than doubled. By early 2026, Talview's AI Threat Index reported that 88% of online exams and assessments now face active AI cheating risk. The tools enabling this behavior are no longer experimental — they are a mature SaaS ecosystem with monthly subscribers, community forums, and open-source forks that anyone can compile and deploy with a custom process name, invisible to existing detection systems.

This report compiles verified data from the platforms, researchers, and surveys closest to the problem. Every statistic is attributed to its primary source. Where figures conflict across sources, we note the discrepancy and report the most conservative figure.

The core finding: one in three candidates on monitored assessments attempted some form of fraud in 2025. Among entry-level technical roles, the rate approaches one in two. Detection-only approaches are failing because the tools now operate below the visibility threshold of every current sensor. A renamed binary calling the OpenAI API looks identical to a legitimately-named one; both send the same HTTPS request to api.openai.com. Process names are irrelevant when the DNS query itself never resolves.

Key findings at a glance

Twelve numbers that define the 2026 assessment-integrity landscape.

Each figure is attributed to its original publisher. Where primary sources disagree we report the most conservative figure. Where estimates are derived rather than directly measured, we label them clearly.

35%
Proctored-assessment fraud attempt rate
Source: CodeSignal, 2025
+119%
Year-over-year jump in fraud attempts (16% → 35%)
Source: CodeSignal, 2024–2025
88%
Assessments now facing active AI cheating risk
Source: Talview AI Threat Index, 2026
83%
Candidates who would use AI if they believed detection was unlikely
Source: TestPartnership, 2025
48%
Cheating rate in technical / engineering roles (19,368 interviews)
Source: Fabric, 2026
61%
Cheaters who pass undetected without behavioral detection
Source: Fabric, 2026
4x
Score-inflation gap: unproctored vs. proctored assessments
Source: CodeSignal, 2025
40%
Entry-level assessment fraud rate (campus / new grad hiring)
Source: CodeSignal, 2025
60–80%
Estimated fraud rate in unproctored assessments (derived)
Source: CodeSignal + Fabric, 2025–2026
$42k–$125k
Cost of one bad hire — junior engineering role
Source: SHRM / US Dept. of Labor, 2025
$200k–$750k+
Cost of one bad hire — senior / VP-level role
Source: Millman Search / SHRM, 2025
0
Detection methods with no known, documented bypass
Source: Academic literature + vendor docs, 2026

§1/Fraud rates

Assessment fraud rates: the 2025–2026 picture

Three independent datasets published between February and March 2026 now tell a consistent story. Each captures a different slice of the assessment funnel — structured coding platforms, AI-conducted interviews, and the broader online exam market — but converge on the same conclusion: cheating has become the default assumption, not the exception.

35%
Of candidates on proctored assessments attempted cheating or fraud in 2025 — up from 16% in 2024
Source: CodeSignal, across millions of assessments, Feb 2026
88%
Of online exams and assessments now face active AI cheating risk
Source: Talview AI Threat Index Report, 2026
38.5%
Of all candidates flagged for cheating behavior in AI-conducted live interviews (19,368 analyzed)
Source: Fabric, Jan 2026

Year-on-year trajectory

PeriodFraud attempt rateSourceNotes
2024 (full year)16% (proctored)CodeSignalBaseline year
H1 2025~15–20%Fabric trajectoryEarly-stage adoption
H2 202535% (Dec 2025)Fabric / CodeSignal3x jump — invisible overlay tools went mainstream
Full year 202535% proctored, 40% entry-levelCodeSignalDoubles from 2024
Early 2026 (all assessments)88% face AI cheating riskTalviewIncludes passive risk — not all are active fraud attempts

Fraud rates by role and seniority

SegmentFraud / cheating rateSource
Technical / engineering roles48%Fabric, 2025–2026
Sales / non-technical roles12%Fabric, 2025–2026
Entry-level (0–5 years experience)40% on proctored; higher unproctoredCodeSignal, 2025
Junior vs. senior~2x junior rate over seniorFabric, 2025–2026
Campus / new grad hiring40% — highest risk segmentCodeSignal, 2025
Take-home / unproctored assessmentsEstimated 60–80% (see §5)Derived from 4x score-gap data

§2/Candidate attitudes

The psychology of the problem

83%
Of candidates say they would use AI assistance in assessments if they believed detection was unlikely
Source: TestPartnership survey, reported by HackerRank, 2025

This figure is the most important single data point in the assessment integrity space. It reveals that the primary inhibitor to cheating is not ethics — it is perceived risk. When that risk drops to near-zero (as modern overlay tools claim), adoption becomes near-universal.

Only 14% of candidates openly admit to having already used generative AI in an assessment context — a gap that researchers call the "dark figure." Based on academic survey methodology, self-reported rates undercount actual behavior by 2.5x to 3x. Applied to the 14% admission rate, the estimated true current usage rate is 35–42% — consistent with CodeSignal's measured detection rate of 35%.

Why candidates cheat: reported motivations

MotivationEvidence
Competitive pressure / FOMO3x cheating surge in H2 2025 coincided with viral social media showing AI-assisted job offers (Fabric)
Low perceived detection risk83% would cheat if undetected — risk calculus is the primary inhibitor (TestPartnership)
Tool accessibility and economics$20–50/month subscription vs. $150k+ salary — the ROI is overwhelming (Fabric)
Normalization of AI in daily work80% of Gen Z use AI daily; the line between tool use and cheating feels arbitrary (CodeSignal)
Prisoner's dilemma effectGenuine candidates feel forced to cheat because they assume competition already is (Fabric)
Perception that tests are arbitraryCandidates feel LeetCode-style assessments don't reflect real job work (CNBC, 2025)

§3/Behavioral signals

What AI-assisted cheating looks like in the data

The key insight from 2025–2026 data is that AI-assisted cheating does not produce the behavioral fingerprints of traditional cheating. There is no visible tab-switch, no obvious paste event, no physical note-taking. The signals are subtler and require multi-dimensional analysis to surface reliably.

CodeSignal: behavioral signals from flagged sessions (2025)

CodeSignal's detection system identified three primary behavioral signatures across all flagged assessment sessions in 2025.

Behavioral signal% of flagged sessionsWhat it indicatesDetection difficulty
Frequent off-screen referencing35%Candidate attention not on the screen — secondary device, phone below webcam, or physical notesMedium — visible via gaze tracking but requires calibration
Unusually linear typing patterns23%Complex code produced with minimal pauses, zero debugging, or error-correction — the pattern of someone reading and transcribing, not problem-solvingHigh — requires keystroke dynamics baseline to distinguish from fast typists
Elevated similarity to known content15%Code matches known AI-generated outputs, GitHub solutions, or flagged submissions from the same periodMedium — but rapidly degrading as AI generates novel code variants each query

Fabric: behavioral signals from 19,368 live interviews (2025–2026)

Fabric's multi-signal analysis of nearly 20,000 AI-conducted interviews identified the following patterns as most predictive of cheating behavior.

  • Gaze deviation patterns: Eyes consistently moving to a fixed off-center position (the overlay location) rather than toward the webcam or screen center — the "reading" eye pattern vs. the "thinking" eye pattern.
  • Response timing anomalies: Answer latency 1–2 seconds after question close, regardless of question complexity. Human cognition produces variable latency; AI pipelines produce consistent latency. When a candidate answers a complex system design question in 1.8 seconds, the same time it takes to answer "what is your name," that is a detection signal.
  • Rubric-perfect answers: AI models are trained on the exact textbooks and documentation hiring managers use to write scoring rubrics. AI-assisted answers are structured to score maximally — STAR format for behavioral questions, O(n log n) complexity analysis for algorithmic questions, CAP theorem citations for distributed systems.
  • Artificial confidence smoothing: Removing the AI overlay from a cheating candidate during testing (via adaptive follow-up questions) immediately exposes stuttering, inability to elaborate, and inability to answer paraphrased versions of the same question.
  • Language register inconsistency: Written or spoken vocabulary significantly exceeding the level demonstrated in unstructured conversation. The candidate who types perfect code but says "um, so I was like…" when asked to explain it verbally.
  • No incremental problem decomposition: Human coders break problems down, test assumptions, hit dead ends, and backtrack. AI-generated solutions appear fully formed.

§4/Threat landscape

The AI cheating toolkit: a complete map

By 2026, AI cheating tools are a mature consumer software market. They have pricing tiers, customer support, feature roadmaps, community forums, and — critically — open-source forks that any developer can compile, customize, and deploy with a completely different process name and binary signature. This last point is the existential threat to detection-first approaches.

Category 1: real-time invisible screen overlays

The primary innovation of 2025 was the mainstream adoption of invisible overlay tools. These use low-level graphics hooks to render content that exists only on the candidate's local display. When a candidate shares their screen via Zoom, Teams, or Google Meet, the video encoding captures the desktop beneath the overlay — the interviewer sees only the code editor or assessment window.

ToolPriceKey claimKnown limitationsStopped by network-layer enforcement?
Cluely$20/monthInvisible overlay + audio pipelineData breach May 2025 exposing 83,000 users' transcriptsYes — all LLM API calls blocked at kernel layer
Interview Coder$100 lifetime100,000+ users; zero documented detections on proper use; invisible in Activity Monitor and dockCannot adapt to follow-up questions mid-solutionYes — requires internet access to generate solutions
Parakeet AI$20–40/monthReal-time audio → structured answers via GPT-4/Claude; 50+ languagesCredit system glitches can expose overlay to IDEYes — transcription + LLM calls both blocked
Ultracode AI$899 lifetimeHandles verbal, system design, and coding; 'invisible even on full screen share'Visible in Windows taskbar; no click-through overlayYes — LLM backend requires internet
LockedIn AI$55–70/monthCloud-based assistantEvery session transcript passes through vendor serversYes — cloud dependency is the point of failure
Final Round AI$149/monthAudio earpiece + overlayTaskbar icon visible to some proctoring softwareYes — transcription and answer generation require internet
Open-source forks (OpenCluely, Pluely, Natively)FreeCustom process name — defeats all signature detectionRequires developer skills to compile; trivial for technical candidatesYes — regardless of process name, LLM calls require internet

Category 2: open-source versions — the undetectable frontier

The most significant development in the threat landscape is not any commercial tool — it is the proliferation of open-source clones that can be compiled with arbitrary process names and signatures. These tools are explicitly designed to be undetectable not just today, but persistently.

ToolPlatformKey capabilityWhy it defeats process-name detection
OpenCluelyGitHub (open source)Invisible overlay for DSA/coding; multi-language; local AI via GeminiFully customizable — compile with any process name or binary signature
PluelyGitHub (open source, Tauri/Rust)10MB; 50% less RAM than Cluely; invisible in Zoom/Teams/Meet; multi-LLM (GPT, Claude, Gemini, Grok)Source code available — any developer can rename and recompile in under 1 hour
NativelyGitHub (open source)Local RAG; BYOK; zero server storage; disguises process as Terminal/Activity Monitor/System SettingsExplicitly designed to disguise process name as system utilities — documented feature
MindWhisperAIGitHub (open source)GPT-5/Claude 4/Gemini 2.0; stealth mode; handles coding, system design, behavioralMIT license — free for any use; no telemetry; fully forkable
DIY Python overlayAny developer, ~2 hoursTesseract OCR + Whisper STT + Gemini API = functional Cluely equivalentNo signature exists for a custom-built tool

TechCrunch, April 2025 / ARES architecture documentation

Category 3: tools by interview type — now spanning all rounds

Early AI cheating tools focused exclusively on LeetCode-style coding questions. By 2026, the ecosystem has expanded to cover every round of the modern hiring funnel, including rounds specifically designed to be AI-resistant.

Interview roundAI assistance availableHow it works
Technical coding (DSA/LeetCode)All toolsOCR captures problem; LLM generates solution with time/space complexity analysis; overlay displays it
System designUltracode, Parakeet, ShadeCoder, all open-source toolsAudio pipeline transcribes prompt; LLM generates architecture with CAP theorem tradeoffs, scaling considerations, component diagrams
Behavioral / STARCluely, Final Round AI, Parakeet, Linkjob AIAudio captures question; LLM generates STAR-format response with specific metrics in ~1 second
PM case studies / product senseParakeet, Ultracode, open-source toolsScreen capture + audio; LLM generates frameworks (Jobs-to-be-Done, CIRCLES), metrics, prioritization rationale
Structured problem-solving (senior/consulting)All audio-capable toolsAudio captures problem framing; LLM generates hypothesis tree, clarification questions, structured synthesis
Take-home projects / code assignmentsChatGPT, Claude, Copilot directlyNo overlay needed — unlimited time, no monitoring

§5/The unproctored gap

Estimating true fraud rates in unmonitored assessments

4x
Score inflation gap between unproctored and proctored assessments — the strongest available evidence of mass cheating in unmonitored environments
Source: CodeSignal, 2025

The statistics cited throughout this report primarily reflect proctored or AI-monitored environments. The majority of technical assessments globally are conducted without dedicated proctoring — take-home assignments, asynchronous coding challenges, and automated screening tests that rely on time limits and the honor system.

CodeSignal's data provides the clearest signal: unproctored assessments showed score increases more than 4x larger than proctored ones when comparing matched cohorts. This is not explained by selection effects — it is the direct signature of unrestricted AI use. Applying the Fabric trajectory (15% → 35% in 6 months in monitored environments) to the unproctored context produces an estimated true fraud rate of 60–80% for asynchronous, unmonitored technical assessments.

Unproctored fraud rate estimates by assessment type

Assessment typeEstimated fraud rateBasis
Live proctored coding assessment35–40%CodeSignal measured rate (2025)
Live AI-conducted interview38.5%Fabric measured rate (2025–2026)
Live human-conducted interview (unmonitored)48%+ for technical rolesFabric measured rate, no behavioral detection
Asynchronous take-home coding challenge60–80% (estimated)4x score gap (CodeSignal) + Fabric trajectory
Automated screening quiz (MCQ, no time pressure)70–85% (estimated)AI tools solve most MCQ assessments in seconds

§6/Regional breakdown

Why the APAC number is a North American problem

48%
Assessment fraud attempt rate in Asia-Pacific — nearly 1 in 2 assessments flagged
Source: CodeSignal, 2025

The geographic data in this space is frequently misread. The 48% fraud attempt rate in Asia-Pacific is often treated as a regional concern — a matter for APAC hiring managers. That reading misses the mechanism entirely.

The overwhelming majority of AI cheating in APAC is targeted at North American employers. US tech companies offer the highest compensation packages in the world — $120,000–$200,000+ base salaries for entry- and mid-level engineering roles. These packages represent a 10x–30x earnings multiple compared to domestic tech salaries in India, the Philippines, Indonesia, and other major candidate-sending countries. Remote-first hiring has created a globally accessible pathway to these packages.

RegionFraud attempt ratePrimary target employerEconomic driver
Asia-Pacific48%US / North American tech companies10x–30x salary differential vs. domestic tech roles
North America27%US employers (domestic competition)Competitive job market
Global average (proctored)35%All major employersAcross all CodeSignal monitored assessments, 2025

§7/Detection tools

Detection tools and their fundamental limitations

The emergence of process-name detection tools (Honrly, Truely, Proctaroo) in 2025 was a direct response to Cluely's viral moment. These tools serve an important market need and do catch unsophisticated users. But they operate on an architecturally flawed premise.

Detection tool comparison

Tool / approachWhat it detectsWhat it missesFundamental limitation
HonrlyKnown tool signatures via process list scanning (Cluely, Interview Coder, ChatGPT native app)Any tool with a renamed/recompiled binary; second-device pipelines; earpiece deliverySignature-based — defeated by any developer with GitHub access and 1 hour
Truely (Validia)Cluely process signatures — triggers alarm on detectionSame as Honrly; signatures-onlyCluely CEO publicly called it 'pointless' (TechCrunch, 2025); open-source clones bypass trivially
ProctarooRunning applications and hidden background processesCustom-compiled open-source tools; second-device setupsProcess name scanning — same fundamental limitation
HonorlockApplication blocking; phone detection via webcam AISecond device outside camera view; earpieceMost robust detection approach — blocks rather than detects; still misses second-device and audio pipelines
Talview (7-layer framework)App blocking + behavioral analytics + audio analysis + identity verification + cross-session intelligenceHighly sophisticated second-device setups with behavioral mimicryClosest to comprehensive — behavioral layer adds signal beyond process names
ARES (network-layer prevention)Entire outbound network stack: WireGuard VPN; nftables 4-layer firewall; DNS filtering with AI domain classifier; SNI/JA3 packet inspection; OS-level overlay detection (6 independent vectors)Candidates on fully offline exam content; second devices physically present; hardware cheating (smart glasses, earpieces)Prevention rather than detection — closes the network path to AI tools rather than analyzing post-hoc evidence

How network-layer enforcement differs architecturally

The ARES row above deserves elaboration because "WireGuard VPN + firewall" undersells the depth. ARES enforces integrity across four independent, orthogonal layers.

LayerMechanismWhat it stops
A — Kernel firewallnftables on the ARES server: QUIC blocked; P2P between candidates blocked; DNS restricted to ARES resolver; new TLS connections queued for packet inspectionQUIC-based AI clients; DNS hijacking; any traffic not going through the exam server
B — Client-side integrityWindows Firewall / macOS pf rules checked and auto-remediated every 15 seconds; FIREWALL_TAMPER triggers immediate REVIEWAttempts to disable local firewall rules after ARES agent installs them
C — DNS filtering + IP allowlistEvery DNS query through the ARES DNS server; unknown domains classified by Gemini AI in real time; only exam-whitelisted IPs can receive TCP connectionsDirect-IP connections bypassing DNS; AI API calls via hardcoded IPs
D — SNI/JA3 deep packet inspectionPython sidecar reads TLS ClientHello; SNI checked against whitelist; JA3 fingerprint matched against 9 known RAT profiles (Cobalt Strike, AsyncRAT, Quasar, DCRat, etc.)TLS connections to non-whitelisted domains; remote access tool C2 channels even if domain resolves

§8/Financial cost

What each bad hire actually costs

30%+
Of first-year salary is the US Department of Labor's minimum estimate for the cost of a bad hire — and that's the floor, not the ceiling
Source: US Department of Labor / SHRM, 2025

The financial consequence of AI-assisted cheating is not a failed test — it is a bad hire. Each bad hire propagates cost across six vectors: direct recruitment, salary paid, onboarding, productivity loss, replacement recruitment, and legal/HR overhead. The total varies substantially by seniority level.

Cost by seniority: junior engineering roles

Cost componentEstimateNotes
Direct recruitment (ads, recruiter time, agency fees)$5,000–$15,000Lower for junior roles; agency fees ~15–20% of salary
Salary + benefits during tenure$20,000–$50,0003–6 months before performance issue recognized
Onboarding and training$5,000–$15,000Senior engineer time, tooling, ramp-up investment
Productivity loss (team coverage)$10,000–$30,00050–100% of salary equivalent for underperformance period
Replacement recruitment$5,000–$15,000Full cycle must be repeated
HR / legal management$2,000–$10,000PIP documentation, potential litigation
TOTAL — junior engineering role$42,000–$125,000US DoL floor: 30% of $80k = $24,000 minimum

Cost by seniority: senior and VP-level roles

The cost curve is non-linear for senior hires. A VP-level bad hire carries a multiplier of 2x–5x their annual salary according to executive search research, reflecting broader organizational impact.

Cost componentEstimateNotes
Direct recruitment (retained search)$30,000–$80,000Retained search fees: 30–35% of base salary for $200k+ roles
Salary + benefits during tenure$100,000–$300,0006–12 months before board-level visibility
Strategic damage / missed opportunities$100,000–$500,000+Projects stalled; A-players hire B/C-players downstream; team attrition
Team morale and retention$50,000–$200,000+High performers leave; replacement cost for each A-player = 0.5–2x salary
Legal / severance$20,000–$100,000Executive contracts frequently include severance provisions
TOTAL — VP/Director-level role$200,000–$750,000+Millman Search: 2–5x annual salary

Portfolio exposure: what this means at scale

Company profileAnnual tech hiresExpected fraud (35%)Passing undetected (61%)Estimated annual bad-hire exposure
50-person startup103.5 attempts~2 passing$84k–$250k (2 junior bad hires)
Mid-market10035 attempts~21 passing$882k–$2.6M
Enterprise500175 attempts~107 passing$4.5M–$13M+

§9/Industry response

How the industry is responding

Major tech employers

  • Google: CEO Sundar Pichai addressed AI cheating at an internal town hall. The company is considering reintroducing mandatory in-person interviews for certain roles. (CNBC, March 2025)
  • Amazon: Candidates required to sign attestation acknowledging unauthorized tool policies before assessments. (CNBC, 2025)
  • 59% of hiring managers now suspect candidates of using AI to misrepresent their abilities during live assessments. (Fabric, 2026)
  • FBI warnings: Formal alerts issued about state-sponsored actors using deepfakes and AI-assisted job applications to infiltrate corporate networks and steal intellectual property.

Assessment platforms

PlatformResponse
CodilityLaunched similarity detection comparing submissions against historical and AI-generated solutions. Also integrated AI Copilot tools to assess legitimate AI collaboration skills. (Codility blog, Jan 2026)
HackerRankReports 93% accuracy using multi-signal behavioral analysis combining ML with keystroke dynamics. (HackerRank, 2025)
CodeSignalProprietary Suspicion Score — 10 years of refinement across millions of assessments — covering plagiarism, proxy test-taking, unauthorized AI use, and identity fraud.
Talview7-layer trust infrastructure: identity verification, secure browser controls, behavioral biometrics, session monitoring, content analysis, cross-session intelligence, human oversight. AI Threat Index 2026 published.

Prevention-layer infrastructure: a new category

Separate from detection tools, a distinct architectural category is emerging: assessment environments that enforce network isolation rather than trying to detect AI use after it occurs. Rather than asking whether a candidate used AI — a question with inherently imperfect answers — these systems enforce that the candidate's machine cannot reach external AI APIs during the session.

Aiseptor ARES routes all candidate traffic through a per-session WireGuard VPN tunnel enforced at the OS level via a native agent installed before the assessment. Egress is controlled by nftables firewall rules server-side. DNS is filtered to a per-exam whitelist with AI-powered real-time domain classification. TLS connections are inspected via SNI deep packet inspection and JA3 fingerprinting. The candidate's machine cannot reach ChatGPT, Claude, Gemini, or any AI API — not because these are detected, but because all outbound traffic is filtered at the kernel layer.

Regulatory developments

  • California (October 2025): Fair Employment & Housing regulations banned AI-based facial expression assessments in hiring.
  • EU AI Act: Classifies certain AI applications in employment as high-risk, requiring disclosure and specific controls.
  • Gartner projection: By 2028, one in four candidate profiles will be entirely fabricated.
  • Experian 2026 Fraud Forecast: Lists deepfake job candidates as one of the top five fraud threats for 2026; 60% of companies reported increased fraud losses from 2024 to 2025.

§10/Forward projections

What 2026 and beyond look like

ProjectionBasisTimeframe
Cheating becomes the statistical norm in technical hiring (>50% attempt rate)Fabric trajectory: 15% → 35% in 6 months; doubling period shorteningLate 2026
Open-source tool proliferation makes signature-based detection obsolete20+ GitHub repos as of March 2026; trivial to customize process nameAlready underway
1 in 4 candidate profiles entirely fabricatedGartner projection; deepfake video + synthetic voice convergence2028
Behavioral multi-signal detection adopted but recognized as insufficientProcess-name detection bypassed by open-source forks; behavioral signals have documented bypasses2026
Network-layer prevention emerges as the architecture that closes the loopNo detection approach survives a determined bypass2026–2027
Identity verification becomes standard step in hiring funnelFBI warnings + Experian 2026 Fraud Forecast + deepfake prevalence2026
Hardware cheating tools (smart glasses, earpieces) go mainstreamCluely CEO publicly stated intent to build hardware bypass products2026–2027
Regulatory mandates for AI disclosure in hiring expand to 10+ US statesCalifornia precedent + all 50 states considered AI legislation2026–2027

§11/The bypass map

Why every detection approach has a bypass — and what does not

This section synthesizes the threat landscape documented in Sections 3 through 7 into a single honest conclusion: every approach designed to detect AI misuse after the fact has a documented, working bypass available to any motivated candidate in 2026. This is not a matter of current tools being poorly implemented. It is a structural property of the detection-after-the-fact paradigm.

The bypass map: what each detection method misses

Detection methodHow it is bypassedVerdict
Process-name signature scanning (Honrly, Truely, Proctaroo)Compile any open-source fork (OpenCluely, Pluely, Natively) with a custom binary name. Under one hour with basic developer skills.Defeated — architecturally irrelevant against technical candidates
Browser lockdown / secure browserOverlay tools are native OS applications, not browser extensions. Browser restrictions have no authority over OS-level processes.Not applicable — wrong abstraction layer
Keystroke dynamics analysisRead AI output from overlay, type manually. arXiv 2601.17280 (2026) confirmed manual transcription produces keystroke patterns statistically indistinguishable from genuine composition.Defeated — motor signals confirm a human typed; not that a human composed
LLM output fingerprinting / perplexity scoringRun AI output through a humanizer pipeline. LeetcodeWizard ships a humanizer by default targeting perplexity normalization.Partially defeated — arms race currently favors evasion
Response timing analysisIntroduce deliberate pauses manually. Works against automated pattern-matching.Defeated at scale — works against automated systems only
Gaze / eye trackingPosition overlay directly below webcam. Reading gaze appears to be eye contact with camera.Defeated by tool positioning
Adaptive follow-up questioning(a) Candidate studies their AI-generated answer before follow-up. (b) Audio pipeline continues assisting during verbal follow-up. (c) Senior candidates have general domain knowledge.Partially effective — best current human-judgment method, incomplete coverage

ARES: remove the network path, not detect its use

Threat vectorARES responseWhy no bypass exists
Candidate queries any AI API (ChatGPT, Claude, Gemini, Copilot)DNS filtering on the ARES server returns NXDOMAIN for AI service domains. Gemini AI classifier categorizes unknown domains in real time.DNS is filtered server-side. Client-side DNS changes blocked by nftables. DNS-over-HTTPS disabled at the registry level.
Overlay tool with renamed binary (defeats signature detection)Overlay calls its LLM backend over HTTPS. WireGuard routes that traffic through the ARES server. SNI DPI reads the TLS ClientHello — target domain visible before encryption.Process name is irrelevant. ARES inspects network traffic, not process lists.
Candidate routes traffic through a second VPN or proxyAll network adapters except the ARES WireGuard tunnel are disabled at join time by the agent.NetworkEnforcer disables all NICs except the VPN adapter.
Candidate uses direct IP connection to bypass DNSLayer C (IP conntrack allowlist) is built from DNS-resolved addresses of whitelisted domains. Direct TCP connections to IPs not in this set trigger IP_BLOCK_EVASION.IP allowlist built server-side from DNS resolution results.
Audio pipeline — AI answers via earpiece from enrolled machineTranscription service and LLM backend both require internet.Both calls blocked before completing.
Separate physical device (phone on cellular)ARES does not control unenrolled hardware. A second physical device with its own cellular connection can access AI tools.The honest gap. OverlayScanner and DisplayGuard provide partial coverage; behavioral signals (gaze, timing) add signal.

§12/Methodology

Data sources and methodology

This report compiles data from primary sources only. All statistics are attributed to their original publisher. Where we note a figure, the source is named.

SourceData typeDate
CodeSignalPlatform data — millions of proctored assessmentsFeb 25, 2026
FabricInterview analysis — 19,368 AI-conducted interviewsJan 2026
TalviewAI Threat Index Report 2026March 2026
TestPartnershipCandidate survey2025
HackerRankPlatform data + industry survey compilationNov 2025
ResumeTemplatesCandidate self-report survey2025
SHRMBad hire cost researchOngoing
US Dept. of LaborBad hire cost benchmark (30% of first-year salary)Ongoing
Millman SearchExecutive bad hire cost (2–5x annual salary for VP)July 2025
Toggl Hire 2025 ReportHR professional survey — bad hire costs2025
TesseonTotal bad hire cost estimate ($240,000+)May 2025
CNBCGoogle / Amazon AI cheating reportingMarch 2025
TechCrunchHonrly / Truely / Cluely detection arms race reportingApril 2025
FBI warningsState-sponsored actor hiring fraud alerts2025
Experian 2026 Fraud ForecastDeepfake job candidates as top fraud threatJan 2026
Gartner1-in-4 fabricated profiles projection by 20282025
GitHub (OpenCluely, Pluely, Natively, et al.)Open-source tool documentation and capabilities2025–2026
AllAboutAIFalse positive rates by demographic2026
arXiv 2601.17280Keystroke dynamics cannot confirm content provenanceJan 2026
LeetcodeWizardHumanizer tool documentation (vendor marketing)2025–2026

Confidence levels

ConfidenceApplies toSource class
HighProctored fraud rates (35%, 40%, 48%, 38.5%, 88%)Platform-measured, millions of sessions or 19,368 interviews
HighBad hire cost rangesSHRM, US DoL, Millman Search — consistent across 3+ primary sources
MediumCandidate attitude data (83%)Single large survey; validated by CodeSignal's measured-vs-admitted gap
Derived estimateUnproctored fraud rate (60–80%)CodeSignal 4x score gap + Fabric trajectory extrapolation
ProjectionForward-looking items in §10Sourced from Gartner, Experian, Fabric trajectory — noted as projections

How to cite

How to cite this report

Use either format below. Both include the canonical URL, which is used for schema.org Dataset and Article metadata.

BibTeX

aiseptor-2026.bib
@techreport{aiseptor2026cheating,
  author      = {Aggarwal, Akshay and Bhanushali, Divya},
  title       = {AI Cheating in Hiring Assessments:
                 The 2026 Statistics Report},
  institution = {Aiseptor},
  year        = {2026},
  month       = {March},
  url         = {https://aiseptor.com/research/
                 ai-cheating-statistics-2026}
}

APA 7th

Aggarwal, A., & Bhanushali, D. (2026, March). AI cheating in hiring assessments: The 2026 statistics report. Aiseptor. https://aiseptor.com/research/ai-cheating-statistics-2026

Endorsement

Endorsed by Graham Hudson, Chairman, eAA Assessment Association.

About the authors

Founder & CEO, Aiseptor

Akshay Aggarwal

10 years in offensive cybersecurity and bug bounty research; $300,000+ in awarded bounties from major platforms. Authored the ARES network-layer enforcement architecture. Leads primary research on kernel-level assessment security. Cornell University.

AI Safety & Red Team Researcher

Divya Bhanushali

Specialist in AI red-teaming and adversarial evaluation. Reverse-engineered the invisible-overlay architectures used by leading cheating tools and documented their network-dependency chain. Leads Aiseptor's threat-intelligence dataset.