Categorical Perception in Large Language Model Hidden States: Structural Warping at Digit-Count Boundaries

Categorical perception (CP) -- enhanced discriminability at category boundaries -- is among the most studied phenomena in perceptual psychology. This paper reports that analogous geometric warping occurs in the hidden-state representations of large l…

Authors: Jon-Paul Cacioli

Categorical Perception in Large Language Model Hidden States: Structural Warping at Digit-Count Boundaries
Categorical P erception in Large Language Mo del Hidden States: Structural W arping at Digit-Coun t Boundaries Jon-P aul Cacioli Indep enden t Researc her, Melbourne, A ustralia Classical Minds, Modern Mac hines Abstract Categorical perception (CP) — enhanced discriminabilit y at category boundaries — is among the most studied phenomena in p erceptual psyc hology . This pap er reports that analogous geometric w arping o ccurs in the hidden-state representations of large language mo dels (LLMs) pro cessing Arabic n umerals. Using represen tational similarit y analysis across six mo dels from v e architec- ture families, the study nds that a CP-additive mo del (log-distance plus a boundary b o ost) ts the represen tational geometry b etter than a purely con tinuous mo del at 100% of primary la yers in every mo del tested. The eect is sp ecic to structurally dened b oundaries (digit-count tran- sitions at 10 and 100), absen t at non-b oundary control positions, and absent in the temperature domain where linguistic categories (hot/cold) lac k a tok enisation discon tin uity . T w o qualitativ ely distinct signatures emerge: “classic CP” (Gemma, Qw en), where mo dels b oth categorise explicitly and show geometric w arping, and “structural CP” (Llama, Mistral, Phi), where geometry w arps at the b oundary but mo dels cannot rep ort the category distinction. This disso ciation is stable across boundaries and is a prop ert y of the architecture, not the stim ulus. Structural input-format discon tinuities are sucient to pro duce categorical p erception geometry in LLMs, indep enden tly of explicit semantic category knowledge. Keyw ords: categorical p erception, large language mo dels, representational similarity analysis, n umerical cognition, hidden states, tok enisation 1. In tro duction Categorical p erception (CP) — the observ ation that stim uli v arying con tinuously along a physical dimension are p erceived as falling in to discrete categories, with enhanced discriminabilit y at cat- egory b oundaries — is one of the most extensively studied phenomena in p erceptual psychology . First demonstrated for sp eec h sounds b y Lib erman, Harris, Homan, and Grith (1957), CP has since b een do cumen ted across domains including colour (Wina wer et al., 2007), facial expressions (Etco & Magee, 1992), and m usical pitc h (Burns & W ard, 1978). The canonical CP signature com- prises three comp onents: (a) a sharp identication function at the category b oundary , (b) a p eak in discrimination p erformance at the b oundary that exceeds what the iden tication function alone w ould predict, and (c) represen tational w arping — increased p erceptual distance b et w een stim uli that straddle the b oundary relative to equidistan t within-category pairs (Harnad, 1987; Goldstone & Hendrickson, 2010). The theoretical signicance of CP lies in what it reveals ab out the relationship b et ween con tinuous sensory input and categorical mental representations. T wo classes of explanation dominate the 1 literature. Under acquired CP accoun ts, exp erience with categorical distinctions warps the under- lying p erceptual space, such that the warping is a consequence of category learning (Goldstone, 1994). Under structural accounts, the warping reects prop erties of the represen tational format it- self — discon tinuities in the input that imp ose categorical structure regardless of explicit category kno wledge (McMurray , 2022). Distinguishing these accounts requires access to the represen tational geometry , which is dicult in biological systems but straightforw ard in articial neural netw orks. Large language mo dels (LLMs) provide an ideal test case. Their hidden-state represen tations are directly observ able, their training data are characterisable, and their arc hitectural prop erties (tok enisation, la yer depth, atten tion structure) create well-dened represen tational discon tinuities that can b e manipulated exp erimen tally . Arabic numerals are a particularly clean domain: the transition from single-digit to double-digit num bers (9 → 10) in volv es a simultaneous change in c haracter count, tok en coun t (for most tok enisers), and lexical form — a structural discontin uit y analogous to the v oice onset time b oundary in sp eec h CP studies. Recen t work has established that LLMs enco de numerical magnitude in their hidden states fol- lo wing a logarithmic compression consistent with W eb er’s Law (Cacioli, 2026a). This geometry is presen t at all lay ers but is causally implicated only at early la yers — a disso ciation b et ween represen tational structure and functional relev ance that parallels ndings in neuroscience (Zhu et al., 2025). Separately , Park, Y un, Lee, and Shin (2024, 2025) demonstrated that LLMs represen t categorical concepts as p olytope structures in hidden-state space, and Bonnasse-Gahot and Nadal (2022) show ed theoretically that CP-like geometry — warping of the Fisher information metric at category b oundaries — emerges in deep lay ers of trained neural net works regardless of output b eha viour. What has not been tested is whether the full CP signature — iden tication, discrimination, and rep- resen tational w arping ev aluated join tly — appears in LLM hidden states for a stim ulus domain with a w ell-dened structural b oundary . The present study lls this gap, applying formal psyc hophys- ical metho dology (representational similarit y analysis, signal detection theory , coun terbalanced forced-c hoice psyc hophysics) to the hidden states of six LLMs pro cessing Arabic numerals. 1.1 The Presen t Study W e test categorical p erception at digit-coun t b oundaries (10 and 100) in six mo dels from ve ar- c hitecture families. The design comprises v e paradigms: P aradigm A (represen tational geometry via RSA), P aradigm B0 (identication), P aradigm B (discrimination), Paradigm C (precision gra- dien t), and Paradigm E (causal interv en tion). T emperature serves as a cross-domain control: a con tinuous physical dimension with a linguistic category b oundary (hot/cold) but no tok enisation discon tinuit y . A nonce-token remapping con trol tests whether the CP eect requires linguistic surface form or can b e induced by ordinal information alone. Eigh t hypotheses w ere pre-registered on the Op en Science F ramework (osf.io/qrxf3) prior to data collection, along with tw elv e exploratory analyses. The primary claim rests on representational geometry (Paradigm A), following the precedent of P ark et al. (2024) and Bonnasse-Gahot and Nadal (2022) who made geometric CP claims without requiring b eha vioural conv ergence. The iden tication task (P aradigm B0) serv es as a secondary analysis documenting the relationship b et w een explicit categorisation and geometric structure. This extends the standard CP proto col b y adding a representational criterion — justied by the unique adv antage of LLM researc h: direct access to the representational geometry , whic h in human studies must b e inferred indirectly from the relationship b et ween iden tication and discrimination (McMurray , 2022). 2 2. Metho d 2.1 Mo dels Six mo dels from ve architecture families were tested, selected to maximise arc hitectural div ersity within the constraint of tting in 16 GB VRAM (AMD RX 7900 GRE): Mo del P arameters F amily Role Llama-3-8B-Instruct 8.0B Meta Primary Mistral-7B-Instruct- v0.3 7.2B Mistral AI Primary Gemma-2-9B-IT 9.2B Go ogle Primary Qw en2.5-7B-Instruct 7.6B Alibaba Primary Phi-3.5-mini-instruct 3.8B Microsoft Primary (scale prob e) Llama-3-8B-Base 8.0B Meta Exploratory (instruction-tuning con trol) All mo dels w ere loaded in FP16 (BF16 for Gemma) with output_hidden_states=True . Phi- 3.5-mini required a communit y fork (Lexius/Phi-3.5-mini-instruct) with man ual DynamicCache compatibilit y patc hes for T ransformers 5.0.0. 2.2 Stim uli 2.2.1 Numerical Domain (Primary) Decade-10 condition. Sev enteen probing v alues (4– 20) spanning the single-digit/double-digit b oundary at 10. The b oundary is dened structurally b y the digit-coun t transition, not empirically from iden tication p erformance (see §1.1 and v0.5 revision rationale). F our of the six mo dels’ tok enisers exhibit a token-coun t discontin uit y at this b oundary: Gemma and Qw en enco de single-digit num b ers as one token and double-digit n umbers as t wo tok ens; Mistral and Phi show the same pattern oset b y a leading-space tok en (2→3 tokens). Llama-3 (b oth instruct and base) is the exception: its BPE vocabulary includes merged tokens for common m ulti-digit num bers, so b oth single- and double-digit num b ers are enco ded as single tok ens (e.g., “9”→ [24] , “10”→ [605] ). That Llama nonetheless shows strong CP geometry despite ha ving no token-coun t discon tinuit y indicates that character-coun t and lexical-form c hanges — not tok en count per se — are sucien t to pro duce the b oundary eect (see Supplemen tary T able S2 for exact token IDs across all mo dels). The 100-b oundary inv olv es an analogous transition (tw o-digit to three-digit character strings) and a token-coun t increase for Gemma, Qwen, Mistral, and Phi, pro ducing the 3.9–12.7  eect-size amplication reported in §3.1.1. Con trol-15 condition. Nine probing v alues (11–19) cen tred on 15, a non-b oundary position in the same numerical range. If CP-like warping app ears equally at 15 as at 10, the eect is general represen tational inhomogeneit y , not categorical p erception (H0 falsication con trol). Decade-100 condition. Thirteen probing v alues (70–130) spanning the double-digit/triple-digit b oundary at 100, with a matc hed con trol at 150 (Con trol-150: v alues 130–170). 3 Eigh t carrier sentences em b edded each probing v alue in natural language con texts (e.g., “Approxi- mately {N} cases w ere observ ed”). F ollo wing Rogers and Da vis (2009), sen tences w ere split: indices 0–3 for iden tication (Paradigm B0) and indices 4–7 for RSA centroids (Paradigm A), preven ting rep etition artefacts from contaminating the geometry . 2.2.2 T emp erature Domain (Secondary) Eigh teen probing v alues (  20 to 100°C) spanning the hot/cold linguistic b oundary (~22°C), with a control condition (35–51°C, no b oundary). T em- p erature has a linguistic category distinction (hot/cold) but no tokenisation discon tinuit y — “21” and “23” tok enise iden tically . The temp erature domain tests whether linguistic category kno wledge alone, without structural input-format discontin uit y , is sucient to induce representational CP . 2.2.3 Nonce-T ok en Remapping Con trol (E10) Seven teen nonce tok ens (“glorp”, “blic ket”, “tazmo”, “fen wic k”, …) mapp ed to ordinal p ositions 1–17 (corresp onding to v alues 4–20). T wo conditions: nonc e_no_or der (nonce tok ens em b edded in carrier sen tences with no ordering infor- mation) and nonc e_or der e d (a preamble establishing the complete ordering preceded each sen tence). Theoretical RDMs used ordinal p osition as the contin uous baseline (d_ij = |rank_i  rank_j|), not log-magnitude. This con trol tests whether CP requires the linguistic/tokenisation structure of real num b ers or can b e induced by ordinal information alone. 2.3 P aradigm A: Represen tational Geometry (RSA) F or eac h probing v alue, hidden states were extracted at all lay ers (em b edding + transformer la yers) from four RSA carrier sentences (indices 4–7). Centroids w ere computed by av eraging across sen tences, yielding one representation p er v alue p er lay er per mo del. P airwise cosine distances b et w een centroids formed the empirical representational dissimilarit y matrix (RDM) at each lay er. Euclidean distance w as computed as a co-primary metric. Five theoretical RDMs were compared: 1. Con tinuous (W eb er/Log): d_ij = |log(x_i)  log(x_j)| 2. CP-A dditive: d_ij = |log(x_i)  log(x_j)| + 󰄗 · 1[dierent category] ( 󰄗 = 1.0 template) 3. CP-Multiplicativ e: d_ij = |log(x_i)  log(x_j)| · (1 + 󰄎 · 1[dieren t category]) 4. Categorical: d_ij = 0 if same category , 1 if dierent 5. Linear: d_ij = |x_i  x_j| RSA w as p erformed via Sp earman rank correlation betw een empirical and theoretical RDMs, with signicance assessed b y Mantel p erm utation tests (10,000 p ermutations per la yer). Lay erwise mul- tiple comparisons were corrected using the Benjamini-Ho c h b erg false discov ery rate (FDR) at 󰄌 = .05. The pre-registered mo del comparison hierarc h y was: (1) Primary: Contin uous vs CP-Additiv e; (2) Mechanism: CP-Additiv e vs CP-Multiplicativ e. 2.4 P aradigm B0: Iden tication Three identication framings per model: “small/large”, “single-digit/multi-digit”, and “one digit/t wo digits” . Eac h framing used a chat-templated forced-c hoice format with counterbalanced A/B option order (follo wing W eb er Appendix A metho dology; Cacioli, 2026a). P(category_b) w as computed as the a verage across b oth orders to correct for p osition bias. Sigmoid functions were t to each iden tication curv e to estimate crossov er p oin t, slop e, and R². 4 2.5 P aradigm B: Beha vioural Discrimination F orced-c hoice discrimination: “Which is larger: A) x  or B) x  ?” for 900 trials (3 p osition categories [cross-b oundary , within-b elow, within-ab o ve]  6 log-distance lev els [0.10–0.60]  50 pairs per cell, with ±10% jitter; collapsed to a 2  6 factorial [cross-b oundary vs within-category  6 distances] for analysis). Both option orders w ere tested (counterbalanced) with logit extraction for A/B tok ens. The primary dep enden t v ariable was |  logit| (decision condence), used as a reaction- time analogue following the rationale that mo dels process all tok ens in a single forw ard pass and condence magnitude is the computational analogue of pro cessing ease (Cacioli, 2026b). 2.6 P aradigm C: Precision Gradien t Lo cal representational precision was dened as 1/||h(n+1)  h(n)|| for adjacen t probing v alues at eac h lay er, following Cacioli (2026a). CP predicts a distance spike (precision dip) at the category b oundary . 2.7 P aradigm E: Causal In terven tion (E5) A ridge-regression prob e was trained on Paradigm A cen troids predicting binary category mem- b ership (< 10 vs  10), dening the “category direction” at each lay er. Direction v alidity w as conrmed via PCA: PC1 correlation with category mem b ership (mean 󰄜 = .83 at primary la y ers) and prob e-PC1 cosine similarity . A ctiv ation patc hing was p erformed via PyT orch forw ard ho oks: at each target lay er, hidden states w ere mo died in-place by adding 󰄌 · ||w|| · v_cat, where v_cat is the unit category direction, ||w|| is the prob e weigh t norm, and 󰄌  {0.25, 0.50, 0.75, 1.00}. T en random-direction controls (same norm) established the sp ecicit y baseline. The dep enden t v ariable was  (|  logit|) — the c hange in discrimination condence under patching. 2.8 H4: Boundary Contribution to Represen tational Distance Hierarc hical regression at each primary lay er tested whether boundary-crossing adds unique v ari- ance b ey ond log-distance. Step 1: regress empirical pairwise distances on |log(x_i)  log(x_j)|. Step 2: add a b oundary-crossing indicator (1 if pair straddles the b oundary , 0 otherwise).  R² and F-test for the addition of the b oundary predictor. 2.9 Pre-Registration and Statistical Analysis All h yp otheses and analysis pip elines w ere pre-registered on OSF (osf.io/qrxf3) prior to data col- lection. Conrmatory tests used Bonferroni correction within h yp othesis families ( 󰄌 = .05). Eect sizes (Cohen’s d,  R², Sp earman 󰄜 ) are rep orted for all comparisons. Bootstrap 95% condence in terv als (10,000 resamples, seed = 42) are rep orted where applicable. All co de, stimuli, and analy- sis scripts are a v ailable at h ttps://anonymous.4open.science/r/weber-B02C (directory m3_pilot/ ). Seed = 42 throughout. Pre-registered exclusion rules: If discrimination accuracy is at ceiling (  99%) or chance (  51%), d’ is uninformative and h yp otheses contingen t on d’ (H2, H2b, H3, H5) are declared not ev aluable. If no iden tication framing pro duces a crosso v er within the probing range, the McMurray (2022) predicted-vs-observ ed test (H2b) is not ev aluable. 5 3. Results Results are organised b y paradigm, follo wing the analysis hierarc hy specied in the pre-registration. Section 3.1 presen ts the primary conrmatory analysis (representational geometry). Sections 3.2– 3.4 present secondary analyses (iden tication, discrimination, and precision gradien t). Sections 3.5–3.7 present the three control conditions (non-b oundary con trol, temperature domain, and nonce- tok en remapping). Section 3.8 presen ts the causal interv en tion. Section 3.9 summarises h yp othesis outcomes. 3.1 Represen tational Geometry: CP-A dditiv e Dominates at All Lay ers (H1) The primary conrmatory test compared ve theoretical RDMs against empirical represen tational dissimilarit y matrices at each lay er of eac h mo del. T able 1 summarises the decade-10 condition (single-digit/double-digit b oundary at 10). T able 1. P aradigm A results: decade-10 b oundary . CP-Additiv e vs Contin uous comparison across six mo dels. “CP > Cont” rep orts the n umber of primary lay ers where the CP-A dditive mo del ac hieved higher Sp earman 󰄜 than the Contin uous mo del. “Mean 󰄜 ” is the mean 󰄜 dierence across primary la yers. All Man tel p erm utation tests (10,000 p erm utations) w ere signican t at p < .001 at all primary lay ers after Benjamini-Ho c hberg FDR correction at 󰄌 = .05. Mo del Primary La yers CP > Cont Mean 󰄜 Max 󰄜 (CP-A dd) Max 󰄜 (Cont) Llama-3-8B- Instruct 17/17 17/17 +0.023 0.940 0.921 Mistral-7B- Instruct 17/17 17/17 +0.060 0.929 0.889 Gemma-2-9B- IT 23/23 23/23 +0.063 0.890 0.834 Qw en2.5-7B- Instruct 15/15 15/15 +0.035 0.832 0.788 Phi-3.5-mini 17/17 17/17 +0.087 0.698 0.600 Llama-3-8B- Base 17/17 17/17 +0.032 0.956 0.927 CP-A dditive ac hiev ed higher Spearman correlation with the empirical RDM than the Contin uous mo del at 100% of primary lay ers for all six mo dels (H1 supp orted). The pre-registered success criterion (  50% of primary lay ers for  3/5 instruct mo dels) was exceeded maximally . Figure 1 sho ws the empirical RDM for Llama-3-8B-Instruct at la yer 16, illustrating the blo ck structure at the 9/10 boundary . The CP-A dditive adv antage is not a trivial consequence of adding a free parameter: the b oundary b o ost 󰄗 was xed at 1.0 (not t to data), the eect was absen t at non-b oundary con trol p ositions (§3.5.1), absen t in the temp erature domain (§3.5.2), and the hierarc hical regression (§3.1.2) conrmed that b oundary-crossing contributes 5–27% unique v ariance b eyond log-distance — a non-trivial geometric feature, not an artefact of mo del complexity . The eect w as robust to distance metric: Euclidean distance pro duced iden tical results (CP-Additiv e > Contin uous at 100% of primary la y ers, all mo dels; Supplementary T able S1). Euclidean max 󰄜 for the CP-A dditive mo del ranged from 0.702 (Phi) to 0.955 (Base). 6 Figure 1: Empirical represen tational dissimilarity matrix (RDM) for Llama-3-8B-Instruct at la yer 16 (decade-10 condition, cosine distance). Eac h cell shows the pairwise cosine distance b et ween hidden state cen troids for tw o probing v alues. The block structure at the 9/10 b oundary is visible as an abrupt increase in cross-b oundary distances. The pre-registered mechanism test compared CP-Additiv e (constant b oundary b oost) with CP- Multiplicativ e (prop ortional scaling). CP-A dditiv e pro vided sup erior t at a majority of primary la yers in ve of six mo dels: Mistral (17/17), Gemma (22/23), Qwen (13/15), Phi (17/17), and Base (12/17). Llama-Instruct was split (8/17). This establishes that the b oundary eect op erates as a constant additiv e displacement in represen tational space rather than a prop ortional scaling of existing distances, consistent with the Fisher information warping predicted by Bonnasse-Gahot and Nadal (2022). 3.1.1 Replication at the 100-Boundary The decade-100 condition (double-digit/triple-digit b oundary at 100, probing v alues 70–130) replicated the decade-10 pattern (T able 1b). CP-Additiv e exceeded Contin uous at 100% of primary la yers for all six mo dels, with substan tially larger eect sizes. T able 1b. P aradigm A results: decade-100 b oundary and control-150. Same format as T able 1. E4 ratio = decade-100 mean CP adv antage / decade-10 mean CP adv an tage. Mo del CP > Cont (100) Mean 󰄜 (100) Max 󰄜 (CP-A dd) Max 󰄜 (Con t) Ctrl-150 󰄜 E4 Ratio Llama-3- 8B-Instruct 17/17 +0.319 0.676 0.442 +0.017 12.7  7 Mo del CP > Cont (100) Mean 󰄜 (100) Max 󰄜 (CP-A dd) Max 󰄜 (Con t) Ctrl-150 󰄜 E4 Ratio Mistral-7B- Instruct 17/17 +0.268 0.323 0.136 +0.042 3.9  Gemma-2- 9B-IT 23/23 +0.279 0.394 0.260 +0.016 4.0  Qw en2.5- 7B-Instruct 15/15 +0.162 0.404 0.241 +0.012 4.6  Phi-3.5- mini 17/17 +0.476 0.580 0.177 +0.012 4.6  Llama-3- 8B-Base 17/17 +0.305 0.619 0.338 +0.016 8.8  The 100-b oundary eect was 3.9–12.7  larger than the decade-10 eect across mo dels (E4). The matc hed control at 150 (probing v alues 130–170) sho w ed negligible CP adv an tage ( 󰄜 = +0.012 to +0.042), far smaller than b oth the decade-100 and decade-10 eects. The scaling of eect size with b oundary magnitude is interpretable: the 100-b oundary in v olves a transition from tw o-digit to three-digit n umbers, which changes b oth c haracter count and token coun t for all tested tok enisers, pro ducing a larger structural discon tinuit y than the 10-b oundary . 3.1.2 Hierarc hical Regression (H4) A t each primary lay er, pairwise represen tational distances w ere regressed on log-distance (Step 1), then a b oundary-crossing indicator was added (Step 2). The b oundary predictor added signicant unique v ariance at 100% of primary la yers for all six mo dels (all F-test p < .001). T able 2 summarises the hierarc hical regression results. T able 2. H4 hierarc hical regression: unique v ariance con tributed by b oundary-crossing b ey ond log-distance. Mo del Sig lay ers Mean  R² Max  R² Llama-3-8B-Instruct 17/17 0.050 0.084 Mistral-7B-Instruct 17/17 0.145 0.200 Gemma-2-9B-IT 23/23 0.174 0.231 Qw en2.5-7B-Instruct 15/15 0.079 0.115 Phi-3.5-mini 17/17 0.267 0.331 Llama-3-8B-Base 17/17 0.067 0.113 Boundary-crossing explains 5–27% of v ariance in representational distance b eyond what log- magnitude compression accoun ts for. The prop ortionally largest  R² appears in Phi (0.267), whic h has the weak est ov erall geometry (max 󰄜 = 0.747). The b oundary eect is prop ortionally dominan t in the smaller mo del even though its absolute geometry is attenuated. 3.2 Iden tication: Classic CP vs Structural CP (H8) Three identication framings (“small/large”, “single-digit/m ulti-digit”, “one digit/t wo digits”) were administered to each instruct model using coun terbalanced A/B forced-choice. T able 3 summarises 8 iden tication outcomes using the digit_count framing, which pro duced the best signal across mo d- els. T able 3. P aradigm B0 results: identication at the decade-10 b oundary . “Boundary at 10?” indicates whether the iden tication function crossed 0.50 within ±1 step of the structural boundary . Mo del Boundary at 10? P attern CP Type Llama-3-8B-Instruct No Gradien t, nev er crosses 0.50 Structural CP Mistral-7B-Instruct No Extreme p osition bias;  0.50 after coun terbalancing Structural CP Gemma-2-9B-IT Y es Sharp step at exactly 10 (slop e = 6.3) Classic CP Qw en2.5-7B-Instruct Y es Step at 11–12 (shifted 1–2 steps from structural b oundary) Classic CP Phi-3.5-mini No Flat at  0.50; no category signal Structural CP Llama-3-8B-Base N/A No chat template; ra w-logit iden tication uninformativ e Structural CP T w o qualitativ ely distinct patterns emerged (Figure 3). Gemma and Qw en sho wed classic CP : both geometric warping and explicit categorisation at the b oundary , with sharp sigmoid iden tication functions. Llama, Mistral, Phi, and the base mo del show ed structur al CP : geometric w arping without explicit iden tication — the category b oundary is imp osed b y represen tational format rather than explicit category knowledge. 9 Figure 2: Identication functions at the decade-10 b oundary for t w o represen tative mo dels. T op: Gemma-2-9B-IT sho ws a sharp step function crossing 0.50 at the structural b oundary under the single-digit/m ulti-digit framing (classic CP). Bottom: Llama-3-8B-Instruct shows a at identi- cation function that never crosses 0.50 under either framing (structural CP). Both mo dels show comparable CP geometry (T able 1) despite the identication disso ciation. This dissociation was stable across b oundaries: at the 100-boundary , the same mo dels that failed to iden tify the 10-b oundary also failed to iden tify the 100-b oundary , and vice versa. The disso ciation is therefore a prop ert y of the arc hitecture, not the stim ulus. Cross-mo del correlation conrmed the dissociation: identication slop e did not predict geometric CP strength (Spearman 󰄜 =  0.14, p = .79, n = 6; E9). Mo dels with the sharp est iden tication b oundaries (Gemma: slop e = 6.34) did not sho w stronger geometric CP than mo dels with no iden tication at all (Phi: slop e = 0.0, y et largest  R² = 0.331 in the hierarc hical regression). This conrms that geometric CP and explicit iden tication are disso ciated, consisten t with H8. 10 Prompt robustness analysis (E6) revealed that only Qwen pro duced consisten t crosso ver lo cations across all three framings (within 0.51 steps). Gemma and Phi show ed framing-dep enden t b oundaries (>2 steps disagreemen t), and Llama, Mistral, and the base mo del produced no crossov er under any framing. 3.3 Beha vioural Discrimination (H2) 3.3.1 Accuracy All four capable instruct mo dels (Llama, Mistral, Gemma, Qwen) achiev ed  99.9% accuracy on the forced-choice “Which is larger?” discrimination task. Phi-3.5-mini and Llama-3-8B-Base show ed chance-lev el accuracy after counterbalancing (pure p osition bias). All instruct mo dels except Phi exhibited strong rst-option (A) bias prior to coun terbalancing; Phi sho wed second-option (B) bias — an architecture-dependent eect that extends the position bias do cumen ted in P aradigm B0 (E12). Per the pre-registered exclusion rule, d  w as uninformativ e and H2 (accuracy-based), H2b (McMurra y strict test), H3 (meta-d  ), and H5 (M-ratio) were declared not ev aluable. 3.3.2 Condence (R T Analogue) Although accuracy was at ceiling, the magnitude of the decision signal — |  logit| betw een the c hosen and unc hosen option — v aried systematically with b oundary p osition (cf. Cacioli, 2026b, for the use of |  logit| as a condence measure in LLM psyc hophysics). This is an operational proxy for the pro cessing-ease signal measured by reaction time in h uman CP studies, not a b eha vioural equiv alen t; its v alidit y rests on the assumption that larger represen tational distance at the decision la y er pro duces larger logit separation, which is conrmed by the Paradigm A geometry . Cross-b oundary pairs pro duced signicantly higher condence than within-category pairs at matched log-distances (T able 4). T able 4. Paradigm B results: condence-based discrimination.  Conf = mean |  logit| for cross- b oundary minus within-category pairs. Cohen’s d and Mann-Whitney U test for the condence dierence. “Sig lev els” reports the n umber of log-distance bins (of 6) where the condence dierence reac hed signicance at p < .05. Mo del Conf(cross) Conf(within)  Conf Cohen’s d p Sig levels Llama-3- 8B-Instruct 8.39 7.77 +0.62 0.45 <.001 4/6 Mistral-7B- Instruct 10.59 10.16 +0.43 0.29 <.001 3/6 Gemma-2- 9B-IT 8.63 8.44 +0.18 0.47 <.001 4/6 Qw en2.5- 7B-Instruct 20.94 20.70 +0.23 0.23 .001 2/6 Phi-3.5- mini 0.005 0.005 +0.0003 0.09 .38 1/6 Llama-3- 8B-Base 0.075 0.061 +0.015 0.55 <.001 4/6 The condence eect was signican t for v e of six mo dels (all except Phi), with medium eect sizes (Cohen’s d = 0.23–0.55). This demonstrates that the geometric warping do cumen ted in Paradigm A propagates to b eha vioural output: the additiv e boundary bo ost in represen tational distance translates to larger evidence magnitude at the decision stage. 11 The distance-controlled pattern w as interpretable: the condence dierence w as signican t at larger log-distances (0.30–0.60) but not at the smallest distances (0.10–0.20), where pairs are close together regardless of b oundary p osition. Llama-3-8B-Base sho w ed the largest eect size (d = 0.55) despite pure p osition bias at the accuracy lev el: the magnitude of its p osition-biased logit v aried with b oundary p osition, consisten t with the P aradigm A nding that base and instruct mo dels share categorical geometric structure. Phi-3.5-mini sho w ed no condence eect (d = 0.09, p = .38), despite exhibiting geometric CP in P aradigm A (CP-A dditiv e > Contin uous at 17/17 primary lay ers). This disso ciation b et w een geom- etry and b eha vioural propagation further supports the geometry-function distinction: the additive b oundary b o ost is present in the representational space of the smaller mo del but is insucient — giv en Phi’s attenuated ov erall geometry (max 󰄜 = 0.698) and near-zero condence magnitudes (|  logit|  0.005) — to propagate to the output la y er as a measurable condence dierence. 3.4 Precision Gradien t (P aradigm C) Lo cal representational precision (1/||h(n+1)  h(n)||) show ed a boundary-sp ecic spik e at 9→10 for all six models. Precision ratios (boundary distance / mean non-b oundary distance) ranged from 1.42 (Phi) to 2.29 (Mistral) at the decade-10 b oundary , compared with ratios near 1.0 (0.92–1.07) at the matched con trol p osition 15. This lo cal expansion of representational distance at the b oundary corresp onds directly to the Fisher information peak predicted b y Bonnasse-Gahot and Nadal (2022): in their framework, optimal categorisation warps the metric tensor of the representational space such that Fisher information is maximal at category b oundaries. The precision gradien t measured here is an empirical proxy for this metric w arping — it quan ties the lo cal expansion of representational space relative to the input space at the b oundary , exactly as their theory predicts. 3.5 Con trol Conditions Both geometric and b ehavioural CP measures used in this study are analytically inv ariant to the sampling temp erature parameter: hidden states are computed deterministically from xed inputs (temp erature aects only the softmax distribution ov er next-token probabilities), and |  logit| is dened on raw logits which are unaected b y temperature scaling (E2). 3.5.1 Non-Boundary Con trol (H0) At the con trol p osition (15), the Con tinuous mo del con- sisten tly outp erformed CP-Additiv e (negative CP adv an tage for all six mo dels). The warping is sp ecic to structurally dened boundaries and absen t at arbitrary non-boundary positions within the same numerical range (H0 not falsied). At con trol p osition 150, CP adv antages w ere negligi- ble (+0.012 to +0.042; see T able 1b), far smaller than the decade-100 eects (+0.162 to +0.476) and comparable in magnitude to the temperature domain negativ es, conrming that the massiv e 100-b oundary eects are b oundary-sp ecic. 3.5.2 T emperature Domain (H6) The temperature domain (  20 to 100°C, hot/cold linguistic b oundary at ~22°C) show ed negativ e CP adv antage for all six mo dels (T able 5b). The Contin uous mo del outperformed CP-Additiv e at the hot/cold boundary , and the temp erature con trol condition (b oundary at 43°C, no linguistic distinction) show ed equally negative adv an tages. T able 5b. T emp erature domain RSA results (H6). CP adv an tage is the mean 󰄜 (CP-Additiv e  Contin uous) across primary la yers. Negative v alues indicate Contin uous ts b etter. 12 Mo del Hot/cold CP wins Hot/cold 󰄜 T emp ctrl 󰄜 Llama-3-8B-Instruct 0/17  0.062  0.062 Mistral-7B-Instruct 1/17  0.038  0.061 Gemma-2-9B-IT 1/23  0.027  0.054 Qw en2.5-7B-Instruct 4/15  0.009  0.063 Phi-3.5-mini 0/17  0.039  0.055 Llama-3-8B-Base 0/17  0.069  0.065 Despite having a linguistic category distinction, temp erature lac ks a tok enisation discon tinuit y — “21” and “23” tok enise iden tically — and therefore do es not produce represen tational w arping. This is a k ey positive null: it demonstrates that LLM categorical p erception requires structural input- format discontin uit y and is not induced by linguistic category kno wledge alone (H6 not supp orted). 3.5.3 Nonce-T ok en Remapping Control (E10) The nonce-tok en exp eriment tested whether categorical geometry can b e induced b y ordinal information alone, without the linguistic surface form of real num b ers. Sev en teen nonce tokens (“glorp”, “blick et”, “tazmo”, …) were mapp ed to ordinal p ositions 1–17 in t w o conditions: no ordering information (nonce_no_order) and an explicit pream ble establishing the ordering (nonce_ordered). T able 5 summarises E10 results. T able 5. E10 nonce-tok en remapping con trol: CP-Additiv e adv an tage across conditions and mo d- els. Mo del nonce_no_order 󰄜 nonce_ordered 󰄜 decade_10 󰄜 Llama-3-8B-Instruct  0.010 +0.012 +0.023 Mistral-7B-Instruct  0.002 +0.005 +0.060 Gemma-2-9B-IT +0.004 +0.004 +0.063 Qw en2.5-7B-Instruct  0.010 +0.017 +0.035 Phi-3.5-mini  0.023 +0.007 +0.087 Llama-3-8B-Base  0.008 +0.004 +0.032 nonce_no_order: Clean null. No signicant geometry and no CP at the arbitrary b oundary p osition. Mean CP adv an tage w as negativ e or negligible for all mo dels (  0.023 to +0.004). The con trol w orks as designed. nonce_ordered: Ordinal geometry emerged strongly , with a small but consistent CP adv an tage at the boundary (mean +0.004 to +0.017 across mo dels). Ho wev er, this CP eect was 3–10  smaller than the decade-10 eect and 10–100  smaller than the decade-100 eect. E10 establishes a three-level hierarch y (Figure 5). Without ordering information, nonce tokens pro duce no geometry and no CP . With ordering information, mo dels build ordinal representations from context alone, with a w eak CP-like eect at the boundary position. With real n um b ers, the tok enisation and digit-coun t discon tinuit y amplies the b oundary eect b y an order of magnitude. The linguistic surface form is the amplier, not the sole cause. 13 Figure 3: E10 nonce-tok en remapping con trol: three-level hierarc hy . Mean CP adv antage ( 󰄜 ) for nonce tok ens without ordering information (grey), nonce tok ens with explicit ordering (blue), and real Arabic numerals at the decade-10 b oundary (red). Real n umbers pro duce 3–10  larger CP eects than ordered nonce tokens, and nonce tokens without ordering pro duce no CP . A notable case is Phi-3.5-mini, which ac hieved near-perfect ordinal geometry in the nonce_ordered condition (max 󰄜 = 0.97 for the contin uous mo del) y et its CP adv antage w as only +0.007. This dis- so ciation b et ween geometry strength and CP magnitude demonstrates that strong ordinal geometry do es not automatically pro duce categorical w arping; the w arping requires a structural discontin uit y . 3.6 Instruction-T uning Con trol (H7) Llama-3-8B-Base sho wed nearly iden tical CP geometry to Llama-3-8B-Instruct (CP adv an tage +0.032 vs +0.023; precision ratio 1.71 vs 1.57). The base mo del achiev ed the highest max 󰄜 in the study (CP-Additiv e 󰄜 = 0.955), conrming that categorical geometric structure is presen t in pretrained representations and is not introduced b y instruction tuning (H7 supp orted). Where base and instruct models div erge is in identication and discrimination b ehaviour: the base model cannot categorise or compare num bers in forced-c hoice tasks (no chat template, pure p osition bias at the accuracy lev el), y et its represen tational geometry con tains the same categorical signature. This is a disso ciation b etw een representational structure and b eha vioural comp etence. 3.7 La y erwise Prole (E1) and Lo cal Manifold Analysis (E7, E8) CP geometry (the CP-Additiv e adv an tage ov er Contin uous) emerged at primary lay ers and was absen t at early and late lay ers for all mo dels (E1; Figure 2). The b oundary p osition induced a lo cal manifold rotation: the rst principal component of the local neigh b ourhoo d rotated 81.6–89.6° at the decade-10 boundary relative to non-b oundary p ositions (E7; range across mo dels: Mistral 81.6°, Gemma 82.3°, Qwen 86.5°, Llama-Base 88.8°, Llama-Instruct 89.1°, Phi 89.6°). Phase-reset 14 analysis conrmed that represen tational similarity tra jectories show ed a signicant discontin uity at the b oundary (Mann-Whitney p < .01, all mo dels; mean ratio of b oundary-to-non-b oundary distance: 1.51–2.37; E8). The relationship b et w een CP strength ( 󰄗 ) and global compression qualit y ( 󰄍 ) v aried across mo dels: larger models (Gemma, Mistral, Llama-Base) show ed anticorrelation at the decade-100 b oundary ( 󰄜 =  0.30 to  0.79), suggesting that CP lo cally disrupts the contin uous magnitude geometry , while Phi sho wed strong positive correlation ( 󰄜 = +0.88 to +0.99), consistent with the b oundary eect b eing the dominan t geometric feature in the smaller mo del (E11). Figure 4: Lay erwise CP-A dditive adv an tage ( 󰄜 = CP-A dditive 󰄜 minus Contin uous 󰄜 ) across all la yers for six mo dels at the decade-10 boundary . The shaded region indicates appro ximate primary la yer range. All mo dels show p ositiv e 󰄜 at primary lay ers. Llama-3-8B-Base (dashed) shows comparable CP geometry to its instruct coun terpart. 3.8 Causal In terv ention (E5) A ctiv ation patc hing tested whether the categorical geometry iden tied in P aradigm A is causally implicated in discrimination behaviour. This analysis was p erformed on Llama-3-8B-Instruct as a pro of-of-concept; full cross-architecture replication is left for future work. A ridge-regression prob e trained on RSA centroids (prob e accuracy = 1.00 at all primary lay ers; PC1–category correlation: mean 󰄜 = .83) dened the “category direction” at each la yer. Patc hing along this direction at graded dose lev els ( 󰄌  {0.25, 0.50, 0.75, 1.00}) pro duced large, specic, dose-dep enden t c hanges in discrimination condence at early lay ers but negligible eects at mid-to-late lay ers where CP geometry is strongest (T able 6; Figure 4). T able 6. E5 causal in terv ention results (Llama-3-8B-Instruct). |  conf| = absolute change in dis- crimination condence under category-direction patc hing. Sp ecicit y = ratio of category-direction eect to mean random-direction eect (10 random controls). 15 La yer Depth Cat  conf Rand 5 Early 0.621 0.009 70.1  Monotonic 8 Early-primary 0.242 0.006 43.6  Monotonic 16 Mid-primary 0.018 0.002 10.8  Monotonic 23 Near p eak RSA 0.058 0.002 34.3  Monotonic 27 Late 0.010 0.001 12.1  Non-monotonic Figure 5: Causal interv en tion via activ ation patching (E5, Llama-3-8B-Instruct). (A) Dose- resp onse: absolute change in discrimination condence (  conf  ) as a function of patching dose ( 󰄌 ) for ve lay ers. Early lay ers (5, 8) show strong, monotonic, dose-dep enden t eects; mid-to-late la yers (16, 23, 27) are at near the random direction band. (B) Sp ecicit y (category direction eect divided by random direction eect) at 󰄌   across la yers. The absolute eect of category-direction patc hing dropp ed by approximately 60  from la y er 5 (|  conf| = 0.621) to lay er 16 (|  conf| = 0.018). The category direction at early lay ers — where geometry is still forming — produced large, specic, dose-dep enden t changes in discrimination con- dence. At mid-to-late la y ers where CP geometry peaks, patching produced negligible eects despite remaining direction-sp ecic (sp ecicit y 10–34  ). This pattern replicates exactly the disso ciation found in the W eb er study (Cacioli, 2026a) for magnitude represen tations: early la yers are causally implicated, late p eak-RSA lay ers are causally inert. A sign rev ersal across la yers w as observ ed: lay er 5 patc hing increased condence (+0.621), while la yer 8 patching decreased it (  0.242). Both eects were monotonic and highly sp ecic, indicating that the category direction has opp osite functional roles at dierent depths. This suggests a programme-level principle: representational structure and functional relev ance disso ciate across depth. The lay er with the most interpretable geometry — the one a researcher w ould b e most tempted to analyse — is not the la yer performing the computational w ork. This holds for b oth magnitude (Cacioli, 2026a) and category (presen t study). 16 3.9 Hyp othesis Summary T able 7 summarises outcomes for all pre-registered hypotheses. T able 7 . Pre-registered hypothesis outcomes. Hyp othesis Outcome Key Evidence H0 (F alsication) Not falsied Con trol p ositions (15, 150) sho w no CP H1 (Representational warping) Supp orted CP-A dd > Cont at 100% of primary lay ers, all 6 mo dels H2 (Behavioural d  ) Not ev aluable Ceiling accuracy (pre-registered exclusion) H2b (McMurray strict test) Not ev aluable No clean identication b oundary in most mo dels H3 (Meta-d  ) Not ev aluable Ceiling accuracy (pre-registered exclusion) H4 (Boundary contribution) Supp orted  R² = 0.05–0.27 at 100% of la yers, all models H5 (M-ratio) Not ev aluable Ceiling accuracy (pre-registered exclusion) H6 (Cross-domain) Not supp orted T emp erature shows no CP (p ositiv e null) H7 (Instruction-tuning) Supp orted Base  Instruct for geometry; div erge for b ehaviour H8 (ID-geometry disso ciation) Supp orted Gemma/Qw en = classic CP; Llama/Mistral/Phi = structural CP F our h yp otheses w ere supp orted (H1, H4, H7, H8). One h yp othesis yielded an informativ e negativ e (H6). F our h yp otheses were not ev aluable due to pre-registered exclusion criteria (ceiling accuracy in the discrimination task). All t welv e pre-registered exploratory analyses were completed. 4. Discussion The present study applied formal psychoph ysical metho dology — representational similarit y analy- sis, signal detection theory , counterbalanced forced-choice psyc hoph ysics, and causal interv en tion — to the hidden states of six large language mo dels pro cessing Arabic n umerals. The results establish that categorical p erception, one of the most extensively studied phenomena in p erceptual psychol- ogy , has a structural analogue in articial neural netw orks. This section considers what the ndings mean for theories of categorical perception, what they rev eal ab out the relationship b et w een rep- resen tational structure and b eha vioural comp etence, and what metho dological implications they carry for the study of b oth biological and articial cognition. 4.1 Categorical P erception Without Category Kno wledge The central nding is that all six mo dels show categorical warping at digit-count b oundaries — CP-A dditive outp erforms Contin uous at 100% of primary la yers, with the eect scaling b y 4–13  17 at the larger b oundary — yet only t wo of six mo dels can explicitly iden tify the category distinction when asked. This disso ciation b et w een representational structure and explicit categorisation (H8) is the most theoretically consequential result. In the human CP literature, the relationship b et ween iden tication and discrimination has b een treated as denitional. McMurra y (2022) argued forcefully that many putative demonstrations of CP fail his strict test: observed discrimination m ust exceed what is predicted from the identication function alone. The logic of this test presupp oses that the only evidence for categorical structure is b eha vioural — because in human studies, the represen tational geometry is not directly observ able. LLMs extend this framework. Direct access to the geometry reveals that categorical warping is presen t in models that cannot rep ort the category . The four “structural CP” mo dels (Llama, Mistral, Phi, Base) exhibit the full geometric signature — additive boundary b o ost, precision gradien t spik e, manifold rotation of 82–90° at the b oundary , phase-reset discon tinuit y — without an y abilit y to p erform the identication task. This is not a failure of the CP paradigm; it is a disso ciation that McMurray’s framework, designed for systems where representations are opaque, w as not p ositioned to detect. The implication for cognitive science is that the standard tw o-task criterion (identication + dis- crimination) is sucien t but not necessary for establishing CP . When represen tational geometry is directly accessible, geometric w arping at the boundary is itself evidence for categorical structure, re- gardless of whether the system can report the category . This aligns with the theoretical p ositions of Bonnasse-Gahot and Nadal (2022), who show ed that Fisher information w arping at category b ound- aries emerges in deep lay ers of trained neural net works regardless of output behaviour, and Park et al. (2024), who demonstrated categorical p olytope structure without requiring identication tasks. The present study pro vides the rst empirical demonstration that this representational-geometric form of CP app ears alongside — and disso ciates from — the classical b eha vioural form within the same set of systems. 4.2 Structural vs A cquired Categorical P erception The distinction b et w een structural and acquired CP — b etw een warping imposed b y prop erties of the representational format and warping that emerges from category learning — has b een dicult to adjudicate in biological systems because the t wo sources of w arping are t ypically confounded. Hu- mans who learn colour categories also ha ve retinal cone tuning curves; humans who learn phoneme categories also hav e co c hlear lter banks. Input structure and learned categorisation cov ary . LLMs oer a cleaner separation. Three con verging results establish that the CP do cumen ted here is primarily structural rather than acquired. First, the temp erature domain control (H6). T emp erature has a linguistic category b oundary (hot/cold) but no tokenisation discontin uit y — “21°C” and “23°C” tokenise identically . If CP w ere driv en primarily b y learned semantic categories, temp erature should show at least some warping. It do es not. CP adv an tage w as negative for all six mo dels in the temperature domain (  0.009 to  0.069), compared with p ositiv e adv an tages of +0.023 to +0.087 for n umbers. This is consistent with the structural account, though it should be noted that “hot/cold” is a relativ ely fuzzy , con text- dep enden t b oundary — a sharp er linguistic category (e.g., grammatical num b er) w ould provide a stronger test. Second, the nonce-tok en remapping control (E10). Nonce tokens mapp ed to ordinal p ositions with an explicit ordering preamble pro duced a w eak CP eect ( 󰄜 = +0.004 to +0.017), but this was 18 3–10  smaller than the decade-10 eect and 10–100  smaller than the decade-100 eect. Ordinal con text alone can induce a trace of b oundary warping — likely reecting the mo del’s ability to learn w eak categorical structure from in-context ordinal information — but the tokenisation and digit-coun t discontin uit y amplies this eect b y an order of magnitude. The Phi disso ciation is particularly informative: near-p erfect ordinal geometry ( 󰄜 = 0.97) with negligible CP adv antage (+0.007), demonstrating that ordered represen tations do not automatically produce categorical w arping. Strong ordinal geometry is not sucient; the structural discon tinuit y in the input format is required to pro duce the large eects do cumen ted for real n umbers. Third, the instruction-tuning control (H7). Llama-3-8B-Base and Llama-3-8B-Instruct show nearly iden tical CP geometry ( 󰄜 = +0.032 vs +0.023; the base mo del actually achiev es the highest max 󰄜 in the study). The categorical structure is presen t in pretrained represen tations b efore an y instruction tuning — it is a prop ert y of ho w the model enco des the distributional statistics of digit strings, not a pro duct of explicit category instruction. T ogether, these three con trols conv erge on the same conclusion: structural prop erties of the input — tokenisation b oundaries, digit-coun t transitions, character-length changes — are sucient to pro duce categorical p erception geometry in LLMs, and the eect do es not require learned seman tic category knowledge. This do es not mean that all CP is structural, nor that semantic categories could never pro duce CP under dieren t conditions. The classic/structural CP disso ciation (H8) suggests that instruction tuning can add an explicit categorisation capacit y on top of the structural geometry , pro ducing the full CP signature in some arc hitectures (Gemma, Qw en) but not others. The structural warping, ho wev er, is universal. This result has implications b ey ond LLMs. It provides a pro of of concept that categorical p erception can arise from representational format constrain ts without category learning — a p ossibilit y that has b een theorised (McMurra y , 2022; Massaro, 1987) but never demonstrated in a system where the geometry , the training data, and the input format are all sim ultaneously observ able. 4.3 The Geometry-F unction Dissociation The causal interv en tion (E5) rev ealed a disso ciation b et w een represen tational structure and func- tional relev ance that replicates across both the magnitude (Cacioli, 2026a) and category domains. A t early lay ers (5 and 8), patching along the category direction pro duced large, sp ecic, dose- dep enden t changes in discrimination condence (|  conf| up to 0.621, sp ecicity 44–70  ab ov e random con trols). At mid-to-late lay ers where CP geometry p eaks (lay ers 16–27), patching pro- duced negligible eects (|  conf| = 0.01–0.06) despite remaining direction-sp ecic. This pattern is problematic for the common in terpretability practice of identifying the lay er with the strongest prob e accuracy or the highest RSA t and declaring it the “represen tation la yer” for a giv en concept. The la yer with the most in terpretable geometry is not the la y er p erforming the computation. This is not a n ull result — the geometry is real, and it is causally grounded at early la yers — but it means that the computational w ork happens where the geometry is still forming, not where it has crystallised into its most legible form. The sign rev ersal at la y ers 5 and 8 (positive and negativ e eects on condence, resp ectiv ely) further suggests that the category direction plays functionally dieren t roles at dieren t depths. This is consisten t with recent w ork on sup erposition and polysemanticit y in neural net w ork represen tations (Elhage et al., 2022): the same geometric feature may supp ort dieren t computations at dierent la yers, and a single “category direction” extracted b y a linear probe ma y conate functionally 19 distinct subspaces. F or the CP literature, this nding complicates the use of representational similarity analysis as a standalone measure of categorical structure. RSA can identify where CP geometry is present, but not where it matters. Causal metho ds are required to establish functional relev ance — a lesson that applies equally to neuroimaging studies of human categorical perception, where the distinction b et w een represen tational presence and causal relev ance is often elided. 4.4 A dditiv e vs Multiplicative W arping The pre-registered mechanism test established that the b oundary eect is additive rather than m ulti- plicativ e: a constan t displacemen t in representational distance at the b oundary , not a prop ortional scaling of existing distances. Five of six mo dels fav oured the additiv e mo del; only Llama-Instruct w as split. This nding connects to the theoretical framework of Bonnasse-Gahot and Nadal (2022), who deriv ed that optimal categorisation pro duces a lo cal spike in Fisher information at the category b oundary — a phenomenon that manifests geometrically as a lo cal additiv e increase in represen- tational distance. The additive nature of the eect is imp ortan t b ecause it means the b oundary op erates as an indep enden t structural feature lay ered on top of the contin uous magnitude geom- etry , rather than mo dulating that geometry . This is consistent with the hierarc hical regression results (H4), where b oundary-crossing explained 5–27% of unique v ariance b ey ond log-distance: the b oundary adds information to the geometry rather than distorting the information already presen t. 4.5 Implications for Theories of Categorical P erception The present study sits at the intersection of classical psychoph ysics, computational neuroscience, and AI interpretabilit y . Three sets of implications follow. F r om c o gnitive scienc e to AI: The psychoph ysical to olkit — iden tication functions, forced-choice discrimination with SDT, represen tational similarity analysis, precision gradients — pro vides a rigorous framew ork for c haracterising how neural netw orks represent categorical structure. This framew ork go es b eyond the binary “do es the mo del know X?” question typical of NLP ev alua- tion and instead asks how knowledge is geometrically enco ded, how it v aries across lay ers, and whether represen tational structure is causally linked to b ehaviour. The nonce-token control (E10) illustrates the v alue of psychoph ysical design: b y manipulating the relationship b et ween ordinal information and linguistic surface form, it isolates the specic con tribution of tokenisation structure to categorical geometry — a precision that standard NLP probing metho ds do not ac hieve. F r om AI to c o gnitive scienc e: LLMs pro vide a system where the acquired-vs-structural distinction can b e cleanly tested b ecause the representational format, the training data, and the resulting geometry are all observ able. The nding that structural input properties (tokenisation, digit coun t) pro duce categorical warping without explicit category learning oers an existence pro of for the structural accoun t of CP that has b een debated for decades in the human literature (Massaro, 1987; McMurray , 2022). It do es not resolv e the debate for biological systems, but it establishes that the structural mec hanism is computationally viable — that format-driven warping can pro duce the full geometric CP signature in a learning system. The present w ork complements a growing b ody of research applying cognitive science constructs to LLM represen tations. Park et al. (2024, 2025) demonstrated that LLMs represent categori- 20 cal concepts as polytop e structures, establishing that categorical geometry exists in hidden states but without testing the psyc hophysical signature (iden tication, discrimination, b oundary-sp ecic w arping) that denes CP . Shani, Marjieh, and Griths (2025) sho w ed that LLM represen tations compress along typicalit y gradients in a manner consisten t with human categorisation, but fo cused on within-category structure rather than the b et w een-category b oundary eects that are diagnostic of CP . The W eber study (Cacioli, 2026a) established that numerical magnitude follows logarithmic compression consisten t with W eb er’s Law, pro viding the contin uous baseline geometry on which the presen t categorical eects are superimp osed. The presen t study integrates these strands: it tests CP sp ecically — the b oundary phenomenon — using the formal psychoph ysical methodology (RSA with theoretical mo del comparison, SDT-based discrimination, identication functions, causal in- terv ention) that the prior w ork did not employ . The result is that LLM categorical structure is not merely present as a geometric fact but op erates through a sp ecic mechanism (additive b ound- ary w arping driven b y input-format discon tinuit y) that can b e dissociated from explicit category kno wledge, quan tied at each lay er, and causally tested. Bidir e ctional: The geometry-function disso ciation (E5) iden ties a metho dological hazard shared b y cognitiv e science and AI in terpretabilit y researc h: the lay er (or brain region, or time windo w) with the most legible represen tational structure ma y not b e the la yer p erforming the relev an t computation. This is not a nov el insigh t in neuroscience — the distinction b etw een neural coding and neural computation has b een drawn b efore (de Wit et al., 2016) — but the present study pro vides an unusually clean demonstration b ecause the entire computational pip eline is observ able and causally manipulable. 4.6 Limitations Sev eral limitations should b e noted. First, the causal in terv en tion w as performed on one model only (Llama-3-8B-Instruct) as a proof-of-concept; although the geometry-function dissociation replicates the pattern found in the W eber study for magnitude represen tations, the generality of the causal nding across architectures remains to b e tested. Second, the n umerical domain is a particularly clean case for structural CP because digit-coun t boundaries pro duce a sim ultaneous change in tok enisation, character count, and lexical form. Whether structural CP extends to domains with subtler input-format discontin uities (e.g., morphological b oundaries in inected languages) is an op en question. Third, the pre-registered discrimination h yp otheses (H2, H3, H5) w ere not ev aluable due to ceiling accuracy — Arabic numeral comparison is trivially easy for capable LLMs. A harder discrimination task (e.g., cross-format comparison: “Which is larger, nine or 12?”) would likely pro duce non-ceiling accuracy and enable the full SDT analysis. F ourth, the control-150 condition sho wed small but non-zero CP adv an tages ( 󰄜 = +0.012 to +0.042), suggesting p ossible leakage of the 100-b oundary eect into the con trol range. These v alues w ere far smaller than the decade- 100 eects (+0.162 to +0.476) and comparable in magnitude to statistical noise in the temp erature domain, but the con trol is not as clean a null as the control-15 condition for the decade-10 b oundary . Fifth, the temp erature domain tested only one non-structural category b oundary , and “hot/cold” is a relatively fuzzy , gradient b oundary; additional domains with sharp er linguistic categories (e.g., grammatical n umber, animacy) w ould strengthen the cross-domain generalisation of the structural accoun t. Sixth, the study tested mo dels in the 3.8–9.2B parameter range; whether the same patterns hold at substan tially larger scales (70B+) is unknown. Seven th, while the present study infers structural input-format discontin uities as the primary mechanism, the Supplementary T able S2 analysis reveals that the relationship b et w een tokenisation and CP is not straightforw ard: Llama- 3’s BPE vocabulary merges multi-digit num b ers in to single tok ens, eliminating the token-coun t 21 discon tinuit y that is presen t in Gemma, Qwen, Mistral, and Phi — yet Llama sho ws comparable CP geometry . This suggests that character-coun t and lexical-form changes (which are common to all mo dels) are sucient, and that tok en-count c hanges amplify but are not necessary for the eect. A same-domain, dierent-tok enisation control (e.g., re-tokenising identical stim uli with alternativ e BPE vocabularies) would pro vide a stronger test and is an imp ortan t direction for future work. 4.7 Conclusion Structural input-format discon tinuities — tok enisation b oundaries, digit-coun t transitions, and c haracter-length c hanges — are sucient to pro duce the full geometric signature of categorical p erception in LLM hidden states, indep enden tly of explicit semantic category kno wledge. The geometric warping at category b oundaries is univ ersal across arc hitectures, present in pretrained represen tations b efore instruction tuning, absent in domains that lack structural input discontin u- ities, and causally grounded at early la yers where represen tations are still forming rather than at later lay ers where the geometry reaches its p eak legibility . The dissociation b et w een geometric CP (universal) and explicit iden tication (arc hitecture- dep enden t) demonstrates that representational structure can outrun b eha vioural comp etence. F or AI interpretabilit y , this cautions against equating the most legible representation with the most functionally relev an t one. F or theories of categorical p erception more broadly , it provides a computational existence pro of that CP geometry can arise from the format of the input represen tation, not only from category learning, in a system where b oth the format and the learning are fully observ able. A c kno wledgments This researc h was conducted as part of the Classical Minds, Mo dern Machines indep enden t researc h programme. No external funding was receiv ed. Use of Generativ e AI: Claude (Anthropic) w as used as a researc h assistan t for co de generation and gure pro duction. The author takes full responsibility for the accuracy of all conten t. LLMs are the ob jects of ev aluation in this study , not components of the experimental metho d. Data A v ailabilit y All co de, stim uli, pre-registration, and analysis scripts are av ailable at h ttps://anonymous.4open.science/r/weber- B02C (directory m3_pilot/ ). Raw hidden-state extractions are a v ailable up on request. Pre- registration: https://osf.io/qrxf3/o v erview?view_only=4afe7fb d28764087a65a4222578ad625. References Cacioli (2026a). W eb er’s Law in large language mo del hidden states. Under review. Cacioli (2026b). Do LLMs kno w what they kno w? Signal detection theory meets metacognition. Under review. Bonnasse-Gahot, L., & Nadal, J.-P . (2022). Categorical p erception: A groundw ork for deep learning. Neur al Computation , 34 (2), 437–475. 22 Burns, E. M., & W ard, W. D. (1978). Categorical p erception — phenomenon or epiphenomenon: Evidence from exp eriments in the perception of melo dic m usical in terv als. Journal of the A c oustic al So ciety of A meric a , 63 (2), 456–468. de Wit, L., Alexander, D., Ekroll, V., & W agemans, J. (2016). Is neuroimaging measuring infor- mation in the brain? Psychonomic Bul letin & R eview , 23 (5), 1415–1428. Elhage, N., Nanda, N., Olsson, C., Henighan, T., Joseph, N., Mann, B., Ask ell, A., Bai, Y., Chen, A., Conerly , T., DasSarma, N., Drain, D., Ganguli, D., Hateld-Do dds, Z., Hernandez, D., Jones, A., Kernion, J., Lovitt, L., Ndousse, K., … Olah, C. (2022). T o y mo dels of sup erp osition. arXiv:2209.10652 . Etco, N. L., & Magee, J. J. (1992). Categorical p erception of facial expressions. Co gnition , 44 (3), 227–240. Goldstone, R. L. (1994). Inuences of categorization on p erceptual discrimination. Journal of Exp erimental Psycholo gy: Gener al , 123 (2), 178–200. Goldstone, R. L., & Hendrickson, A. T. (2010). Categorical p erception. Wiley Inter disciplinary R eviews: Co gnitive Scienc e , 1 (1), 69–78. Harnad, S. (Ed.). (1987). Cate goric al p er c eption: The gr oundwork of c o gnition . Cambridge Uni- v ersity Press. Lib erman, A. M., Harris, K. S., Homan, H. S., & Grith, B. C. (1957). The discrimination of sp eec h sounds within and across phoneme b oundaries. Journal of Exp erimental Psycholo gy , 54 (5), 358–368. Massaro, D. W. (1987). Categorical partition: A fuzzy-logical mo del of categorization b ehavior. In S. Harnad (Ed.), Cate goric al p er c eption: The gr oundwork of c o gnition (pp. 254–283). Cambridge Univ ersity Press. McMurra y , B. (2022). The myth of categorical p erception. Journal of the A c oustic al So ciety of A meric a , 152 (6), 3819–3842. P ark, K., Y un, S., Lee, J., & Shin, J. (2024). The geometry of categorical and hierarchical concepts in large language mo dels. . P ark, K., Y un, S., Lee, J., & Shin, J. (2025). The linear representation of categorical and hierarchi- cal concepts in LLMs. In Pr o c e e dings of the International Confer enc e on L e arning R epr esentations (ICLR 2025) (Oral). Rogers, J. C., & Da vis, M. H. (2009). Categorical p erception of sp eec h without stim ulus repetition. In Pr o c e e dings of Intersp e e ch 2009 (pp. 376–379). Shani, C., Marjieh, R., & Griths, T. L. (2025). Compression in LLMs mirrors human typicalit y gradien ts. . Wina wer, J., Witthoft, N., F rank, M. C., W u, L., W ade, A. R., & Boro ditsky , L. (2007). Russian blues reveal eects of language on color discrimination. Pr o c e e dings of the National A c ademy of Scienc es , 104 (19), 7780–7785. Zh u, F., Dai, D., & Sui, Z. (2025). Language mo dels enco de the v alue of num bers linearly . In Pr o c e e dings of the International Confer enc e on Computational Linguistics (COLING 2025) (pp. 693– 709). 23 Supplemen tary Materials Supplemen tary T able S2: T ok eniser Enco ding of Boundary-Relev ant Numbers T ok en coun ts for num bers spanning the decade-10 (9->10) and decade-100 (99->100) boundaries. Three tokenisation patterns emerge: (i) Llama-3 (instruct and base) uses merged BPE tokens for m ulti-digit num b ers, pro ducing no token-coun t discontin uit y at either b oundary; (ii) Gemma and Qw en use p er-digit tok enisation with single-digit n umbers as one tok en, yielding a 1->2 tok en jump at the decade-10 boundary and 2->3 at decade-100; (iii) Mistral and Phi prep end a leading-space tok en and tokenise p er-digit, yielding 2->3 and 3->4 token jumps resp ectively . Decade-10 b oundary: Num b er Digits Llama-3 Mistral Gemma Qwen Phi 9 1 1 tok 2 tok 1 tok 1 tok 2 tok 10 2 1 tok 3 tok 2 tok 2 tok 3 tok  tok ens 0 +1 +1 +1 +1 Decade-100 b oundary: Num b er Digits Llama-3 Mistral Gemma Qwen Phi 99 2 1 tok 3 tok 2 tok 2 tok 3 tok 100 3 1 tok 4 tok 3 tok 3 tok 4 tok  tok ens 0 +1 +1 +1 +1 F ull t ok en IDs (decade-10 range): Num b er Llama-3 (In- struct/Base) Gemma-2-9B Qw en2.5-7B Mistral-7B Phi-3.5-mini 4 [19] (1 tok) [235310] (1 tok) [19] (1 tok) [_, 4] (2 tok) [_, 4] (2 tok) 5 [20] (1 tok) [235308] (1 tok) [20] (1 tok) [_, 5] (2 tok) [_, 5] (2 tok) 6 [21] (1 tok) [235318] (1 tok) [21] (1 tok) [_, 6] (2 tok) [_, 6] (2 tok) 7 [22] (1 tok) [235324] (1 tok) [22] (1 tok) [_, 7] (2 tok) [_, 7] (2 tok) 8 [23] (1 tok) [235321] (1 tok) [23] (1 tok) [_, 8] (2 tok) [_, 8] (2 tok) 9 [24] (1 tok) [235315] (1 tok) [24] (1 tok) [_, 9] (2 tok) [_, 9] (2 tok) 24 Num b er Llama-3 (In- struct/Base) Gemma-2-9B Qw en2.5-7B Mistral-7B Phi-3.5-mini 10 [605] (1 tok) [1, 0] (2 tok) [1, 0] (2 tok) [_, 1, 0] (3 tok) [_, 1, 0] (3 tok) 11 [806] (1 tok) [1, 1] (2 tok) [1, 1] (2 tok) [_, 1, 1] (3 tok) [_, 1, 1] (3 tok) 12 [717] (1 tok) [1, 2] (2 tok) [1, 2] (2 tok) [_, 1, 2] (3 tok) [_, 1, 2] (3 tok) 15 [868] (1 tok) [1, 5] (2 tok) [1, 5] (2 tok) [_, 1, 5] (3 tok) [_, 1, 5] (3 tok) 20 [508] (1 tok) [2, 0] (2 tok) [2, 0] (2 tok) [_, 2, 0] (3 tok) [_, 2, 0] (3 tok) Note: Llama-3 (b oth Instruct and Base) shares the same tok eniser; tok en IDs are iden tical. The _ symbol denotes a leading-space token prep ended b y Mistral and Phi tokenisers. Despite the absence of a token-coun t discon tinuit y in Llama-3, all mo dels show comparable CP geometry at the decade-10 b oundary (T able 1), indicating that c haracter-count and lexical-form changes are sucien t to pro duce the eect. Dr aft gener ate d [date]. Corr esp ondenc e: Jon-Paul Cacioli, synthium@hotmail.c om. 25

Original Paper

Loading high-quality paper...

Comments & Academic Discussion

Loading comments...

Leave a Comment