Under review at ICML 2026 Workshop on Mechanistic Interpretability, 2026
We argue for a disciplined import of network-neuroscience tools rather than a loose brain analogy, specify the transformer graph contract, and state eight testable translations with failure criteria.
Across three open instruction models, holding the intervention fixed and varying the demonstration environment moves the self-report between target and source mechanisms. Self-report benchmarks should require environment-shift invariance under fixed intervention.
Under review at ICML 2026 Workshop on Mechanistic Interpretability, 2026
Moral framing is linearly decodable in pretrained Gemma-3-4B but has no causal effect on its judgment; in the instruction-tuned checkpoint that same representation becomes causally usable. Within-model framing-judgment alignment is 8.4x larger in IT than in the matched pretrained checkpoint.
An evaluation protocol, documentation object, executable audit package, and claim-specific evidential standard for safe fine-tuning defenses. In a 46-cell audit on Gemma-2-2B-it, no cell satisfies the strict conjunction.
We localize an eval-vs-deploy routing signal to a narrow mid-depth attention window and a low-dimensional subspace installed by fine-tuning. Clamping the subspace at inference reduces the gap in 11 of 12 architecture-behavior cells.
A matched-action evaluation protocol that formalises grounded auditing as an interface policy I = (O, R, A, G). Evidence access improves correction; citation gates reduce over-trust mainly by inducing abstention.
AI deployment in sensitive domains is often treated as unsafe to authorize until model internals can be explained. We argue the gate should be calibrated verification, and propose Verification Coverage, a six-component reportable standard.
A reference architecture for agentic hybrid retrieval combining BM25 lexical search with dense-embedding retrieval via reciprocal rank fusion, orchestrated by an LLM controller.
CAKE is 188 expert-validated questions spanning four cognitive levels and five cloud-native topics, evaluated across 22 model configurations from four families.
We survey agentic coding tools and identify five mechanisms by which they make implicit architectural choices, then analyze prompt-architecture coupling. Six recurring patterns arise. We call this vibe architecting.
29th International Conference on Knowledge-Based and Intelligent Information & Engineering Systems (KES 2025), 2025
A three-class segmentation framework distinguishing sea, inland water, and land. DeepLabv3+ and a hybrid ResNet-UNet outperformed the other eight models evaluated.
A systematic mapping of algorithmic trends to GI imaging techniques, with quantitative analysis of dataset-size to performance and translational enablers.
Biomedical Signal Processing and Control, 2025 (Under Revision), 2025
A carefully controlled experiment on segmenting the layers of the artery wall from only nine annotated histology images. Standard CNNs pretrained on a large histology corpus vs a vision foundation model under a systematic prompting curriculum.
Under review at Computers and Electronics in Agriculture, 2025
We benchmark 19 traditional ML algorithms on dual-task prediction using the DeepHS Fruit dataset across five species. ExtraTrees with stratified resplit achieves 75.00% overall accuracy, surpassing Fruit-HSNet.
FYI is a browser extension that bridges automated and manual fact-checking through four complementary tools. In an N=22 think-aloud study, participants adopted three workflow archetypes.