Kristian Kuznetsov
2026
AudioSAE: Towards Understanding of Audio-Processing Models with Sparse AutoEncoders
Georgii Aparin | Tasnima Sadekova | Alexey Rukhovich | Assel Yermekova | Laida Kushnareva | Vadim Popov | Kristian Kuznetsov | Irina Piontkovskaya
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Georgii Aparin | Tasnima Sadekova | Alexey Rukhovich | Assel Yermekova | Laida Kushnareva | Vadim Popov | Kristian Kuznetsov | Irina Piontkovskaya
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Sparse Autoencoders (SAEs) are powerful tools for interpreting neural representations, yet their use in audio remains underexplored. We train SAEs across all encoder layers of Whisper and HuBERT, provide an extensive evaluation of their stability, interpretability, and show their practical utility. Over 50% of the features remain consistent across random seeds, and reconstruction quality is preserved. SAE features capture general acoustic and semantic information as well as specific events, including environmental noises and paralinguistic sounds (e.g. laughter, whispering) and disentangle them effectively, requiring removal of only 19-27% of features to erase a concept. Feature steering reduces Whisper’s false speech detections by 70% with negligible WER increase, demonstrating real-world applicability. Finally, we find SAE features correlated with human EEG activity during speech perception, indicating alignment with human neural processing. The code and checkpoints are available at https://github.com/audiosae/audiosae_demo.
Unveiling Intrinsic Dimension of Texts: from Academic Abstract to Creative Story
Pedashenko Vladislav | Laida Kushnareva | Yana Khassan Nibal | Eduard Tulchinskii | Kristian Kuznetsov | Vladislav Zharchinskii | Yury Maximov | Irina Piontkovskaya
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Pedashenko Vladislav | Laida Kushnareva | Yana Khassan Nibal | Eduard Tulchinskii | Kristian Kuznetsov | Vladislav Zharchinskii | Yury Maximov | Irina Piontkovskaya
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Intrinsic dimension (ID) is an important tool in modern LLM analysis, informing studies of training dynamics, scaling behavior, and dataset structure, yet its textual determinants remain underexplored. We provide the first comprehensive study grounding ID in interpretable text properties through cross-encoder analysis, linguistic features, and sparse autoencoders (SAEs). In this work, we establish three key findings. First, ID is complementary to entropy-based metrics: after controlling for length, the two are uncorrelated, with ID capturing geometric complexity orthogonal to prediction quality. Second, ID exhibits robust genre stratification: scientific prose shows low ID (∼ 8), encyclopedic content medium ID (∼ 9), and creative/opinion writing high ID (∼ 10.5) across all models tested. This reveals that contemporary LLMs find scientific text "representationally simple" while fiction requires additional degrees of freedom. Third, using SAEs, we identify causal features: scientific signals (formal tone, report templates, statistics) reduce ID; humanized signals (personalization, emotion, narrative) increase it. Steering experiments confirm these effects are causal. Thus, for contemporary models, scientific writing appears comparatively "easy", whereas fiction, opinion, and affect add representational degrees of freedom. Our multi-faceted analysis provides practical guidance for the proper use of ID and the sound interpretation of ID-based results.
2025
Feature-Level Insights into Artificial Text Detection with Sparse Autoencoders
Kristian Kuznetsov | Laida Kushnareva | Anton Razzhigaev | Polina Druzhinina | Anastasia Voznyuk | Irina Piontkovskaya | Evgeny Burnaev | Serguei Barannikov
Findings of the Association for Computational Linguistics: ACL 2025
Kristian Kuznetsov | Laida Kushnareva | Anton Razzhigaev | Polina Druzhinina | Anastasia Voznyuk | Irina Piontkovskaya | Evgeny Burnaev | Serguei Barannikov
Findings of the Association for Computational Linguistics: ACL 2025
Artificial Text Detection (ATD) is becoming increasingly important with the rise of advanced Large Language Models (LLMs). Despite numerous efforts, no single algorithm performs consistently well across different types of unseen text or guarantees effective generalization to new LLMs. Interpretability plays a crucial role in achieving this goal. In this study, we enhance ATD interpretability by using Sparse Autoencoders (SAE) to extract features from Gemma-2-2B’s residual stream. We identify both interpretable and efficient features, analyzing their semantics and relevance through domain- and model-specific statistics, a steering approach, and manual or LLM-based interpretation of obtained features. Our methods offer valuable insights into how texts from various models differ from human-written content. We show that modern LLMs have a distinct writing style, especially in information-dense domains, even though they can produce human-like outputs with personalized prompts. The code for this paper is available at https://github.com/pyashy/SAE_ATD.
2024
Robust AI-Generated Text Detection by Restricted Embeddings
Kristian Kuznetsov | Eduard Tulchinskii | Laida Kushnareva | German Magai | Serguei Barannikov | Sergey Nikolenko | Irina Piontkovskaya
Findings of the Association for Computational Linguistics: EMNLP 2024
Kristian Kuznetsov | Eduard Tulchinskii | Laida Kushnareva | German Magai | Serguei Barannikov | Sergey Nikolenko | Irina Piontkovskaya
Findings of the Association for Computational Linguistics: EMNLP 2024
Growing amount and quality of AI-generated texts makes detecting such content more difficult. In most real-world scenarios, the domain (style and topic) of generated data and the generator model are not known in advance. In this work, we focus on the robustness of classifier-based detectors of AI-generated text, namely their ability to transfer to unseen generators or semantic domains. We investigate the geometry of the embedding space of Transformer-based text encoders and show that clearing out harmful linear subspaces helps to train a robust classifier, ignoring domain-specific spurious features. We investigate several subspace decomposition and feature selection strategies and achieve significant improvements over state of the art methods in cross-domain and cross-generator transfer. Our best approaches for head-wise and coordinate-based subspace removal increase the mean out-of-distribution (OOD) classification score by up to 9% and 14% in particular setups for RoBERTa and BERT embeddings respectively. We release our code and data: https://github.com/SilverSolver/RobustATD