• ↑↓ pour naviguer
  • pour ouvrir
  • pour sélectionner
  • ⌘ ⌥ ↵ pour ouvrir dans un panneau
  • esc pour rejeter
⌘ '
raccourcis clavier

During my undergraduate year, I watched my roommate Noah struggle with assigned readings. He wasn’t lacking in intelligence or curiosity - quite the opposite. But the dense walls of academic text would cause his attention to scatter like marbles on a hardwood floor. By the time he reached the bottom of a page, the beginning had already evaporated from his mind.

Noah’s experience isn’t unique. Attention-Deficit/Hyperactivity-Disorder (ADHD) affects approximately 5–8% of children globally (Polanczyk et al., 2015; Wolraich et al., 2019), with symptoms persisting into adulthood for 60% of cases (Faraone et al., 2006; Sibley et al., 2017). This neurodevelopmental condition, characterized by inattention, hyperactivity, and impulsivity, creates significant challenges in tasks requiring sustained focus—particularly reading. For individuals with ADHD, dense text environments often lead to cognitive overload, reduced comprehension, and frustration, exacerbating educational disparities, workplace inequities, and mental health struggles (Chang et al., 2014; Dalsgaard et al., 2014; Skirrow & Asherson, 2013). Traditional interventions, such as medication and behavioral therapy, are effective but fail to address real-time cognitive engagement during reading. Additionally, accessibility tools for reading focus, while mainly on surface-level modifications - font sizes, color schemes, text-to-speech conversion, lack the intelligence to respond to fluctuating attention or scaffold comprehension in real time.

To bridge this gap, our project reimagines Kay (2011) ‘s Dynabook, a “metamedium” for creative thought, as a generative system that interacts with ADHD users: when attention wanes, it extrapolates key concepts from the text and generates interfaces to re-engage the reader. This approach aligns with ADHD’s neurocognitive profile, where novelty and interactivity enhance dopamine-driven focus.

why machine learning?

One salient property of modern large language models system is their emergent ability to comprehend knowledge: they can process and understand text at multiple levels of abstraction simultaneously, from an unprecedented amount of data.

This capability isn’t just impressive - it’s remarkably similar to how we process information. Elhage et al. (2022) posits that these models store way more features in linear representation than it has dimensions, a phenomenon called superposition. Just as a human might understand “apple” simultaneously as a fruit, a tech company, and a symbol of knowledge, LLMs develop rich, interconnected representations of concepts. Our core hypothesis is that we can leverage this emergent property in a novel way. Instead of using these models as black boxes for generating text (which is how most current applications work, ChatGPT, Copilot), we can use sparse autoencoders (SAEs) to “peek inside” and extract these higher-dimensional concept representations. Think of it like creating a map of how ideas connect and relate, but in many more dimensions than traditional concept mapping. SAEs address this by acting as “concept sieves.” Trained on LLM activations, they apply sparsity constraints to isolate distinct features from the model’s superpositional latent space.

This approach is particularly promising for ADHD readers because: ADHD minds often excel at seeing unexpected connections and patterns. These higher-dimensional concept maps could provide multiple “entry points” into complex material. We can then generate dynamic, personalized scaffolding that matches individual thinking patterns.

For example, when processing a dense academic text, our system could:

  • Extract the hierarchical concept structure
  • Identify parallel ideas and metaphors
  • Generate multiple complementary representations of key ideas
  • Create dynamic pathways through the material based on individual engagement patterns

This is fundamentally different from traditional approaches because we’re not just transforming the presentation - we’re leveraging the same kind of deep pattern recognition that makes modern AI systems so powerful to support human pattern recognition where it might struggle.

dataset.

To achieve this, we plan to replicate Anthropic April Update on LMSys Chat 1M (Zheng et al., 2024), with synthetic dataset from distilabel from high-quality tokens from DeepSeek R1 (DeepSeek-AI et al., 2025)