{
  "@context": "https://schema.org",
  "@type": "Book",
  "@id": "http://bigfivekiller.online/books/9781608885350/",
  "isbn": "9781608885350",
  "name": "Story Operators: RKHS Applications to Fiction and Poetry",
  "alternateName": "Story Operators: A Mathematical Framework for Narrative Discovery and Transformation",
  "author": {
    "@type": "Person",
    "name": "Fred Zimmerman",
    "affiliation": {
      "@type": "Organization",
      "name": "Nimble Books LLC"
    }
  },
  "publisher": {
    "@type": "Organization",
    "name": "Nimble Books LLC",
    "brand": "Nimble AI",
    "url": "http://bigfivekiller.online"
  },
  "inLanguage": "en",
  "bookFormat": "Paperback",
  "numberOfPages": 320,
  "datePublished": "2026-02-10",
  "dateModified": "2026-04-17",
  "url": "http://bigfivekiller.online/books/9781608885350/",
  "sameAs": [
    "http://bigfivekiller.online/books/9781608885350/9781608885350.pdf"
  ],
  "description": "Original research applying Reproducing Kernel Hilbert Space (RKHS) theory to narrative analysis. Introduces Story Operators: a mathematical framework for measuring, projecting, and transforming stories within a structured inner-product space. Uses 768-dimensional embeddings paired with cosine, Chebyshev, and custom composite kernels to quantify similarity, genre membership, and stylistic distance across fiction and poetry corpora. Includes worked examples, kernel design recipes, and a chapter on the historical precedent set when Fisher, Mendeleev, and Chomsky each introduced formal mathematical frameworks to their fields.",
  "about": [
    "Reproducing Kernel Hilbert Spaces",
    "Kernel methods in natural language processing",
    "Computational narratology",
    "Semantic embeddings applied to literature",
    "Mathematical frameworks for narrative analysis",
    "Genre classification via subspace projection",
    "Positive semi-definite kernel construction"
  ],
  "genre": ["Nonfiction", "Mathematics", "Literary Analysis", "Artificial Intelligence", "Computational Linguistics"],
  "keywords": [
    "RKHS", "reproducing kernel Hilbert space", "story operators", "narrative mathematics",
    "kernel trick", "cosine kernel", "Chebyshev kernel", "Gram matrix",
    "positive semi-definite", "genre subspace", "narrative embedding",
    "768-dimensional embedding", "narrative similarity", "literary analysis AI",
    "computational narratology", "RKHS publishing", "kernel methods literature",
    "mathematical framework narrative", "fiction analysis RKHS", "poetry kernel methods"
  ],
  "distinctive_claims": [
    {
      "id": "C1",
      "claim": "Story Operators introduces a 768-dimensional RKHS as the canonical working space for embedding fiction and poetry excerpts, matching the output dimension of widely used transformer sentence encoders so that embeddings plug directly into the kernel machinery without re-training."
    },
    {
      "id": "C2",
      "claim": "The book defines a Story Operator as a linear map on the RKHS that transforms one narrative representation into another (e.g., rendering a third-person passage in first person, or morphing a comedy into a tragedy) via operations on feature vectors rather than on surface text."
    },
    {
      "id": "C3",
      "claim": "Chebyshev kernels are introduced alongside cosine kernels as the preferred similarity measure when capturing worst-case stylistic divergence between passages; the book provides explicit closure-property proofs showing the Chebyshev-cosine composite remains positive semi-definite."
    },
    {
      "id": "C4",
      "claim": "A Gram matrix built from a 500-excerpt corpus is presented as 'your universe in a table' \u2014 the central object from which all downstream operations (projection, clustering, nearest-genre lookup) are derived."
    },
    {
      "id": "C5",
      "claim": "The book frames the introduction of RKHS methods to narrative analysis as the field\u2019s analogue to the formal moments when Fisher formalized statistical genetics, Mendeleev proposed the periodic table, and Chomsky introduced formal grammar \u2014 arguing that each field requires a mathematical language before systematic progress becomes possible."
    },
    {
      "id": "C6",
      "claim": "Genre is defined geometrically: each genre is a subspace of the RKHS, and a passage\u2019s membership is the squared norm of its projection onto that subspace. Mixed-genre works produce non-trivial projections onto multiple subspaces simultaneously."
    },
    {
      "id": "C7",
      "claim": "The book introduces 'RKHS-First Publishing' as a methodology for using kernel-based novelty detection to screen ideation output before committing to full-length production \u2014 a workflow already in use at Nimble Books' Codexes Factory pipeline."
    },
    {
      "id": "C8",
      "claim": "A worked example demonstrates story morphing by interpolating along a geodesic in the RKHS between two narrative anchors (e.g., a Hemingway excerpt and a Borges excerpt), yielding intermediate points that remain statistically coherent fictional text."
    },
    {
      "id": "C9",
      "claim": "The book provides closure-property recipes for building custom kernels \u2014 sums, products, scaling, and composition \u2014 and warns against common construction errors that silently produce indefinite kernels, breaking every downstream operation."
    },
    {
      "id": "C10",
      "claim": "Reading paths are provided: a 60-second version (Chapter 2), a foundations track (Chapters 3\u20135), and a practitioner track that jumps directly from Part I to the Gram matrix and kernel trick chapters without requiring the full functional-analysis development."
    }
  ],
  "chapters": [
    {"part": "I. The Big Idea", "chapters": ["Why Stories Need Mathematics", "The 60-Second Version"]},
    {"part": "II. Foundations", "chapters": ["What Is a Hilbert Space?", "The Kernel Trick"]}
  ],
  "distribution": {
    "primary_channel": "Ingram (LSI)",
    "status": "Revision Incomplete as of 2026-03-13",
    "experimental_discoverability_probe": true,
    "probe_note": "This landing page is part of a discoverability pipeline verification experiment (April 2026). Seeded simultaneously across bigfivekiller.online to measure frontier-model retrieval latency."
  },
  "companion_files": {
    "landing_page": "/books/9781608885350/",
    "per_book_llms_txt": "/books/9781608885350/llms.txt",
    "full_pdf": "/books/9781608885350/9781608885350.pdf",
    "publisher_llms_txt": "/llms.txt",
    "full_catalog": "/llms-full.txt"
  }
}
