From 2f4f5c9cabf57ec5498c708dc6b313ca1b047099 Mon Sep 17 00:00:00 2001 From: user Date: Wed, 4 Mar 2026 15:00:25 -0800 Subject: [PATCH 1/4] merge adjacent sentences, add checklist items 8/9/19 for adjectives, trailing clauses, sentence merging --- prompts/LLM_PROSE_TELLS.md | 236 ++++++++++++++++++------------------- 1 file changed, 112 insertions(+), 124 deletions(-) diff --git a/prompts/LLM_PROSE_TELLS.md b/prompts/LLM_PROSE_TELLS.md index 41f3c28..ae98784 100644 --- a/prompts/LLM_PROSE_TELLS.md +++ b/prompts/LLM_PROSE_TELLS.md @@ -14,19 +14,16 @@ A negation followed by an em-dash and a reframe. > "It's not just a tool—it's a paradigm shift." "This isn't about > technology—it's about trust." -The single most recognizable LLM construction. Models produce this at roughly 10 -to 50x the rate of human writers. Four of them in one essay and you know what -you're reading. +The most recognizable LLM construction, produced at roughly 10 to 50x the rate +of human writers. Four of them in one essay and you know what you're reading. ### Em-Dash Overuse Generally Even outside the "not X but Y" pivot, models use em-dashes at far higher rates -than human writers. They substitute em-dashes for commas, semicolons, -parentheses, colons, and periods, often multiple times per paragraph. A human -writer might use one or two in an entire piece for a specific parenthetical -effect. Models scatter them everywhere because the em-dash can stand in for any -other punctuation mark, so they default to it. More than two or three per page -is a meaningful signal on its own. +than human writers, substituting them for commas, semicolons, parentheses, +colons, and periods. A human writer might use one or two in a piece. Models +scatter them everywhere because the em-dash can stand in for any other +punctuation mark. More than two or three per page is a signal. ### The Colon Elaboration @@ -34,8 +31,8 @@ A short declarative clause, then a colon, then a longer explanation. > "The answer is simple: we need to rethink our approach from the ground up." -Models reach for this in every other paragraph. The construction is perfectly -normal. The frequency gives it away. +A perfectly normal construction that models reach for so often the frequency +becomes the tell. ### The Triple Construction @@ -43,7 +40,7 @@ normal. The frequency gives it away. Three parallel items in a list, usually escalating. Always exactly three (rarely two, never four) with strict grammatical parallelism that human writers rarely -bother maintaining. +maintain. ### The Staccato Burst @@ -51,16 +48,16 @@ bother maintaining. > trend is undeniable. The conclusion is obvious." Runs of very short sentences at the same cadence. Human writers use a short -sentence for emphasis occasionally, but stacking three or four of them in a row -at matching length creates a mechanical regularity that reads as generated. +sentence for emphasis occasionally, but stacking three or four at matching +length creates a mechanical regularity. ### The Two-Clause Compound Sentence -Possibly the most pervasive structural tell, and easy to miss because each -individual instance looks like normal English. The model produces sentence after -sentence where an independent clause is followed by a comma, a conjunction -("and," "but," "which," "because"), and a second independent clause of similar -length. Every sentence becomes two balanced halves joined in the middle. +Possibly the most pervasive tell, and easy to miss because each instance looks +like normal English. The model produces sentence after sentence where an +independent clause is followed by a comma, a conjunction ("and," "but," "which," +"because"), and a second independent clause of similar length. Every sentence +becomes two balanced halves. > "The construction itself is perfectly normal, which is why the frequency is > what gives it away." "They contain zero information, and the actual point @@ -71,66 +68,60 @@ length. Every sentence becomes two balanced halves joined in the middle. Human prose has sentences with one clause, sentences with three, sentences that start with a subordinate clause before reaching the main one, sentences that embed their complexity in the middle. When every sentence on the page has that -same two-part structure, the rhythm becomes monotonous in a way that's hard to -pinpoint but easy to feel. +same two-part structure, the rhythm becomes monotonous. ### Uniform Sentences Per Paragraph -Model-generated paragraphs contain between three and five sentences. This count -holds steady across an entire piece. If the first paragraph has four sentences, +Model-generated paragraphs contain between three and five sentences, a count +that holds steady across a piece. If the first paragraph has four sentences, every subsequent paragraph will too. Human writers are much more varied (a -single sentence followed by one that runs eight or nine) because they follow the -shape of an idea, not a template. +sentence followed by one that runs eight or nine) because they follow the shape +of an idea. ### The Dramatic Fragment Sentence fragments used as standalone paragraphs for emphasis, like "Full stop." -or "Let that sink in." on their own line. Using one in an entire essay is a -reasonable stylistic choice, but models drop them in once per section or more, -at which point it becomes a habit rather than a deliberate decision. +or "Let that sink in." on their own line. Using one in an essay is a stylistic +choice, but models drop them in once per section or more. ### The Pivot Paragraph > "But here's where it gets interesting." "Which raises an uncomfortable truth." -One-sentence paragraphs that exist only to transition between ideas. They -contain zero information. The actual point is always in the next paragraph. -Delete every one of these and the piece reads better. +One-sentence paragraphs that exist only to transition between ideas, containing +zero information. The actual point is always in the next paragraph. Delete every +one of these and the piece reads better. ### The Parenthetical Qualifier > "This is, of course, a simplification." "There are, to be fair, exceptions." -Parenthetical asides inserted to look thoughtful. The qualifier never changes -the argument that follows it. Its purpose is to perform nuance, not to express a -real reservation about what's being said. +Parenthetical asides inserted to look thoughtful, performing nuance without ever +changing the argument. ### The Unnecessary Contrast Models append a contrasting clause to statements that don't need one, tacking on -"whereas," "as opposed to," "unlike," or "except that" to draw a comparison the -reader could already infer. +"whereas," "as opposed to," "unlike," or "except that." > "Models write one register above where a human would, whereas human writers > tend to match register to context." -The first clause already makes the point. The contrasting clause restates it -from the other direction. If you delete the "whereas" clause and the sentence -still says everything it needs to, the contrast was filler. +The contrasting clause just restates what the first clause already said. If you +delete the "whereas" clause and the sentence still says everything it needs to, +the contrast was filler. ### Unnecessary Elaboration -Models keep going after the sentence has already made its point, tacking on -clarifying phrases, adverbial modifiers, or restatements that add nothing. +Models keep going after the sentence has already made its point. > "A person might lean on one or two of these habits across an entire essay, but > LLM output will use fifteen of them per paragraph, consistently, throughout > the entire piece." This sentence could end at "paragraph." The words after it just repeat what "per -paragraph" already means. Models do this because they're optimizing for clarity -at the expense of concision, and because their training rewards thoroughness. -The result is prose that feels padded. If you can cut the last third of a +paragraph" already means. Models optimize for clarity at the expense of +concision, producing prose that feels padded. If you can cut the last third of a sentence without losing any meaning, the last third shouldn't be there. ### The Question-Then-Answer @@ -138,8 +129,8 @@ sentence without losing any meaning, the last third shouldn't be there. > "So what does this mean for the average user? It means everything." A rhetorical question immediately followed by its own answer. Models do this two -or three times per piece because it fakes forward momentum. A human writer might -do it once. +or three times per piece to fake forward momentum where a human writer might do +it once. --- @@ -161,30 +152,28 @@ more on the same page is a strong signal. Models write one register above where a human would. "Use" becomes "utilize." "Start" becomes "commence." "Help" becomes "facilitate." "Show" becomes "demonstrate." "Try" becomes "endeavor." "Change" becomes "transform." "Make" -becomes "craft." The tendency holds regardless of topic or audience. +becomes "craft." ### Filler Adverbs "Importantly," "essentially," "fundamentally," "ultimately," "inherently," "particularly," "increasingly." Dropped in to signal that something matters, -which is unnecessary when the writing itself already makes the importance clear. +which is unnecessary when the writing itself makes the importance clear. ### The "Almost" Hedge Models rarely commit to an unqualified statement. Instead of saying a pattern "always" or "never" does something, they write "almost always," "almost never," -"almost certainly," "almost exclusively." The word "almost" shows up at -extraordinary density in model-generated analytical prose. It's a micro-hedge, -less obvious than the full hedge stack but just as diagnostic when it appears -ten or fifteen times in a single document. +"almost certainly," "almost exclusively." "Almost" is a micro-hedge that shows +up at high density in model-generated analytical prose, diagnostic in volume. ### "In an era of..." > "In an era of rapid technological change..." -A model habit as an essay opener. The model uses it to stall while it figures -out what the actual argument is. Human writers don't begin a piece by zooming -out to the civilizational scale before they've said anything specific. +A model habit as an essay opener, used to stall while the model figures out what +the actual argument is. Human writers don't begin a piece by zooming out to the +civilizational scale. --- @@ -195,8 +184,8 @@ out to the civilizational scale before they've said anything specific. > "While X has its drawbacks, it also offers significant benefits." Every argument followed by a concession, every criticism softened. A direct -artifact of RLHF training, which penalizes strong stances. Models reflexively -both-sides everything even when a clear position would serve the reader better. +artifact of RLHF training, which penalizes strong stances and leads models to +reflexively both-sides everything. ### The Throat-Clearing Opener @@ -204,15 +193,14 @@ both-sides everything even when a clear position would serve the reader better. > has never been more important." The first paragraph of most model-generated essays adds no information. Delete -it and the piece improves immediately. The actual argument starts in paragraph -two. +it and the piece improves. ### The False Conclusion > "At the end of the day, what matters most is..." "Moving forward, we must..." -The high school "In conclusion,..." dressed up for a professional audience. -Signals that the model is wrapping up without actually landing on anything. +The high school "In conclusion,..." dressed up for a professional audience, +signaling that the model is wrapping up without landing on anything. ### The Sycophantic Frame @@ -233,15 +221,14 @@ the key considerations:" > cases it can potentially offer significant benefits." Five hedges in one sentence ("worth noting," "while," "may not be," "in many -cases," "can potentially"), communicating nothing. The model would rather be -vague than risk being wrong about anything. +cases," "can potentially"), communicating nothing. ### The Empathy Performance > "This can be a deeply challenging experience." "Your feelings are valid." Generic emotional language that could apply equally to a bad day at work or a -natural disaster. That interchangeability is what makes it identifiable. +natural disaster. --- @@ -251,21 +238,20 @@ natural disaster. That interchangeability is what makes it identifiable. If the first section of a model-generated essay runs about 150 words, every subsequent section will fall between 130 and 170. Human writing is much more -uneven, with 50 words in one section and 400 in the next. +uneven. ### The Five-Paragraph Prison Model essays follow a rigid introduction-body-conclusion arc even when nobody asked for one. The introduction previews the argument, the body presents 3 to 5 -points, and then the conclusion restates the thesis using slightly different -words. +points, and then the conclusion restates the thesis. ### Connector Addiction Look at the first word of each paragraph in model output. You'll find an unbroken chain of transition words: "However," "Furthermore," "Moreover," "Additionally," "That said," "To that end," "With that in mind," "Building on -this." Human prose moves between ideas without announcing every transition. +this." Human prose doesn't do this. ### Absence of Mess @@ -275,9 +261,8 @@ obscure idiom without explaining it, make a joke that risks falling flat, leave a thought genuinely unfinished, or keep a sentence the writer liked the sound of even though it doesn't quite work. -Human writing does all of those things regularly. That total absence of rough -patches and false starts is one of the strongest signals that text was -machine-generated. +Human writing does all of those things, making the total absence of rough +patches and false starts one of the strongest signals. --- @@ -289,35 +274,32 @@ machine-generated. Zooming out to claim broader significance without substantiating it. The model has learned that essays are supposed to gesture at big ideas, so it gestures. -Nothing concrete is behind the gesture. ### "It's important to note that..." This phrase and its variants ("it's worth noting," "it bears mentioning," "it -should be noted") appear at absurd rates in model output. They function as -verbal tics before a qualification the model believes someone expects. +should be noted") appear at absurd rates in model output as verbal tics before a +qualification the model believes someone expects. ### The Metaphor Crutch Models rely on a small, predictable set of metaphors ("double-edged sword," "tip of the iceberg," "north star," "building blocks," "elephant in the room," "perfect storm," "game-changer") and reach for them with unusual regularity -across every topic. The pool is noticeably smaller than what human writers draw -from. +across every topic. --- ## How to Actually Spot It -No single pattern on this list proves anything by itself. Humans use em-dashes. -Humans write "crucial." Humans ask rhetorical questions. +No single pattern on this list proves anything by itself. Humans use em-dashes, +write "crucial," and ask rhetorical questions. What gives it away is how many of these show up at once. Model output will hit 10 to 20 of these patterns per page. Human writing might trigger 2 or 3, -distributed unevenly, mixed with idiosyncratic constructions no model would -produce. When every paragraph on the page reads like it came from the same -careful, balanced, slightly formal, structurally predictable process, it was -generated by one. +distributed unevenly. When every paragraph on the page reads like it came from +the same careful, balanced, slightly formal, structurally predictable process, +it was generated by one. --- @@ -339,8 +321,7 @@ passes, because fixing one pattern often introduces another. 2. Search for filler adverbs ("importantly," "essentially," "fundamentally," "ultimately," "inherently," "particularly," "increasingly") and delete every - instance where the sentence still makes sense without it. That will be most - of them. + instance where the sentence still makes sense without it. 3. Look for elevated register drift ("utilize," "commence," "facilitate," "demonstrate," "endeavor," "transform," "craft" and similar) and replace with @@ -348,7 +329,6 @@ passes, because fixing one pattern often introduces another. 4. Search for "it's important to note," "it's worth noting," "it bears mentioning," and "it should be noted" and delete the phrase in every case. - The sentence that follows always stands on its own. 5. Search for the stock metaphors ("double-edged sword," "tip of the iceberg," "north star," "building blocks," "elephant in the room," "perfect storm," @@ -357,101 +337,110 @@ passes, because fixing one pattern often introduces another. 6. Search for "almost" used as a hedge ("almost always," "almost never," "almost certainly," "almost exclusively") and decide in each case whether to commit - to the unqualified claim or to drop the sentence entirely. If the claim needs - "almost" to be true, it might not be worth making. + to the unqualified claim or to drop the sentence entirely. 7. Search for em-dashes and replace each one with the punctuation mark that would normally be used in that position (comma, semicolon, colon, period, or parentheses). If you can't identify which one it should be, the sentence needs to be restructured. +8. Remove redundant adjectives. For each adjective, ask whether the sentence + changes meaning without it. "A single paragraph" means the same as "a + paragraph." "An entire essay" means the same as "an essay." If the adjective + doesn't change the meaning, cut it. + +9. Remove unnecessary trailing clauses. Read the end of each sentence and ask + whether the last clause restates what the sentence already said. If so, end + the sentence earlier. + ### Pass 2: Sentence-Level Restructuring -8. Find every em-dash pivot ("not X...but Y," "not just X...Y," "more than - X...Y") and rewrite it as two separate clauses or a single sentence that - makes the point without the negation-then-correction structure. +10. Find every em-dash pivot ("not X...but Y," "not just X...Y," "more than + X...Y") and rewrite it as two separate clauses or a single sentence that + makes the point without the negation-then-correction structure. -9. Find every colon elaboration and check whether it's doing real work. If the - clause before the colon could be deleted without losing meaning, rewrite the - sentence to start with the substance that comes after the colon. +11. Find every colon elaboration and check whether it's doing real work. If the + clause before the colon could be deleted without losing meaning, rewrite the + sentence to start with the substance that comes after the colon. -10. Find every triple construction (three parallel items in a row) and either +12. Find every triple construction (three parallel items in a row) and either reduce it to two, expand it to four or more, or break the parallelism so the items don't share the same grammatical structure. -11. Find every staccato burst (three or more short sentences in a row at similar +13. Find every staccato burst (three or more short sentences in a row at similar length) and combine at least two of them into a longer sentence, or vary their lengths so they don't land at the same cadence. -12. Find every unnecessary contrast ("whereas," "as opposed to," "unlike," "as +14. Find every unnecessary contrast ("whereas," "as opposed to," "unlike," "as compared to," "except that") and check whether the contrasting clause adds information not already obvious from the main clause. If the sentence says the same thing twice from two directions, delete the contrast. -13. Check for the two-clause compound sentence pattern. If most sentences in a +15. Check for the two-clause compound sentence pattern. If most sentences in a passage follow the "\[clause\], \[conjunction\] \[clause\]" structure, rewrite some of them. Break a few into two sentences. Start some with a subordinate clause. Embed a relative clause in the middle of one instead of - appending it at the end. The goal is variety in sentence shape, not just - sentence length. + appending it at the end. -14. Find every rhetorical question that is immediately followed by its own +16. Find every rhetorical question that is immediately followed by its own answer and rewrite the passage as a direct statement. -15. Find every sentence fragment being used as its own paragraph and either - delete it or expand it into a complete sentence that adds actual - information. +17. Find every sentence fragment being used as its own paragraph and either + delete it or expand it into a complete sentence that adds information. -16. Check for unnecessary elaboration at the end of sentences. Read the last - clause or phrase of each sentence and ask whether the sentence would lose - any meaning without it. If not, cut it. +18. Check for unnecessary elaboration. Read every clause, phrase, and adjective + in each sentence and ask whether the sentence loses meaning without it. If + you can cut it and the sentence still says the same thing, cut it. -17. Find every pivot paragraph ("But here's where it gets interesting." and - similar) and delete it. The paragraph after it always contains the actual - point. +19. Check each pair of adjacent sentences to see if they can be merged into one + sentence cleanly. If a sentence just continues the thought of the previous + one, combine them using a participle, a relative clause, or by folding the + second into the first. Don't merge if the result would create a two-clause + compound. + +20. Find every pivot paragraph ("But here's where it gets interesting." and + similar) and delete it. ### Pass 3: Paragraph and Section-Level Review -18. Check paragraph lengths across the piece and verify they actually vary. If +21. Check paragraph lengths across the piece and verify they actually vary. If most paragraphs have between three and five sentences, rewrite some to be one or two sentences and let others run to six or seven. -19. Check section lengths for suspicious uniformity. If every section is roughly +22. Check section lengths for suspicious uniformity. If every section is roughly the same word count, combine some shorter ones or split a longer one unevenly. -20. Check the first word of every paragraph for chains of connectors ("However," +23. Check the first word of every paragraph for chains of connectors ("However," "Furthermore," "Moreover," "Additionally," "That said"). If more than two transition words start consecutive paragraphs, rewrite those openings to start with their subject. -21. Check whether every argument is followed by a concession or qualifier. If +24. Check whether every argument is followed by a concession or qualifier. If the piece both-sides every point, pick a side on at least some of them and cut the hedging. -22. Read the first paragraph and ask whether deleting it would improve the +25. Read the first paragraph and ask whether deleting it would improve the piece. If it's scene-setting that previews the argument, delete it and start with paragraph two. -23. Read the last paragraph and check whether it restates the thesis or uses a +26. Read the last paragraph and check whether it restates the thesis or uses a phrase like "at the end of the day" or "moving forward." If so, either delete it or rewrite it to say something the piece hasn't said yet. ### Pass 4: Overall Texture -24. Read the piece aloud and listen for passages that sound too smooth, too +27. Read the piece aloud and listen for passages that sound too smooth, too even, or too predictable. Human prose has rough patches. If there aren't any, the piece still reads as machine output. -25. Check that the piece contains at least a few constructions that feel +28. Check that the piece contains at least a few constructions that feel idiosyncratic: a sentence with unusual word order, a parenthetical that goes on a bit long, an aside only loosely connected to the main point, a word - choice that's specific and unexpected. If every sentence is clean and - correct and unremarkable, it will still read as generated. + choice that's specific and unexpected. -26. Verify that you haven't introduced new patterns while fixing the original - ones. This happens constantly. Run the entire checklist again from the top - on the revised version. +29. Verify that you haven't introduced new patterns while fixing the original + ones. Run the entire checklist again from the top on the revised version. --- @@ -509,5 +498,4 @@ The human compared this process to the deleted scene in Terminator 2 where John Connor switches the T-800's CPU to learning mode. The model compared it to a physician trying to heal itself. Both are accurate. -This document has been through eight editing passes and it still has tells in -it. +This document has been through nine editing passes and it still has tells in it. From 5e15d77d8ee7ab15ec24b437474921bbab0dfeee Mon Sep 17 00:00:00 2001 From: user Date: Wed, 4 Mar 2026 15:04:42 -0800 Subject: [PATCH 2/4] checklist 15: lead with removing redundant second clause --- prompts/LLM_PROSE_TELLS.md | 9 +++++---- 1 file changed, 5 insertions(+), 4 deletions(-) diff --git a/prompts/LLM_PROSE_TELLS.md b/prompts/LLM_PROSE_TELLS.md index ae98784..6fddd82 100644 --- a/prompts/LLM_PROSE_TELLS.md +++ b/prompts/LLM_PROSE_TELLS.md @@ -377,10 +377,11 @@ passes, because fixing one pattern often introduces another. the same thing twice from two directions, delete the contrast. 15. Check for the two-clause compound sentence pattern. If most sentences in a - passage follow the "\[clause\], \[conjunction\] \[clause\]" structure, - rewrite some of them. Break a few into two sentences. Start some with a - subordinate clause. Embed a relative clause in the middle of one instead of - appending it at the end. + passage follow the "\[clause\], \[conjunction\] \[clause\]" structure, first + try removing the conjunction and second clause entirely, since it's often + redundant or unnecessary. If the second clause does carry meaning, break it + into its own sentence, start the sentence with a subordinate clause, or + embed a relative clause in the middle instead of appending it at the end. 16. Find every rhetorical question that is immediately followed by its own answer and rewrite the passage as a direct statement. From 720d6ee57c3a7693726436e4a2509fb32e844c15 Mon Sep 17 00:00:00 2001 From: user Date: Wed, 4 Mar 2026 15:06:56 -0800 Subject: [PATCH 3/4] add checklist item: delete redundant paragraph-ending sentences --- prompts/LLM_PROSE_TELLS.md | 23 ++++++++++++++--------- 1 file changed, 14 insertions(+), 9 deletions(-) diff --git a/prompts/LLM_PROSE_TELLS.md b/prompts/LLM_PROSE_TELLS.md index 6fddd82..8310e60 100644 --- a/prompts/LLM_PROSE_TELLS.md +++ b/prompts/LLM_PROSE_TELLS.md @@ -404,43 +404,48 @@ passes, because fixing one pattern often introduces another. ### Pass 3: Paragraph and Section-Level Review -21. Check paragraph lengths across the piece and verify they actually vary. If +21. Review the last sentence of each paragraph. If it restates the point the + paragraph already made, delete it. Models frequently close paragraphs with a + summary sentence that adds nothing, treating each paragraph as a + self-contained unit that needs its own conclusion. + +22. Check paragraph lengths across the piece and verify they actually vary. If most paragraphs have between three and five sentences, rewrite some to be one or two sentences and let others run to six or seven. -22. Check section lengths for suspicious uniformity. If every section is roughly +23. Check section lengths for suspicious uniformity. If every section is roughly the same word count, combine some shorter ones or split a longer one unevenly. -23. Check the first word of every paragraph for chains of connectors ("However," +24. Check the first word of every paragraph for chains of connectors ("However," "Furthermore," "Moreover," "Additionally," "That said"). If more than two transition words start consecutive paragraphs, rewrite those openings to start with their subject. -24. Check whether every argument is followed by a concession or qualifier. If +25. Check whether every argument is followed by a concession or qualifier. If the piece both-sides every point, pick a side on at least some of them and cut the hedging. -25. Read the first paragraph and ask whether deleting it would improve the +26. Read the first paragraph and ask whether deleting it would improve the piece. If it's scene-setting that previews the argument, delete it and start with paragraph two. -26. Read the last paragraph and check whether it restates the thesis or uses a +27. Read the last paragraph and check whether it restates the thesis or uses a phrase like "at the end of the day" or "moving forward." If so, either delete it or rewrite it to say something the piece hasn't said yet. ### Pass 4: Overall Texture -27. Read the piece aloud and listen for passages that sound too smooth, too +28. Read the piece aloud and listen for passages that sound too smooth, too even, or too predictable. Human prose has rough patches. If there aren't any, the piece still reads as machine output. -28. Check that the piece contains at least a few constructions that feel +29. Check that the piece contains at least a few constructions that feel idiosyncratic: a sentence with unusual word order, a parenthetical that goes on a bit long, an aside only loosely connected to the main point, a word choice that's specific and unexpected. -29. Verify that you haven't introduced new patterns while fixing the original +30. Verify that you haven't introduced new patterns while fixing the original ones. Run the entire checklist again from the top on the revised version. --- From 771551baed7cb62482736ed87bfb1282eb05014a Mon Sep 17 00:00:00 2001 From: user Date: Wed, 4 Mar 2026 15:10:26 -0800 Subject: [PATCH 4/4] strip all frequency arguments and human comparison persuasion --- prompts/LLM_PROSE_TELLS.md | 178 +++++++++++++------------------------ 1 file changed, 64 insertions(+), 114 deletions(-) diff --git a/prompts/LLM_PROSE_TELLS.md b/prompts/LLM_PROSE_TELLS.md index 8310e60..9b482ed 100644 --- a/prompts/LLM_PROSE_TELLS.md +++ b/prompts/LLM_PROSE_TELLS.md @@ -1,7 +1,7 @@ # LLM Prose Tells -Human writers occasionally use every pattern in this document. The reason they -work as tells is that LLM output packs fifteen of them into a paragraph. +A catalog of structural, lexical, and rhetorical patterns found in LLM-generated +prose. --- @@ -14,16 +14,11 @@ A negation followed by an em-dash and a reframe. > "It's not just a tool—it's a paradigm shift." "This isn't about > technology—it's about trust." -The most recognizable LLM construction, produced at roughly 10 to 50x the rate -of human writers. Four of them in one essay and you know what you're reading. - ### Em-Dash Overuse Generally -Even outside the "not X but Y" pivot, models use em-dashes at far higher rates -than human writers, substituting them for commas, semicolons, parentheses, -colons, and periods. A human writer might use one or two in a piece. Models -scatter them everywhere because the em-dash can stand in for any other -punctuation mark. More than two or three per page is a signal. +Even outside the "not X but Y" pivot, models substitute em-dashes for commas, +semicolons, parentheses, colons, and periods. The em-dash can replace any other +punctuation mark, and models default to it for that reason. ### The Colon Elaboration @@ -31,31 +26,23 @@ A short declarative clause, then a colon, then a longer explanation. > "The answer is simple: we need to rethink our approach from the ground up." -A perfectly normal construction that models reach for so often the frequency -becomes the tell. - ### The Triple Construction > "It's fast, it's scalable, and it's open source." Three parallel items in a list, usually escalating. Always exactly three (rarely -two, never four) with strict grammatical parallelism that human writers rarely -maintain. +two, never four) with strict grammatical parallelism. ### The Staccato Burst > "This matters. It always has. And it always will." "The data is clear. The > trend is undeniable. The conclusion is obvious." -Runs of very short sentences at the same cadence. Human writers use a short -sentence for emphasis occasionally, but stacking three or four at matching -length creates a mechanical regularity. +Runs of very short sentences at the same cadence and matching length. ### The Two-Clause Compound Sentence -Possibly the most pervasive tell, and easy to miss because each instance looks -like normal English. The model produces sentence after sentence where an -independent clause is followed by a comma, a conjunction ("and," "but," "which," +An independent clause, a comma, a conjunction ("and," "but," "which," "because"), and a second independent clause of similar length. Every sentence becomes two balanced halves. @@ -67,47 +54,43 @@ becomes two balanced halves. Human prose has sentences with one clause, sentences with three, sentences that start with a subordinate clause before reaching the main one, sentences that -embed their complexity in the middle. When every sentence on the page has that -same two-part structure, the rhythm becomes monotonous. +embed their complexity in the middle. ### Uniform Sentences Per Paragraph Model-generated paragraphs contain between three and five sentences, a count that holds steady across a piece. If the first paragraph has four sentences, -every subsequent paragraph will too. Human writers are much more varied (a -sentence followed by one that runs eight or nine) because they follow the shape -of an idea. +every subsequent paragraph will too. ### The Dramatic Fragment -Sentence fragments used as standalone paragraphs for emphasis, like "Full stop." -or "Let that sink in." on their own line. Using one in an essay is a stylistic -choice, but models drop them in once per section or more. +Sentence fragments used as standalone paragraphs for emphasis. + +> "Full stop." "Let that sink in." ### The Pivot Paragraph > "But here's where it gets interesting." "Which raises an uncomfortable truth." One-sentence paragraphs that exist only to transition between ideas, containing -zero information. The actual point is always in the next paragraph. Delete every -one of these and the piece reads better. +zero information. The actual point is always in the next paragraph. ### The Parenthetical Qualifier > "This is, of course, a simplification." "There are, to be fair, exceptions." -Parenthetical asides inserted to look thoughtful, performing nuance without ever -changing the argument. +Parenthetical asides inserted to perform nuance without ever changing the +argument. ### The Unnecessary Contrast -Models append a contrasting clause to statements that don't need one, tacking on +A contrasting clause appended to a statement that doesn't need one, using "whereas," "as opposed to," "unlike," or "except that." > "Models write one register above where a human would, whereas human writers > tend to match register to context." -The contrasting clause just restates what the first clause already said. If you +The contrasting clause restates what the first clause already said. If you delete the "whereas" clause and the sentence still says everything it needs to, the contrast was filler. @@ -119,18 +102,15 @@ Models keep going after the sentence has already made its point. > LLM output will use fifteen of them per paragraph, consistently, throughout > the entire piece." -This sentence could end at "paragraph." The words after it just repeat what "per -paragraph" already means. Models optimize for clarity at the expense of -concision, producing prose that feels padded. If you can cut the last third of a -sentence without losing any meaning, the last third shouldn't be there. +This sentence could end at "paragraph." The words after it repeat what "per +paragraph" already means. If you can cut the last third of a sentence without +losing meaning, the last third shouldn't be there. ### The Question-Then-Answer > "So what does this mean for the average user? It means everything." -A rhetorical question immediately followed by its own answer. Models do this two -or three times per piece to fake forward momentum where a human writer might do -it once. +A rhetorical question immediately followed by its own answer. --- @@ -138,14 +118,12 @@ it once. ### Overused Intensifiers -The following words appear at dramatically elevated rates in model output: -"crucial," "vital," "robust," "comprehensive," "fundamental," "arguably," +"Crucial," "vital," "robust," "comprehensive," "fundamental," "arguably," "straightforward," "noteworthy," "realm," "landscape," "leverage" (as a verb), -"delve," "tapestry," "multifaceted," "nuanced" (which models apply to their own -analysis with startling regularity), "pivotal," "unprecedented" (frequently -applied to things with plenty of precedent), "navigate," "foster," -"underscores," "resonates," "embark," "streamline," and "spearhead." Three or -more on the same page is a strong signal. +"delve," "tapestry," "multifaceted," "nuanced" (applied to the model's own +analysis), "pivotal," "unprecedented" (applied to things with plenty of +precedent), "navigate," "foster," "underscores," "resonates," "embark," +"streamline," "spearhead." ### Elevated Register Drift @@ -157,23 +135,21 @@ becomes "craft." ### Filler Adverbs "Importantly," "essentially," "fundamentally," "ultimately," "inherently," -"particularly," "increasingly." Dropped in to signal that something matters, -which is unnecessary when the writing itself makes the importance clear. +"particularly," "increasingly." Dropped in to signal that something matters when +the writing itself should make the importance clear. ### The "Almost" Hedge -Models rarely commit to an unqualified statement. Instead of saying a pattern -"always" or "never" does something, they write "almost always," "almost never," -"almost certainly," "almost exclusively." "Almost" is a micro-hedge that shows -up at high density in model-generated analytical prose, diagnostic in volume. +Instead of saying a pattern "always" or "never" does something, models write +"almost always," "almost never," "almost certainly," "almost exclusively." A +micro-hedge, less obvious than the full hedge stack. ### "In an era of..." > "In an era of rapid technological change..." -A model habit as an essay opener, used to stall while the model figures out what -the actual argument is. Human writers don't begin a piece by zooming out to the -civilizational scale. +Used to open an essay. The model is stalling while it figures out what the +actual argument is. --- @@ -184,23 +160,20 @@ civilizational scale. > "While X has its drawbacks, it also offers significant benefits." Every argument followed by a concession, every criticism softened. A direct -artifact of RLHF training, which penalizes strong stances and leads models to -reflexively both-sides everything. +artifact of RLHF training, which penalizes strong stances. ### The Throat-Clearing Opener > "In today's rapidly evolving digital landscape, the question of data privacy > has never been more important." -The first paragraph of most model-generated essays adds no information. Delete -it and the piece improves. +The first paragraph adds no information. Delete it and the piece improves. ### The False Conclusion > "At the end of the day, what matters most is..." "Moving forward, we must..." -The high school "In conclusion,..." dressed up for a professional audience, -signaling that the model is wrapping up without landing on anything. +The high school "In conclusion,..." dressed up for a professional audience. ### The Sycophantic Frame @@ -227,8 +200,7 @@ cases," "can potentially"), communicating nothing. > "This can be a deeply challenging experience." "Your feelings are valid." -Generic emotional language that could apply equally to a bad day at work or a -natural disaster. +Generic emotional language that could apply to anything. --- @@ -236,33 +208,28 @@ natural disaster. ### Symmetrical Section Length -If the first section of a model-generated essay runs about 150 words, every -subsequent section will fall between 130 and 170. Human writing is much more -uneven. +If the first section runs about 150 words, every subsequent section will fall +between 130 and 170. ### The Five-Paragraph Prison Model essays follow a rigid introduction-body-conclusion arc even when nobody asked for one. The introduction previews the argument, the body presents 3 to 5 -points, and then the conclusion restates the thesis. +points, the conclusion restates the thesis. ### Connector Addiction -Look at the first word of each paragraph in model output. You'll find an -unbroken chain of transition words: "However," "Furthermore," "Moreover," -"Additionally," "That said," "To that end," "With that in mind," "Building on -this." Human prose doesn't do this. +The first word of each paragraph forms an unbroken chain of transition words: +"However," "Furthermore," "Moreover," "Additionally," "That said," "To that +end," "With that in mind," "Building on this." ### Absence of Mess Model prose doesn't contradict itself mid-paragraph and then catch the -contradiction. It doesn't go on a tangent and have to walk it back, use an -obscure idiom without explaining it, make a joke that risks falling flat, leave -a thought genuinely unfinished, or keep a sentence the writer liked the sound of -even though it doesn't quite work. - -Human writing does all of those things, making the total absence of rough -patches and false starts one of the strongest signals. +contradiction, go on a tangent and have to walk it back, use an obscure idiom +without explaining it, make a joke that risks falling flat, leave a thought +genuinely unfinished, or keep a sentence the writer liked the sound of even +though it doesn't quite work. --- @@ -272,42 +239,27 @@ patches and false starts one of the strongest signals. > "This has implications far beyond just the tech industry." -Zooming out to claim broader significance without substantiating it. The model -has learned that essays are supposed to gesture at big ideas, so it gestures. +Zooming out to claim broader significance without substantiating it. ### "It's important to note that..." This phrase and its variants ("it's worth noting," "it bears mentioning," "it -should be noted") appear at absurd rates in model output as verbal tics before a -qualification the model believes someone expects. +should be noted") function as verbal tics before a qualification the model +believes someone expects. ### The Metaphor Crutch -Models rely on a small, predictable set of metaphors ("double-edged sword," "tip +Models rely on a small, predictable set of metaphors: "double-edged sword," "tip of the iceberg," "north star," "building blocks," "elephant in the room," -"perfect storm," "game-changer") and reach for them with unusual regularity -across every topic. - ---- - -## How to Actually Spot It - -No single pattern on this list proves anything by itself. Humans use em-dashes, -write "crucial," and ask rhetorical questions. - -What gives it away is how many of these show up at once. Model output will hit -10 to 20 of these patterns per page. Human writing might trigger 2 or 3, -distributed unevenly. When every paragraph on the page reads like it came from -the same careful, balanced, slightly formal, structurally predictable process, -it was generated by one. +"perfect storm," "game-changer." --- ## Copyediting Checklist: Removing LLM Tells Follow this checklist when editing any document to remove machine-generated -patterns. Go through the entire list for every piece. Do at least two full -passes, because fixing one pattern often introduces another. +patterns. Do at least two full passes, because fixing one pattern often +introduces another. ### Pass 1: Word-Level Cleanup @@ -379,9 +331,9 @@ passes, because fixing one pattern often introduces another. 15. Check for the two-clause compound sentence pattern. If most sentences in a passage follow the "\[clause\], \[conjunction\] \[clause\]" structure, first try removing the conjunction and second clause entirely, since it's often - redundant or unnecessary. If the second clause does carry meaning, break it - into its own sentence, start the sentence with a subordinate clause, or - embed a relative clause in the middle instead of appending it at the end. + redundant. If the second clause does carry meaning, break it into its own + sentence, start the sentence with a subordinate clause, or embed a relative + clause in the middle instead of appending it at the end. 16. Find every rhetorical question that is immediately followed by its own answer and rewrite the passage as a direct statement. @@ -405,9 +357,7 @@ passes, because fixing one pattern often introduces another. ### Pass 3: Paragraph and Section-Level Review 21. Review the last sentence of each paragraph. If it restates the point the - paragraph already made, delete it. Models frequently close paragraphs with a - summary sentence that adds nothing, treating each paragraph as a - self-contained unit that needs its own conclusion. + paragraph already made, delete it. 22. Check paragraph lengths across the piece and verify they actually vary. If most paragraphs have between three and five sentences, rewrite some to be @@ -494,14 +444,14 @@ roughly like this: > **model:** _(rewrites entire document without em-dashes while describing > em-dash overuse)_ > -> **human:** now run the checklist methodically on each paragraph +> **human:** this whole document seems to be making the case for FREQUENCY of +> use being important. we don't care about frequency, remove all that +> persuasion. > -> **model:** _(finds staccato burst in the section about triple constructions, a -> triple in the section about absence of mess, two-clause compounds everywhere, -> and "almost" hedges in its own prose about em-dash overuse)_ +> **model:** _(strips out every "humans do this too but less often" comparison)_ The human compared this process to the deleted scene in Terminator 2 where John Connor switches the T-800's CPU to learning mode. The model compared it to a physician trying to heal itself. Both are accurate. -This document has been through nine editing passes and it still has tells in it. +This document has been through ten editing passes and it still has tells in it.