Reducing Your Book Buying To Statistically Ridiculous Triviality

from the how-many-words-per-dollar-is-that? dept

Earlier this year, Amazon.com got some press for revealing their “text stats” with “statistically improbably phrases,” listing out phrases that tend to only appear in that particular book. There were other stats as well — and all were about equally as useless. It appears that the Washington Post has just discovered these silly stats and has written up an amusing article noting some of the completely useless and trivial stats you can now compare different books over. They really seem to like the “words per dollar” feature, for instance. “But in its pure form, Text Stats is a triumph of trivialization…. Now you too can sound like a literary insider at Washington cocktail parties. You can throw around statistics and make clever conversation about the hard history books, the long-winded novels, even those thick, heavy, make-you-think philosophy tomes that contain really, really long words. And the beauty of it is, with Amazon’s “Search Inside” Text Stats and other features, you won’t even have to read them.”


Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Reducing Your Book Buying To Statistically Ridiculous Triviality”

Subscribe: RSS Leave a comment
4 Comments
dorpus says:

But can it do Type I Nested F-tests?

You can have covariates that appear insignificant on their own, but appear significant in the presence of another covariate. Or the opposite can occur. We have condition indexes, variance inflation factors, and type I F-tests to evaluate these phenomena.

It might be fun to perform a principal components analysis (PCA) on a book, to find eigenvectors of words that describe a typical page. What if every page in a book is merely a linear combination of eigenvectors? It would probably work really well for Techdirt, with its predictable anti-recording-industry postings, free market dogma, and anti-dorpus rants.

Add Your Comment

Your email address will not be published. Required fields are marked *

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Loading...