{"id":18781,"date":"2024-12-10T08:30:03","date_gmt":"2024-12-10T08:30:03","guid":{"rendered":"\/?p=18781"},"modified":"2024-12-10T08:45:09","modified_gmt":"2024-12-10T08:45:09","slug":"many-fairly-random-and-chaotic-looking-things-are-often-nothing-surprising-at-all","status":"publish","type":"post","link":"\/?p=18781","title":{"rendered":"Many fairly random and chaotic looking things are often nothing surprising at all."},"content":{"rendered":"<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"bpqdu\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"bpqdu-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"bpqdu-0-0\">\n<p>https:\/\/x.com\/keenanisalive\/status\/1866251675440460234<\/p>\n<p><strong>Many fairly random and chaotic looking things are often nothing surprising at all.<\/strong><\/p>\n<p>Take any sequence and count unique values, the probability distribution ( as counts) tells you how many unique tokens are needed to store it losslessly. A log function fits for estimation, but so does the tail of a Poisson or normal distribution.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"6osei\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"6osei-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"6osei-0-0\"><span data-offset-key=\"6osei-0-0\">I just keep all the data now and only looks at statistics when I do not have the lossless original data. Much of STEMCFQ (science technology engineering mathematics computing finance quantiative_things) on the Internet uses analytic expressions from a past where it was hard to memorize whole tables of data. <\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"3q5te\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"3q5te-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"3q5te-0-0\"><span data-offset-key=\"3q5te-0-0\">\u00a0<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"ahd47\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"ahd47-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"ahd47-0-0\"><span data-offset-key=\"ahd47-0-0\">To me it is simply a &#8220;compression algorithm&#8221; because of the finite and incomplete memory of humans (and other species) . The more unique things there are to manage, the larger the memory required. If there are many all alike, but arriving more or less randomly, there is a cost to manage ones &#8220;almost all alike&#8221; and if they are normally distributed on another dimension, a mean and standard deviation and count is often useful.<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"m0gb\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"m0gb-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"m0gb-0-0\"><span data-offset-key=\"m0gb-0-0\">\u00a0<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"bprh9\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"bprh9-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"bprh9-0-0\"><span data-offset-key=\"bprh9-0-0\">The arguments over entropy come from its lack of precision and care. It sounds really magical, so it gets trotted out in many places on the Internet. As though anyone actually uses it.<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"fmsjo\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"fmsjo-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"fmsjo-0-0\"><span data-offset-key=\"fmsjo-0-0\">\u00a0<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"8dc4o\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"8dc4o-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"8dc4o-0-0\"><span data-offset-key=\"8dc4o-0-0\">The only place it has a specific meaning is in thermodynamics where any ratio of energy to temperature (Joules\/Kelvin) is an &#8220;entropy&#8221; Where I see it used it is simply to avoid doing the detailed statistics. When working with temperatures as proxies for energy contained or used or absorbed. Other than a unit, which does have meaning if people do a good job of tracking from the raw data to stuff put on the Internet. Generally I would avoid using it at all.<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"brqdk\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"brqdk-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"brqdk-0-0\">\n<p><span data-offset-key=\"brqdk-0-0\">As for surprise, that has to do with how much you have to memorize to not be surprised. The same training data (for machine learning) can be compressed significantly and if you can model relations between unique entities in the data, fairly random looking and chaotic looking things take on meaning, are controllable, and nothing surprising at all.<\/span><\/p>\n<p>Temperature is the problem because it does not directly relate to power or energy except where people are very careful and precise. That care and precision is rare on the Internet. If anyone saying entropy want to be precise, they need to show how they measure and define energy and power, spatial and temporal data in their systems. None of this is really hard, it is just tedious and takes great care.<\/p>\n<\/div>\n<\/div>\n<\/div>\n<div data-offset-key=\"brqdk-0-0\">\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"37tur\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"37tur-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"37tur-0-0\">( &#8220;entropy&#8221;) on google has 120 Million entries. That is NOT a concise and consistent global resource term. It is not hard to fix, just voluminous and tedious and no one gives a hoot. And you want to talk it out &#8211; in a &#8220;limited to text mostly&#8221; chat not capable of sharing mathematical models and real data?\u00a0 Better to work on &#8220;poverty&#8221; &#8220;knowledge for all&#8221; or &#8220;solar system exploration and development&#8221;.<\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"iqt2\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"iqt2-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"iqt2-0-0\"><span data-offset-key=\"iqt2-0-0\">\u00a0<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"iidg\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"iidg-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"iidg-0-0\"><a href=\"https:\/\/books.google.com\/ngrams\/graph?content=entropy%2C+entropic&amp;year_start=1800&amp;year_end=2022&amp;corpus=en&amp;smoothing=3\"><span data-offset-key=\"iidg-0-0\">https:\/\/books.google.com\/ngrams\/graph?content=entropy%2C+entropic&amp;year_start=1800&amp;year_end=2022&amp;corpus=en&amp;smoothing=3<\/span><\/a><\/div>\n<\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"c2tkp\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"c2tkp-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"c2tkp-0-0\"><span data-offset-key=\"c2tkp-0-0\">\u00a0<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"5\" data-rbd-draggable-id=\"eqt3t\">\n<div class=\"\" data-block=\"true\" data-editor=\"4s0fb\" data-offset-key=\"eqt3t-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"eqt3t-0-0\"><span data-offset-key=\"eqt3t-0-0\">Richard Collins, The Internet Foundation<\/span><\/div>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/x.com\/keenanisalive\/status\/1866251675440460234 Many fairly random and chaotic looking things are often nothing surprising at all. Take any sequence and count unique values, the probability distribution ( as counts) tells you how many unique tokens are needed to store it losslessly. A log function fits for estimation, but so does the tail of a Poisson or normal <br \/><a class=\"read-more-button\" href=\"\/?p=18781\">Read More &raquo;<\/a><\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[73],"tags":[],"class_list":["post-18781","post","type-post","status-publish","format-standard","hentry","category-all-knowledge"],"_links":{"self":[{"href":"\/index.php?rest_route=\/wp\/v2\/posts\/18781","targetHints":{"allow":["GET"]}}],"collection":[{"href":"\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=18781"}],"version-history":[{"count":6,"href":"\/index.php?rest_route=\/wp\/v2\/posts\/18781\/revisions"}],"predecessor-version":[{"id":18787,"href":"\/index.php?rest_route=\/wp\/v2\/posts\/18781\/revisions\/18787"}],"wp:attachment":[{"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=18781"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=18781"},{"taxonomy":"post_tag","embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=18781"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}