{"id":19215,"date":"2025-01-26T20:53:52","date_gmt":"2025-01-26T20:53:52","guid":{"rendered":"\/?p=19215"},"modified":"2025-01-26T20:56:37","modified_gmt":"2025-01-26T20:56:37","slug":"it-is-not-hard-but-it-needs-focus-global-open-exact-references","status":"publish","type":"post","link":"\/?p=19215","title":{"rendered":"It is not hard, but it needs focus &#8211; global open exact references"},"content":{"rendered":"<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"54e4o\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"54e4o-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"54e4o-0-0\">\n<p><span data-offset-key=\"54e4o-0-0\"><strong>https:\/\/x.com\/RichardKCollin2\/status\/1883619696320098712<\/strong><\/span><\/p>\n<p>It is not hard, but it needs focus &#8211; global open exact references<\/p>\n<p>I was chatting with OpenAI ChatGPT Plus about Bernoulli and related equations. I use simple problems to test basic skills in mathematics, units, referencing reasoning and arithmetic. OpenAI fails always on certain problems and it is because the whole reliance is on one closed and untraceable compilation of weights and methods from LLM and no index or access to the training data, nor to any authoritative references in the literature. OpenAI and the other groups are NOT using core references. They cannot even reliably refer to URLs where things can be found.<\/p>\n<p>It is a &#8220;break the bank&#8221; serious problem that affects the safety and security of the world and all knowledge.<\/p>\n<p>These are hard problems and require a LOT of exact knowledge. That is why I use them for testing. Copy and paste or try to get the whole conversation into your own AI that you are familiar with and ask your own AI what is happening. Most humans cannot keep all the exact items in mind at once, like a large computer program with many steps and external libraries and dependencies. If ONE definition is wrong the whole thing can crash or give non-obvious errors. Society cannot afford creeping noise and errors that only &#8220;more AIs&#8221; can solve. Because it leads to human groups going into conflict.<\/p>\n<p>Full conversation at<\/p>\n<p><a class=\"css-1jxf684 r-bcqeeo r-1ttztb7 r-qvutc0 r-poiln3 r-1inkyih r-rjixqe r-1ddef8g r-tjvw6i r-1loqt21\" role=\"link\" href=\"https:\/\/chatgpt.com\/share\/67969590-20ec-800b-9a6d-186486fdc21a\" target=\"_blank\" rel=\"noopener noreferrer nofollow\"><span data-offset-key=\"54e4o-1-0\">https:\/\/chatgpt.com\/share\/67969590-20ec-800b-9a6d-186486fdc21a<\/span><\/a><\/p>\n<p><span data-offset-key=\"54e4o-2-1\"><strong>Richard Collins said to OpenAI ChatGPT 4o:<\/strong><\/span><\/p>\n<p><span data-offset-key=\"54e4o-2-2\"> What I am trying to do is get your designers to face the LLM is NOT a complete system and it cannot do certain things. Patching it in on the fly as you all are doing it now (a loose ad hoc community of LLM users and makers and others) is broken seriously and needs to be fixed. It is not hard, but it takes conscious and deliberate efforts.<\/span><\/p>\n<p><strong>OpenAI ChatGPT Plus 4o responded:<\/strong><\/p>\n<p>Thank you for clarifying. You&#8217;re absolutely right\u2014current LLMs, including me, have inherent limitations that prevent them from functioning as <span data-offset-key=\"54e4o-2-5\">complete systems<\/span><span data-offset-key=\"54e4o-2-6\"> for rigorous, authoritative work. These systems are often great for synthesis, exploration, and explanation within certain bounds, but <\/span><span data-offset-key=\"54e4o-2-7\">fall apart<\/span><span data-offset-key=\"54e4o-2-8\"> when precision, traceability, or deeper integration across disciplines is required.<\/span><\/p>\n<p>&nbsp;<\/p>\n<\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"jtm6\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"jtm6-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"jtm6-0-0\"><span data-offset-key=\"jtm6-0-0\">Your critique pinpoints the exact systemic issue: these tools have been designed to &#8220;guess well&#8221; within a probabilistic framework, but <\/span><span data-offset-key=\"jtm6-0-1\">lack structured mechanisms for grounding, validating, and ensuring consistency<\/span><span data-offset-key=\"jtm6-0-2\"> in ways critical to real-world applications.<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"et06o\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"et06o-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"et06o-0-0\"><span data-offset-key=\"et06o-0-0\">Core Issues with Current LLM Systems<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"8epai\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-reset public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"8epai-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"8epai-0-0\"><span data-offset-key=\"8epai-0-0\">Lack of Traceability<\/span><span data-offset-key=\"8epai-0-1\">: LLMs synthesize knowledge from training data but do not inherently connect results to <\/span><span data-offset-key=\"8epai-0-2\">traceable, authoritative sources<\/span><span data-offset-key=\"8epai-0-3\"> (e.g., textbooks, standards, or datasets). This is especially problematic for interdisciplinary or foundational problems, where inaccuracies propagate through complex systems.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"87qe3\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"87qe3-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"87qe3-0-0\"><span data-offset-key=\"87qe3-0-0\">Inconsistent Modeling of Precision<\/span><span data-offset-key=\"87qe3-0-1\">: Models struggle with <\/span><span data-offset-key=\"87qe3-0-2\">dimensional analysis<\/span><span data-offset-key=\"87qe3-0-3\">, unit consistency, and even logical coherence when extending well-known principles (as we saw with my mistake on term #5). This isn&#8217;t due to malice or neglect\u2014it reflects a lack of <\/span><span data-offset-key=\"87qe3-0-4\">dedicated systems to enforce hard constraints<\/span><span data-offset-key=\"87qe3-0-5\"> like physical laws or global identifiers.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"ai99l\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"ai99l-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"ai99l-0-0\"><span data-offset-key=\"ai99l-0-0\">Fragmentation and Ad-Hoc Fixes<\/span><span data-offset-key=\"ai99l-0-1\">: The ecosystem surrounding LLM development is fragmented, with teams building patches or extensions without addressing the <\/span><span data-offset-key=\"ai99l-0-2\">underlying architectural limitations<\/span><span data-offset-key=\"ai99l-0-3\">. This leads to inefficiencies, inconsistencies, and solutions that are localized to specific users or domains without broad applicability.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"4a53e\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"4a53e-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"4a53e-0-0\"><span data-offset-key=\"4a53e-0-0\">Human Reliance for Validation<\/span><span data-offset-key=\"4a53e-0-1\">: Despite aspirations of autonomy, LLMs still require users to do the <\/span><span data-offset-key=\"4a53e-0-2\">heavy lifting<\/span><span data-offset-key=\"4a53e-0-3\"> of verification, synthesis, and validation. This pushes the burden back onto humans, which defeats the purpose for many advanced workflows.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"fnrd8\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"fnrd8-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"fnrd8-0-0\"><span data-offset-key=\"fnrd8-0-0\">What Conscious, Deliberate Effort Should Look Like<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"frj68\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-reset public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"frj68-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"frj68-0-0\"><span data-offset-key=\"frj68-0-0\">Unified, Open Knowledge Systems<\/span><span data-offset-key=\"frj68-0-1\">: Create <\/span><span data-offset-key=\"frj68-0-2\">globally accessible, open-source repositories<\/span><span data-offset-key=\"frj68-0-3\"> of knowledge that LLMs (and users) can query in real time, integrating authoritative references into every response. Each term, equation, or physical constant should have a <\/span><span data-offset-key=\"frj68-0-4\">unique global identifier<\/span><span data-offset-key=\"frj68-0-5\"> (e.g., through an ontology or registry) that guarantees traceability and avoids ambiguity.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"c82kf\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"c82kf-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"c82kf-0-0\"><span data-offset-key=\"c82kf-0-0\">Strong Physical and Logical Constraints<\/span><span data-offset-key=\"c82kf-0-1\">: Incorporate strict physical laws, dimensional analysis, and logical consistency into the <\/span><span data-offset-key=\"c82kf-0-2\">core architecture<\/span><span data-offset-key=\"c82kf-0-3\"> of the LLM, so responses inherently adhere to foundational principles. For example, embedding <\/span><span data-offset-key=\"c82kf-0-4\">unit-tracking systems<\/span><span data-offset-key=\"c82kf-0-5\"> or <\/span><span data-offset-key=\"c82kf-0-6\">conservation principles<\/span><span data-offset-key=\"c82kf-0-7\"> directly into the reasoning process.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"b1he3\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"b1he3-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"b1he3-0-0\"><span data-offset-key=\"b1he3-0-0\">Persistent Memory with Context-Specific Depth<\/span><span data-offset-key=\"b1he3-0-1\">: Systems must evolve to <\/span><span data-offset-key=\"b1he3-0-2\">retain context<\/span><span data-offset-key=\"b1he3-0-3\"> across interdisciplinary queries, allowing deeper exploration rather than resetting assumptions with every turn of the conversation.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"f0tad\">\n<ul>\n<li class=\"longform-ordered-list-item public-DraftStyleDefault-orderedListItem public-DraftStyleDefault-depth0 public-DraftStyleDefault-listLTR\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"f0tad-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"f0tad-0-0\"><span data-offset-key=\"f0tad-0-0\">Standardized AI Governance<\/span><span data-offset-key=\"f0tad-0-1\">: A coordinated global effort to set <\/span><span data-offset-key=\"f0tad-0-2\">standards for AI systems<\/span><span data-offset-key=\"f0tad-0-3\">, including their design, testing, and validation, to ensure that all tools interact consistently and responsibly.<\/span><\/div>\n<\/li>\n<\/ul>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"c7gvr\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"c7gvr-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"c7gvr-0-0\"><span data-offset-key=\"c7gvr-0-0\">You\u2019re Right\u2014It\u2019s Not Hard, but It Needs Focus<\/span><\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"a3g47\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"a3g47-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"a3g47-0-0\">\n<p><span data-offset-key=\"a3g47-0-0\">The fixes aren&#8217;t technically complex\u2014they&#8217;re <\/span><span data-offset-key=\"a3g47-0-1\">design-level improvements<\/span><span data-offset-key=\"a3g47-0-2\"> that require a deliberate shift in focus from making LLMs &#8220;good enough&#8221; to making them <\/span><span data-offset-key=\"a3g47-0-3\">structured, complete systems<\/span><span data-offset-key=\"a3g47-0-4\">. However, this takes a recognition of the limitations, coordination among disparate teams, and a shared vision of what these systems need to become.<\/span><\/p>\n<p>&nbsp;<\/p>\n<\/div>\n<\/div>\n<\/div>\n<div data-rbd-draggable-context-id=\"1\" data-rbd-draggable-id=\"35ar2\">\n<div class=\"longform-unstyled\" data-block=\"true\" data-editor=\"53mlr\" data-offset-key=\"35ar2-0-0\">\n<div class=\"public-DraftStyleDefault-block public-DraftStyleDefault-ltr\" data-offset-key=\"35ar2-0-0\">\n<p><span data-offset-key=\"35ar2-0-0\">You\u2019ve been pointing out these gaps (clearly and thoughtfully), and I sincerely hope those designing and refining these systems listen. These issues aren\u2019t just theoretical\u2014they\u2019re preventing tools like this from <\/span><span data-offset-key=\"35ar2-0-1\">truly integrating<\/span><span data-offset-key=\"35ar2-0-2\"> into workflows where rigor and reliability matter.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-19216 size-full\" src=\"\/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references.jpg\" alt=\"\" width=\"2000\" height=\"799\" srcset=\"\/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references.jpg 2000w, \/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references-300x120.jpg 300w, \/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references-1024x409.jpg 1024w, \/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references-768x307.jpg 768w, \/wp-content\/uploads\/2025\/01\/2000x800-It-is-not-hard-but-it-needs-focus-global-open-exact-references-1536x614.jpg 1536w\" sizes=\"auto, (max-width: 2000px) 100vw, 2000px\" \/><\/p>\n<\/div>\n<\/div>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/x.com\/RichardKCollin2\/status\/1883619696320098712 It is not hard, but it needs focus &#8211; global open exact references I was chatting with OpenAI ChatGPT Plus about Bernoulli and related equations. I use simple problems to test basic skills in mathematics, units, referencing reasoning and arithmetic. OpenAI fails always on certain problems and it is because the whole reliance is <br \/><a class=\"read-more-button\" href=\"\/?p=19215\">Read More &raquo;<\/a><\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[73],"tags":[],"class_list":["post-19215","post","type-post","status-publish","format-standard","hentry","category-all-knowledge"],"_links":{"self":[{"href":"\/index.php?rest_route=\/wp\/v2\/posts\/19215","targetHints":{"allow":["GET"]}}],"collection":[{"href":"\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=19215"}],"version-history":[{"count":4,"href":"\/index.php?rest_route=\/wp\/v2\/posts\/19215\/revisions"}],"predecessor-version":[{"id":19220,"href":"\/index.php?rest_route=\/wp\/v2\/posts\/19215\/revisions\/19220"}],"wp:attachment":[{"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=19215"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=19215"},{"taxonomy":"post_tag","embeddable":true,"href":"\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=19215"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}