{"id":263,"date":"2025-11-19T06:24:58","date_gmt":"2025-11-19T14:24:58","guid":{"rendered":"https:\/\/scienceblog.com\/neuroedge\/?p=263"},"modified":"2025-11-19T06:24:58","modified_gmt":"2025-11-19T14:24:58","slug":"he-taught-ai-to-say-i-dont-know","status":"publish","type":"post","link":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/","title":{"rendered":"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;"},"content":{"rendered":"<p>Artificial intelligence can diagnose disease, write essays, and generate art. But it often refuses to admit when it&#8217;s wrong. Now, a University of Arizona astronomer has found a way to change that.<\/p>\n<p>In a preprint posted to <em>arXiv<\/em>, Peter Behroozi introduces a new method for reducing hallucinations in large-scale AI models by making them aware of their own uncertainty. Drawing on computer graphics techniques used in Pixar-style ray tracing and grounded in Bayesian mathematics, the approach allows neural networks with trillions of parameters to flag when their predictions might be unreliable, without requiring massive computational power. The research was supported by the National Science Foundation and includes public code for immediate reuse.<\/p>\n<h2>&#8220;Wrong but Confident&#8221; Is an AI Epidemic<\/h2>\n<blockquote><p>&#8220;There are many examples of neural networks &#8216;hallucinating,&#8217; or making up nonexistent facts, research papers and books to back up their incorrect conclusions,&#8221; Behroozi said. &#8220;This leads to real human suffering, including incorrect medical diagnoses, declined rental applications and facial recognition gone wrong.&#8221;<\/p><\/blockquote>\n<p>Trained as a cosmologist, Behroozi built the <em>Universe Machine<\/em> to simulate galaxy formation by testing billions of model variants. But as his models grew more complex, standard tools for quantifying uncertainty broke down. That\u2019s when a student\u2019s physics question about how light bends through Earth\u2019s atmosphere sparked a breakthrough.<\/p>\n<p>Behroozi realized he could adapt ray tracing\u2014a graphics method that calculates light\u2019s paths through 3D scenes\u2014for use in the billion-dimensional mathematical spaces that govern AI training. The resulting technique approximates the gold-standard Bayesian sampling method, which typically requires evaluating thousands of model variants to capture the full range of possible outcomes. His adaptation, described in detail across Figures 1\u20134 of the preprint, makes this feasible for even the largest language and vision models.<\/p>\n<p>Instead of trusting a single prediction, the method gathers results from a representative set of models. If their answers diverge, that signals a high-uncertainty region, essentially a warning light that the AI may not know what it\u2019s talking about. On standard benchmarks like CIFAR-10, ImageNet, and GLUE, the technique showed stronger calibration and better performance under distribution shifts than baseline models, all while remaining scalable.<\/p>\n<h2>Uncertainty Can Be a Feature, Not a Flaw<\/h2>\n<blockquote><p>&#8220;Suppose a doctor ordered a routine scan and decided that you needed to begin treatment for cancer immediately, even though you had no other symptoms,&#8221; Behroozi said. &#8220;Many people in this situation would seek a second opinion. The new method would have a similar effect: instead of the opinion of one AI doctor, it would give the range of plausible opinions.&#8221;<\/p><\/blockquote>\n<p>Trust in AI has faltered, in part because many systems deliver outputs with false precision. For scientists, the implications are stark. AI tools are now used to generate hypotheses, analyze images, even write sections of papers\u2014but they often do so with no sense of uncertainty. As Behroozi writes, this &#8220;undermines public trust in scientific output&#8221; and forces researchers to waste time on costly validations that a better-calibrated system might avoid.<\/p>\n<p>The broader promise lies in making AI more cautious when it should be. From criminal sentencing algorithms to autonomous vehicles, AI decisions increasingly affect real lives. Behroozi\u2019s method doesn\u2019t fix bias or eliminate errors, but it could give users a critical signal: when not to trust the machine.<\/p>\n<p>And for Behroozi\u2019s own research, it opens the door to more ambitious questions. Rather than generating simulations that vaguely resemble our universe, he now hopes to reconstruct the actual initial conditions that gave rise to the cosmic web. &#8220;What this technique allows us to do is figure out what were the initial conditions of the actual universe,&#8221; he said.<\/p>\n<p><a href=\"https:\/\/doi.org\/10.48550\/arXiv.2510.25824\">arXiv: 10.48550\/arXiv.2510.25824<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Artificial intelligence can diagnose disease, write essays, and generate art. But it often refuses to admit when it&#8217;s wrong. Now, a University of Arizona astronomer has found a way to change that. In a preprint posted to arXiv, Peter Behroozi introduces a new method for reducing hallucinations in large-scale AI models by making them aware &#8230; <a title=\"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;\" class=\"read-more\" href=\"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/\" aria-label=\"Read more about He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;\">Read more<\/a><\/p>\n","protected":false},"author":1297,"featured_media":264,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"_jetpack_newsletter_access":"","_jetpack_dont_email_post_to_subs":false,"_jetpack_newsletter_tier_id":0,"_jetpack_memberships_contains_paywalled_content":false,"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[4],"tags":[],"class_list":["post-263","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-computational-innovation","generate-columns","tablet-grid-50","mobile-grid-100","grid-parent","grid-50"],"yoast_head":"<!-- This site is optimized with the Yoast SEO Premium plugin v27.4 (Yoast SEO v27.4) - https:\/\/yoast.com\/product\/yoast-seo-premium-wordpress\/ -->\n<title>He Taught AI to Say &quot;I Don&#039;t Know&quot; - NeuroEdge<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"He Taught AI to Say &quot;I Don&#039;t Know&quot;\" \/>\n<meta property=\"og:description\" content=\"Artificial intelligence can diagnose disease, write essays, and generate art. But it often refuses to admit when it&#8217;s wrong. Now, a University of Arizona astronomer has found a way to change that. In a preprint posted to arXiv, Peter Behroozi introduces a new method for reducing hallucinations in large-scale AI models by making them aware ... Read more\" \/>\n<meta property=\"og:url\" content=\"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/\" \/>\n<meta property=\"og:site_name\" content=\"NeuroEdge\" \/>\n<meta property=\"article:published_time\" content=\"2025-11-19T14:24:58+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp\" \/>\n\t<meta property=\"og:image:width\" content=\"900\" \/>\n\t<meta property=\"og:image:height\" content=\"474\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/webp\" \/>\n<meta name=\"author\" content=\"NeuroEdge\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"NeuroEdge\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"3 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/\"},\"author\":{\"name\":\"NeuroEdge\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#\\\/schema\\\/person\\\/a13c664778e7eb97cb71e3e1ad356d2e\"},\"headline\":\"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;\",\"datePublished\":\"2025-11-19T14:24:58+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/\"},\"wordCount\":617,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/11\\\/AI-Behroozi-feature-image.png.webp\",\"articleSection\":[\"Computational Innovation\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#respond\"]}],\"copyrightYear\":\"2025\",\"copyrightHolder\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/#organization\"}},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/\",\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/\",\"name\":\"He Taught AI to Say \\\"I Don't Know\\\" - NeuroEdge\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/11\\\/AI-Behroozi-feature-image.png.webp\",\"datePublished\":\"2025-11-19T14:24:58+00:00\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#primaryimage\",\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/11\\\/AI-Behroozi-feature-image.png.webp\",\"contentUrl\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/11\\\/AI-Behroozi-feature-image.png.webp\",\"width\":900,\"height\":474,\"caption\":\"Light rays are propagating smoothly through a noisy, high-dimensional space in this artist\u2019s impression. The new ray tracing algorithm improves on previous methods by better averaging information over trajectories, making it many orders of magnitude faster for quantifying uncertainties in large neural networks.\"},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/2025\\\/11\\\/19\\\/he-taught-ai-to-say-i-dont-know\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#website\",\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/\",\"name\":\"NeuroEdge\",\"description\":\"A data-driven look at neuroscience and AI, for investors, policymakers, and innovators.\",\"publisher\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#organization\",\"name\":\"NeuroEdge\",\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/04\\\/cropped-neuroedge_logo.jpg\",\"contentUrl\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/wp-content\\\/uploads\\\/sites\\\/14\\\/2025\\\/04\\\/cropped-neuroedge_logo.jpg\",\"width\":955,\"height\":191,\"caption\":\"NeuroEdge\"},\"image\":{\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#\\\/schema\\\/logo\\\/image\\\/\"}},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/#\\\/schema\\\/person\\\/a13c664778e7eb97cb71e3e1ad356d2e\",\"name\":\"NeuroEdge\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g\",\"caption\":\"NeuroEdge\"},\"url\":\"https:\\\/\\\/scienceblog.com\\\/neuroedge\\\/author\\\/neuroedge\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO Premium plugin. -->","yoast_head_json":{"title":"He Taught AI to Say \"I Don't Know\" - NeuroEdge","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/","og_locale":"en_US","og_type":"article","og_title":"He Taught AI to Say \"I Don't Know\"","og_description":"Artificial intelligence can diagnose disease, write essays, and generate art. But it often refuses to admit when it&#8217;s wrong. Now, a University of Arizona astronomer has found a way to change that. In a preprint posted to arXiv, Peter Behroozi introduces a new method for reducing hallucinations in large-scale AI models by making them aware ... Read more","og_url":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/","og_site_name":"NeuroEdge","article_published_time":"2025-11-19T14:24:58+00:00","og_image":[{"width":900,"height":474,"url":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","type":"image\/webp"}],"author":"NeuroEdge","twitter_card":"summary_large_image","twitter_misc":{"Written by":"NeuroEdge","Est. reading time":"3 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#article","isPartOf":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/"},"author":{"name":"NeuroEdge","@id":"https:\/\/scienceblog.com\/neuroedge\/#\/schema\/person\/a13c664778e7eb97cb71e3e1ad356d2e"},"headline":"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;","datePublished":"2025-11-19T14:24:58+00:00","mainEntityOfPage":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/"},"wordCount":617,"commentCount":0,"publisher":{"@id":"https:\/\/scienceblog.com\/neuroedge\/#organization"},"image":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#primaryimage"},"thumbnailUrl":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","articleSection":["Computational Innovation"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#respond"]}],"copyrightYear":"2025","copyrightHolder":{"@id":"https:\/\/scienceblog.com\/#organization"}},{"@type":"WebPage","@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/","url":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/","name":"He Taught AI to Say \"I Don't Know\" - NeuroEdge","isPartOf":{"@id":"https:\/\/scienceblog.com\/neuroedge\/#website"},"primaryImageOfPage":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#primaryimage"},"image":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#primaryimage"},"thumbnailUrl":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","datePublished":"2025-11-19T14:24:58+00:00","breadcrumb":{"@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#primaryimage","url":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","contentUrl":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","width":900,"height":474,"caption":"Light rays are propagating smoothly through a noisy, high-dimensional space in this artist\u2019s impression. The new ray tracing algorithm improves on previous methods by better averaging information over trajectories, making it many orders of magnitude faster for quantifying uncertainties in large neural networks."},{"@type":"BreadcrumbList","@id":"https:\/\/scienceblog.com\/neuroedge\/2025\/11\/19\/he-taught-ai-to-say-i-dont-know\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/scienceblog.com\/neuroedge\/"},{"@type":"ListItem","position":2,"name":"He Taught AI to Say &#8220;I Don&#8217;t Know&#8221;"}]},{"@type":"WebSite","@id":"https:\/\/scienceblog.com\/neuroedge\/#website","url":"https:\/\/scienceblog.com\/neuroedge\/","name":"NeuroEdge","description":"A data-driven look at neuroscience and AI, for investors, policymakers, and innovators.","publisher":{"@id":"https:\/\/scienceblog.com\/neuroedge\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/scienceblog.com\/neuroedge\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/scienceblog.com\/neuroedge\/#organization","name":"NeuroEdge","url":"https:\/\/scienceblog.com\/neuroedge\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/scienceblog.com\/neuroedge\/#\/schema\/logo\/image\/","url":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/cropped-neuroedge_logo.jpg","contentUrl":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/cropped-neuroedge_logo.jpg","width":955,"height":191,"caption":"NeuroEdge"},"image":{"@id":"https:\/\/scienceblog.com\/neuroedge\/#\/schema\/logo\/image\/"}},{"@type":"Person","@id":"https:\/\/scienceblog.com\/neuroedge\/#\/schema\/person\/a13c664778e7eb97cb71e3e1ad356d2e","name":"NeuroEdge","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/28782ec992e8763e1f8d41ddc10864e7d8cd4cb99bacea6224c4abe634bbabec?s=96&d=mm&r=g","caption":"NeuroEdge"},"url":"https:\/\/scienceblog.com\/neuroedge\/author\/neuroedge\/"}]}},"jetpack_featured_media_url":"https:\/\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/11\/AI-Behroozi-feature-image.png.webp","jetpack_likes_enabled":true,"jetpack_sharing_enabled":true,"jetpack-related-posts":[{"id":94,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/04\/28\/quantum-neural-hybrid-solves-impossible-math\/","url_meta":{"origin":263,"position":0},"title":"Quantum-Neural Hybrid Solves Impossible Math","author":"NeuroEdge","date":"April 28, 2025","format":false,"excerpt":"The worlds of quantum mechanics and neural networks have collided in a new system that's setting benchmarks for solving previously intractable optimization problems. A multi-university team led by Shantanu Chakrabartty at Washington University in St. Louis has introduced NeuroSA, a neuromorphic architecture that leverages quantum tunneling mechanisms to reliably discover\u2026","rel":"","context":"In &quot;Physics&quot;","block_context":{"text":"Physics","link":"https:\/\/scienceblog.com\/neuroedge\/category\/physics\/"},"img":{"alt_text":"NeuroSA is a neuromorphic architecture with quantum capabilities that allow it to solve optimization problems more reliably than state-of-the-art methods. (Image created by Shantanu Chakrabartty using Google Gemini)","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/gemini-neuro-quantum.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/gemini-neuro-quantum.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/gemini-neuro-quantum.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/gemini-neuro-quantum.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]},{"id":22,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/04\/04\/ai-slashes-fluid-simulation-times-fifteenfold\/","url_meta":{"origin":263,"position":1},"title":"AI Slashes Fluid Simulation Times Fifteenfold","author":"NeuroEdge","date":"April 4, 2025","format":false,"excerpt":"Osaka researchers have developed an AI model that performs complex fluid simulations in minutes instead of hours, potentially transforming offshore engineering while maintaining high accuracy. This advancement could accelerate development cycles for maritime technologies and enable real-time monitoring systems previously considered computationally impossible. Traditional particle-based fluid simulations, essential for predicting\u2026","rel":"","context":"In &quot;Computational Innovation&quot;","block_context":{"text":"Computational Innovation","link":"https:\/\/scienceblog.com\/neuroedge\/category\/computational-innovation\/"},"img":{"alt_text":"Simulated waves","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/ocen-simulation.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":273,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/12\/01\/brain-like-ai-emerges-without-training-data-in-new-study\/","url_meta":{"origin":263,"position":2},"title":"Brain-Like AI Emerges Without Training Data in New Study","author":"NeuroEdge","date":"December 1, 2025","format":false,"excerpt":"Before these systems ever see a single cat photo or traffic sign, some AI models are already humming in tune with the visual cortex. In new work from Johns Hopkins University, scientists showed that carefully designed, biologically inspired architectures can mimic activity in human and primate visual brain areas even\u2026","rel":"","context":"In &quot;Brain Health&quot;","block_context":{"text":"Brain Health","link":"https:\/\/scienceblog.com\/neuroedge\/category\/brain-health\/"},"img":{"alt_text":"circuit-board-brain","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/circuit-board-brain.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/circuit-board-brain.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/circuit-board-brain.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/circuit-board-brain.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]},{"id":284,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/12\/29\/brain-model-discovers-neurons-that-reliably-predict-mistakes\/","url_meta":{"origin":263,"position":3},"title":"Brain Model Discovers Neurons That Reliably Predict Mistakes","author":"NeuroEdge","date":"December 29, 2025","format":false,"excerpt":"About 20 percent of neurons in a learning brain seem to be doing something counterintuitive. When these cells become more active, mistakes follow. A new computational model of the brain, built to mirror real neural circuits rather than optimize performance, stumbled onto this pattern while learning a simple visual task.\u2026","rel":"","context":"In &quot;Brain Health&quot;","block_context":{"text":"Brain Health","link":"https:\/\/scienceblog.com\/neuroedge\/category\/brain-health\/"},"img":{"alt_text":"neuron networks","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/Screenshot-2025-12-29-at-8.52.01-AM.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/Screenshot-2025-12-29-at-8.52.01-AM.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/Screenshot-2025-12-29-at-8.52.01-AM.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/12\/Screenshot-2025-12-29-at-8.52.01-AM.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]},{"id":90,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/04\/28\/brain-decoder-controls-spinal-cord-stimulation\/","url_meta":{"origin":263,"position":4},"title":"Brain decoder controls spinal cord stimulation","author":"NeuroEdge","date":"April 28, 2025","format":false,"excerpt":"In a significant development for spinal cord injury treatment, researchers at Washington University in St. Louis have created a neural decoder that bridges the communication gap between brain and spine, potentially opening new avenues for rehabilitation. The research team, led by Ismael Se\u00e1\u00f1ez, assistant professor of biomedical engineering at WashU,\u2026","rel":"","context":"In &quot;Computational Innovation&quot;","block_context":{"text":"Computational Innovation","link":"https:\/\/scienceblog.com\/neuroedge\/category\/computational-innovation\/"},"img":{"alt_text":"EEG Cap and walking man","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/eeg-cap.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/eeg-cap.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/eeg-cap.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/eeg-cap.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]},{"id":73,"url":"https:\/\/scienceblog.com\/neuroedge\/2025\/04\/24\/ai-fails-to-read-human-social-cues\/","url_meta":{"origin":263,"position":5},"title":"AI Fails To Read Human Social Cues","author":"NeuroEdge","date":"April 24, 2025","format":false,"excerpt":"Despite rapid advances in artificial intelligence, humans still maintain a significant edge when it comes to understanding social interactions, according to new research from Johns Hopkins University that reveals fundamental limitations in AI's ability to interpret human behavior. The study, presented at the International Conference on Learning Representations, found that\u2026","rel":"","context":"In &quot;Computational Innovation&quot;","block_context":{"text":"Computational Innovation","link":"https:\/\/scienceblog.com\/neuroedge\/category\/computational-innovation\/"},"img":{"alt_text":"A man covering his eyes in embarassment","src":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/man-379800_1280.jpg?resize=350%2C200&ssl=1","width":350,"height":200,"srcset":"https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/man-379800_1280.jpg?resize=350%2C200&ssl=1 1x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/man-379800_1280.jpg?resize=525%2C300&ssl=1 1.5x, https:\/\/i0.wp.com\/scienceblog.com\/neuroedge\/wp-content\/uploads\/sites\/14\/2025\/04\/man-379800_1280.jpg?resize=700%2C400&ssl=1 2x"},"classes":[]}],"_links":{"self":[{"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/posts\/263","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/users\/1297"}],"replies":[{"embeddable":true,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/comments?post=263"}],"version-history":[{"count":1,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/posts\/263\/revisions"}],"predecessor-version":[{"id":265,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/posts\/263\/revisions\/265"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/media\/264"}],"wp:attachment":[{"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/media?parent=263"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/categories?post=263"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scienceblog.com\/neuroedge\/wp-json\/wp\/v2\/tags?post=263"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}