{"id":7684,"date":"2025-10-14T16:59:28","date_gmt":"2025-10-14T07:59:28","guid":{"rendered":"https:\/\/www.jaif.or.jp\/en\/?p=7684"},"modified":"2025-10-14T17:02:18","modified_gmt":"2025-10-14T08:02:18","slug":"the-limits-of-ai-and-its-potential-applications-in-the-nuclear-industry-lecture-by-prof-washio-takashi-kansai-university","status":"publish","type":"post","link":"https:\/\/www.jaif.or.jp\/en\/news\/regulation-safety\/7684","title":{"rendered":"The Limits of AI and Its Potential Applications in the Nuclear Industry \u2014 Lecture by Prof. WASHIO Takashi (Kansai University)"},"content":{"rendered":"<p class=\"p1\">At the outset, he outlined key approaches such as machine learning and deep learning, emphasizing that <i>\u201cAI demonstrates excellent interpolation ability in data-rich domains, but it cannot make correct judgments when faced with unknown conditions.\u201d<\/i> He continued, <i>\u201cAI can find optimal solutions within given data, but it is powerless when data do not exist. Therefore, we should not aim for full automation through AI\u2014it should be positioned as a collaborative tool that humans must monitor and evaluate.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">Turning to generative AI exemplified by ChatGPT, Prof. Washio explained its underlying mechanisms: <i>\u201cGenerative AI is essentially a massive probabilistic model\u2014it does not engage in creative thinking like humans.\u201d<\/i> He cautioned that <i>\u201calthough such systems can generate text that sounds plausible, they may produce errors when applied to unfamiliar contexts.\u201d<\/i> Furthermore, he warned, <i>\u201cThe \u2018answers\u2019 produced by AI are merely extensions of the most statistically likely word sequences. Unless users understand this nature, misuse could undermine safety culture itself.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">As an example of practical application, Prof. Washio introduced his joint research with the National Institute of Advanced Industrial Science and Technology (AIST) and NEC Corporation on analyzing <i>stray light<\/i> in space telescopes. Using an AI algorithm that automatically explores risk conditions, the project identified hazardous scenarios 100,000 times more efficiently than random searches. He noted, <i>\u201cThis methodology could be applied to automatically extract unanticipated accident scenarios in nuclear power plants.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">He further described collaborations with Osaka University, where AI optimized chemical reaction conditions and derived high-yield results from limited experimental data. In another project with Nissan Motor Co., AI analyzed plant operation data to automatically adjust simulation models, improving their consistency with real-world operations. Prof. Washio stated, <i>\u201cProcess optimization and high-precision operational planning through AI can contribute to the safe operation of nuclear facilities.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">During the Q&amp;A session, a Swedish researcher asked whether AI could eventually <i>\u201cinherit human competence.\u201d<\/i> Prof. Washio replied clearly: <i>\u201cAI can handle knowledge and data, but it cannot replicate human judgment or insight.\u201d<\/i> He added, <i>\u201cThe key issue is how we design AI outputs and connect them to decision-making in human society. This depends not only on technology but also on organizational systems, social institutions, and human dialogue.\u201d<\/i> Another participant suggested that <i>\u201cAI might become an entity that nurtures competence through education or meetings,\u201d<\/i> to which Prof. Washio responded that such discussion <i>\u201cbelongs to future philosophical and ethical debates.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">An engineer from the United States raised a practical question about the difficulty of reading data from obsolete media such as floppy disks and CD-ROMs. Prof. Washio stressed that <i>\u201cAI and database maintenance should not be left solely to corporate self-responsibility. In the future, public management by governments will become necessary.\u201d<\/i> He concluded that <i>\u201cinformation and AI models should be preserved as part of social infrastructure.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">Finally, when asked whether AI can explore the unknown, Prof. Washio answered, <i>\u201cThe fundamental limitation of AI lies in its inability to quantify the unknown.\u201d<\/i> He explained, <i>\u201cAI can assist in discovering the unknown, but it cannot create it on its own. That is precisely why combining human scientific intuition with AI\u2019s analytical power is essential.\u201d<\/i><i><\/i><\/p>\n<p class=\"p1\">In closing, Prof. Washio remarked, <i>\u201cThe nuclear industry tends to be overly cautious about introducing new technologies due to its emphasis on safety. Yet to enhance safety, we must pursue innovation with a \u2018conservatively exploratory\u2019 mindset.\u201d<\/i> He concluded, <i>\u201cBy understanding the limitations of AI and integrating its strengths with human judgment, we can build the foundation for the next generation of safety culture.\u201d<\/i> The lecture ended with resounding applause from the audience.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>At the outset, he outlined key approaches such as machine learning and deep learning, emphasizing that \u201cAI demonstrates excellent interpolation ability in data-rich domains, but it cannot make correct judgments when faced with unknown conditions.\u201d He continued, \u201cAI can find optimal solutions within given data, but it is powerless when data do not exist. Therefore, [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":7686,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[9,1250,7],"tags":[1361,1404,1403,1401,1400,1393,1402,498,1325,1405,516,497,967,103,1399],"class_list":["post-7684","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-industry","category-rd","category-regulation-safety","tag-ai","tag-aist","tag-chatgpt","tag-generative-ai","tag-kansai-university","tag-knowledge-management","tag-machine-learning","tag-nea","tag-nec","tag-nissan","tag-nuclear-industry","tag-oecd","tag-oecd-nea","tag-safety-culture","tag-washio-takashi"],"acf":[],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/posts\/7684","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/comments?post=7684"}],"version-history":[{"count":1,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/posts\/7684\/revisions"}],"predecessor-version":[{"id":7685,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/posts\/7684\/revisions\/7685"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/media\/7686"}],"wp:attachment":[{"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/media?parent=7684"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/categories?post=7684"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.jaif.or.jp\/en\/wp-json\/wp\/v2\/tags?post=7684"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}