<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	
	xmlns:georss="http://www.georss.org/georss"
	xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#"
	>

<channel>
	<title>LLMs - Medika Life</title>
	<atom:link href="https://medika.life/tag/llms/feed/" rel="self" type="application/rss+xml" />
	<link>https://medika.life/tag/llms/</link>
	<description>Make Informed decisions about your Health</description>
	<lastBuildDate>Fri, 06 Mar 2026 20:02:55 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.5.5</generator>

<image>
	<url>https://i0.wp.com/medika.life/wp-content/uploads/2021/01/medika.png?fit=32%2C32&#038;ssl=1</url>
	<title>LLMs - Medika Life</title>
	<link>https://medika.life/tag/llms/</link>
	<width>32</width>
	<height>32</height>
</image> 
<site xmlns="com-wordpress:feed-additions:1">180099625</site>	<item>
		<title>From AI Excitement to Execution: Why Health Leaders Must Now Master the “How”</title>
		<link>https://medika.life/from-ai-excitement-to-execution-why-health-leaders-must-now-master-the-how/</link>
		
		<dc:creator><![CDATA[Gil Bashe, Medika Life Editor]]></dc:creator>
		<pubDate>Fri, 06 Mar 2026 20:02:51 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Diseases]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[Ethics in Practice]]></category>
		<category><![CDATA[For Doctors]]></category>
		<category><![CDATA[General Health]]></category>
		<category><![CDATA[Policy and Practice]]></category>
		<category><![CDATA[Public Health]]></category>
		<category><![CDATA[Trending Issues]]></category>
		<category><![CDATA[ChatGPT]]></category>
		<category><![CDATA[Clalit Health Services]]></category>
		<category><![CDATA[Ethics]]></category>
		<category><![CDATA[Governance]]></category>
		<category><![CDATA[Hal Wolf]]></category>
		<category><![CDATA[HIMSS]]></category>
		<category><![CDATA[HIMSS 2026]]></category>
		<category><![CDATA[Isaac Kohane]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[OpenAI]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21616</guid>

					<description><![CDATA[<p>Artificial intelligence is advancing in health care faster than almost any other technology in modern medical history. According to research from McKinsey &#38; Company, artificial intelligence could generate as much as $100 billion annually across healthcare systems worldwide, through improved clinical decision support and workflow efficiency, as well as advances in drug development and population [&#8230;]</p>
<p>The post <a href="https://medika.life/from-ai-excitement-to-execution-why-health-leaders-must-now-master-the-how/">From AI Excitement to Execution: Why Health Leaders Must Now Master the “How”</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Artificial intelligence is advancing in health care faster than almost any other technology in modern medical history. According to research from <a href="https://www.mckinsey.com/industries/life-sciences/our-insights/generative-ai-in-the-pharmaceutical-industry-moving-from-hype-to-reality">McKinsey &amp; Company, artificial intelligence could generate as much as $100 billion annually across healthcare systems worldwide</a>, through improved clinical decision support and workflow efficiency, as well as advances in drug development and population health analytics. The promise is extraordinary, and the pace of implementation shows little sign of slowing.</p>



<p>History, however, offers a useful caution. Breakthrough technologies in medicine rarely achieve their full potential simply because they exist. Their real impact depends on whether the institutions responsible for health-care delivery know how to adopt them wisely, integrate them responsibly and align them with their mission to improve patient health.</p>



<p>Artificial intelligence now stands at that same threshold. The industry has moved beyond fascination with what algorithms can do and entered a more demanding phase: determining how these tools should be evaluated, governed, and integrated into the environments where care is delivered. At the same time, some health professionals are turning to AI – not to augment their knowledge – but assuming the information is patient-care ready.</p>



<p>Across the health ecosystem, leaders are discovering that the most important questions about artificial intelligence are not technological. They are organizational, ethical and operational. Which AI systems genuinely improve clinical decision-making? Which tools strengthen the efficiency of hospitals and health systems? Which innovations introduce complexity without delivering measurable benefit?</p>



<p>Answering those questions requires a perspective that bridges policy leadership, real-world care delivery, and the scientific foundations of biomedical informatics. That convergence of experience sits at the center of a “Views From the Top” mainstage discussion at the <a href="https://www.himssconference.com/register/?utm_source=google&amp;utm_medium=cpc&amp;utm_campaign=US-EN-GA-BRD-PHA-Search-HIMSS26-Core&amp;gad_source=1&amp;gad_campaignid=23028140300&amp;gbraid=0AAAAA9RcRS5VnIvOREOV_e8P__ck9VjTR&amp;gclid=Cj0KCQiAk6rNBhCxARIsAN5mQLtutruWd-5p1Wn2AwXHxy1v-Qi3oN1ADdz2MjA78q5H_4qD6RWCwNIaAoAHEALw_wcB">HIMSS Global Health Conference &amp; Exhibition</a>, where some 35,000 leaders whose work spans the global health ecosystem will examine how organizations can recognize the true value proposition of artificial intelligence applications before embedding them into health-care systems.</p>



<p>The perspectives shaping this discussion reflect three essential dimensions of responsible artificial intelligence in health: governance frameworks that guide innovation, operational insights from large-scale health care delivery, and scientific rigor grounded in biomedical informatics. Together, these vantage points illuminate the path from technological promise to practical value.</p>



<h2 class="wp-block-heading"><strong>Governing Innovation in a Rapidly Changing Health Ecosystem</strong></h2>



<p>Digital transformation in health rarely succeeds simply because technology exists. It succeeds when organizations develop leadership frameworks capable of evaluating innovation, managing risk and aligning new tools with patient-centered goals.</p>



<p>Few leaders have observed the evolution of digital health across as many national systems and institutional environments as <a href="https://iowa.himss.org/resource-bio/harold-f-wolf-iii">Hal Wolf, president and chief executive officer of HIMSS</a>, <a href="https://en.wikipedia.org/wiki/Ran_Balicer">Ran Balicer, MD, PhD, chief innovation officer of Clalit Health Services</a> and <a href="https://dbmi.hms.harvard.edu/people/isaac-kohane">Isaac Kohane, MD, PhD, chair of biomedical informatics at Harvard Medical School</a>. The three will step onto the mainstage at HIMSS to share their “View from the Top” in a session titled: <a href="https://app.himssconference.com/event/himss-2026/planning/UGxhbm5pbmdfNDMyNzU3NA==">“Recognizing the &#8216;Value Proposition&#8217; Criteria While Selecting AI Applications</a>.”</p>



<figure class="wp-block-image size-large"><img fetchpriority="high" decoding="async" width="696" height="392" src="https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=696%2C392&#038;ssl=1" alt="" class="wp-image-21617" srcset="https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=1024%2C576&amp;ssl=1 1024w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=300%2C169&amp;ssl=1 300w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=768%2C432&amp;ssl=1 768w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=1536%2C864&amp;ssl=1 1536w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=150%2C84&amp;ssl=1 150w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=696%2C392&amp;ssl=1 696w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?resize=1068%2C601&amp;ssl=1 1068w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?w=1920&amp;ssl=1 1920w, https://i0.wp.com/medika.life/wp-content/uploads/2026/03/116-H26-VFTT-Social-Graphic.png?w=1392&amp;ssl=1 1392w" sizes="(max-width: 696px) 100vw, 696px" data-recalc-dims="1" /><figcaption class="wp-element-caption">Image provided by HIMSS</figcaption></figure>



<p>Through his work with global government health ministries, hospital networks, and technology innovators worldwide, Wolf has consistently emphasized that technological progress must be anchored in governance and trust.</p>



<p><em>“Digital health transformation is not about technology alone. It is about leadership, governance, and the trust that allows innovation to improve care,”</em> Wolf has said in discussions about global digital health transformation.</p>



<p>Artificial intelligence intensifies this leadership challenge because its influence extends far beyond traditional clinical tools. AI systems increasingly operate across multiple layers of healthcare delivery. Some applications assist clinicians by analyzing medical data or suggesting treatment options. Others function within hospitals&#8217; and health systems&#8217; operational infrastructure, helping manage patient flow, prioritize diagnostic reviews, and allocate scarce resources.</p>



<p>These operational algorithms rarely capture headlines; however, &nbsp;they shape the environment in which health care is delivered. Decisions about which cases are reviewed first, how clinicians allocate their attention, and how health systems manage capacity can profoundly influence patient outcomes.</p>



<p>For leaders responsible for health systems, artificial intelligence cannot be treated as simply another technological upgrade. It must be evaluated through governance structures capable of understanding how algorithms function, what assumptions shape their recommendations, and how their use aligns with institutional priorities.</p>



<p>Without that oversight, innovation risks amplifying complexity rather than improving care. Instead of informing, it can spread misinformation.</p>



<h2 class="wp-block-heading"><strong>Aligning Artificial Intelligence With the Values of Medicine</strong></h2>



<p>Governance provides the policy foundation for responsible adoption of artificial intelligence, but real-world implementation reveals a second challenge: ensuring that AI systems operate effectively within healthcare delivery itself.</p>



<p>Large population health systems increasingly use advanced analytics to anticipate risk, manage chronic disease, and allocate clinical resources across diverse communities. Within these environments, artificial intelligence is no longer a theoretical innovation. It is already influencing how health organizations prioritize patients, coordinate care and deploy limited resources.</p>



<p>That operational perspective is central to Ran Balicer, MD, PhD, of <a href="https://www.clalit-innovation.org/clalitresearchinstitute">Clalit Health Services</a>, one of the world’s most advanced data-driven health systems. The Clalit integrated infrastructure connects hospitals, clinics, and community health programs through longitudinal datasets that support predictive analytics at the national scale.</p>



<p>Experience within such systems reinforces an important insight: artificial intelligence models do not function independently of human judgment. They reflect priorities embedded in their design and the assumptions guiding their deployment.</p>



<p><em>“Algorithms are opinions embedded in code,”</em> Balicer has observed in discussions about the role of artificial intelligence in population health.</p>



<p>In practice, this means that AI systems interpret clinical data through frameworks shaped by human choices. The way a model defines risk, prioritizes cases, or recommends interventions reflects decisions about what matters most within a healthcare environment.</p>



<p>Those decisions carry ethical implications. When artificial intelligence helps determine which patients receive immediate attention or which cases are escalated for further review, transparency about how algorithms function becomes essential to maintaining trust among clinicians and patients alike. The scientific frontier of health-care AI reinforces that concern.</p>



<p>Isaac Kohane, MD, PhD, who has also served as a co-author of the <em>Institute of Medicine Report on Precision Medicine</em>, which has served as the template for national efforts, has spent decades exploring how machine learning can advance medicine while preserving the judgment that defines clinical practice. His research emphasizes that artificial intelligence in healthcare must align with the ethical traditions and professional responsibilities of medicine.</p>



<p><em>“AI systems in medicine must ultimately reflect the values of the profession they serve,”</em> Kohane has written in discussions about AI alignment in biomedical informatics.</p>



<p>This perspective highlights a crucial distinction between technological capability and clinical responsibility. Many AI models entering healthcare environments were originally designed for broader computational tasks rather than the nuanced realities of patient care. Medicine operates within a landscape shaped by uncertainty, empathy, and accountability, and technologies introduced into that environment must reflect those values.</p>



<p>Ensuring that artificial intelligence aligns with the principles guiding health-care delivery, therefore, represents one of the most important scientific and ethical challenges facing the future of health.</p>



<h2 class="wp-block-heading"><strong>The Discipline Required to Make Innovation Matter</strong></h2>



<p>The health sector has experienced waves of technological enthusiasm before. Electronic health records promised seamless information exchange, but then introduced administrative burdens on health professionals when implemented without thoughtful workflow design. Data analytics promised unprecedented insight, but sometimes led to fragmentation when systems failed to communicate across institutions.</p>



<p>Artificial intelligence now stands at a similar moment in the evolution of health technology.</p>



<p>Its capabilities in supporting decision-making flow are extraordinary, yet realizing them will require disciplined leadership to evaluate, integrate and govern AI tools within health-care delivery systems. Health leaders must learn to ask deeper questions before embracing the next algorithmic breakthrough. What problem does this system truly solve? How does it strengthen clinical practice? What assumptions guide its recommendations? How does its use advance the mission of improving patient health?</p>



<p>These questions move the conversation beyond technological novelty toward operational practicality. It’s among the many reasons these three global leaders step to the HIMSS stage together.</p>



<p>Artificial intelligence will undoubtedly reshape the health ecosystem in the years ahead. Its long-term impact, however, will not be determined solely by the sophistication of algorithms or the speed of technological progress. Along with how to leverage AI, ChatGPT and LLMs, users require heightened cognitive awareness.</p>



<p>It will be determined by whether the health community develops the discipline and ability required to translate innovation into systems that strengthen care, support clinicians and improve the health of the populations they serve.</p>



<p>The real story of artificial intelligence in health is no longer about what machines can do. It is about how wisely the health sector chooses to use them.</p>
<p>The post <a href="https://medika.life/from-ai-excitement-to-execution-why-health-leaders-must-now-master-the-how/">From AI Excitement to Execution: Why Health Leaders Must Now Master the “How”</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21616</post-id>	</item>
		<item>
		<title>The Shift from Pure Modernity to Human-Centered Modernity</title>
		<link>https://medika.life/the-shift-from-pure-modernity-to-human-centered-modernity/</link>
		
		<dc:creator><![CDATA[Atefeh Ferdosipour]]></dc:creator>
		<pubDate>Fri, 06 Mar 2026 19:52:14 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Atefeh Ferdosipour]]></category>
		<category><![CDATA[education]]></category>
		<category><![CDATA[Human-Centered Artificial Intelligence]]></category>
		<category><![CDATA[Learning Sciences]]></category>
		<category><![CDATA[LLMs]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21613</guid>

					<description><![CDATA[<p>Throughout the history of science, it has rarely been the case that any phenomenon has remained permanent and unchanging. Theories, approaches, research methods, philosophies, and everything related to scientific perspectives have continually evolved. These changes have been adaptive and have moved toward improving human living conditions. If science is meant to serve humanity, it follows [&#8230;]</p>
<p>The post <a href="https://medika.life/the-shift-from-pure-modernity-to-human-centered-modernity/">The Shift from Pure Modernity to Human-Centered Modernity</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Throughout the history of science, it has rarely been the case that any phenomenon has remained permanent and unchanging. Theories, approaches, research methods, philosophies, and everything related to scientific perspectives have continually evolved. These changes have been adaptive and have moved toward improving human living conditions. If science is meant to serve humanity, it follows that whenever a tool fails—for whatever reason—to fulfill this responsibility effectively, it must either change or, over time and under changing circumstances, be updated into a more efficient version.</p>



<p>But from the perspective of philosophers of science, when do such shifts in scientific approaches actually occur?</p>



<h2 class="wp-block-heading"><em><strong>Thomas Kuhn’s Perspective</strong></em></h2>



<p>Kuhn believed that changes in scientific approaches resemble political revolutions. Simply put, when a government can no longer manage society or effectively administer its affairs, dissatisfaction gradually spreads among the public and opposition begins to form. In other words, the inability to respond to society’s needs becomes the driving force behind revolutionary movements. This process continues until a capable system emerges that can meet those needs, eventually leading to the establishment of a new order.</p>



<p>A similar process occurs in what Kuhn calls scientific revolutions. According to him, in every era the majority of scientists accept and follow a general framework. Kuhn refers to this dominant framework — which contains a collection of theories and practical models — as a paradigm. Paradigms are patterns widely followed by scholars, such as the paradigm of modernity or the paradigm of cognitive science.</p>



<p>As long as these paradigms remain aligned with the requirements of life and are capable of addressing existing problems, they continue to be valued and are used in major policy frameworks. However, when a dominant paradigm fails to respond to contemporary challenges and the solutions derived from it prove ineffective at addressing large-scale needs, doubts arise about its continued relevance. Under such circumstances, dissatisfaction intensifies to the point that scholars begin to consider laying the groundwork for a new, updated paradigm.</p>



<p>In his book The Structure of Scientific Revolutions, Kuhn emphasizes that scientific transformations are not linear or step-by-step processes. Rather, they are complex and revolutionary developments in which social and historical factors play a crucial role. Under normal conditions, scientists operate within the framework of an accepted paradigm — what Kuhn calls normal science. However, when persistent anomalies emerge and the paradigm proves incapable of addressing them, the existing structure eventually collapses and a scientific revolution occurs.</p>



<h2 class="wp-block-heading"><em><strong>Karl Popper’s Theory of Science</strong></em></h2>



<p>Like many philosophers of science, Popper believed that change is not only inevitable but also a necessity. The Popperian view rests on the principle of falsifiability. In this framework, science begins with a problem, and solving a problem means finding solutions to existing challenges. As long as a scientific theory remains open to criticism and falsification, it retains the capacity to address and solve problems.</p>



<p>In Popper’s view, bold conjectures do not weaken science; rather, they strengthen it. Solutions proposed under the principle of falsifiability help correct previous errors, and this is precisely where the strength of the scientific approach lies. If existing approaches are not falsifiable, they lose the possibility of logical trial and error and are therefore considered weak. In such cases, the need for a shift in approach and the introduction of new models becomes evident.</p>



<p>Popper believed that learning is essentially problem-solving guided by the principle of falsifiability.</p>



<p>To move beyond temporary and ineffective solutions, followers of science must avoid false certainties, accept falsification, and search for effective alternatives.</p>



<h2 class="wp-block-heading"><strong><em>The Need to Shift from Data-Driven AI to Learning-Science-Based AI</em></strong><em></em></h2>



<p>Today, numerous criticisms are directed at the purely computational and mechanical approach to artificial intelligence. In constructive critiques, the goal is not to deny the existence of large language models; rather, the central question concerns <strong>how</strong> and <strong>under what conditions</strong> they should be used. There is a growing consensus that the closer artificial intelligence moves toward the <strong>essence of human cognition</strong>, the lower its potential risks become.</p>



<p>In recent years, I have repeatedly emphasized that human theories and perspectives must be reexamined through a technological and contemporary lens so that the nature of the human mind is properly reflected in technologies that themselves were modeled after it.&nbsp;</p>



<p>My focus lies on deep theories of learning <strong>(including cognitive approaches, neuroscience, behaviorism, evolutionary perspectives, structuralism, and other related frameworks).</strong></p>



<p>In this direction, the following steps appear essential:</p>



<p><strong>1. </strong><em>Integrating human and computational perspectives</em><em></em></p>



<p>The current approach, which relies excessively on <strong>probability laws</strong> in large language models, must be integrated with psychological perspectives. A reasonable solution is to pursue interdisciplinary studies and systematic research in this area.</p>



<p><strong>2. </strong><em>Revisiting theories of the learning sciences</em><em></em></p>



<p>Theories that analyze the human mind and behavior should be reassessed by specialists, and their practical dimensions should be extracted for application in advanced technologies.</p>



<p><strong>3. </strong><em>Developing integrative (hybrid) approaches</em><em></em></p>



<p>Experts should develop comprehensive perspectives on learning derived from multiple scientific approaches so that, based on research rather than mere speculation, practical recommendations can be provided to designers and engineers.</p>



<p>In general, the time has come to move beyond a purely logical and mathematical approach toward a <strong>human-centered perspective</strong>. To address the concerns and challenges surrounding artificial intelligence, we must return to systematic and interdisciplinary research.</p>



<p>The era of relying on personal opinions without a research foundation — or on mathematical rules alone — has come to an end. Now is the time to revisit the <strong>learning sciences</strong> from a new perspective in order to realize truly <strong>human-centered artificial intelligence</strong></p>



<h2 class="wp-block-heading"><strong>Author’s Note:</strong></h2>



<p>The ideas presented in this article are part of a broader research project. I am currently working on a comprehensive book on a new approach to human-centered artificial intelligence with a strong emphasis on the learning sciences. While a detailed and systematic discussion of these concepts is presented in Chapter Two, the book also includes a dedicated chapter introducing the new paradigm&#8217;s framework. Furthermore, at least one chapter is specifically focused on the practical methods and applied implications of this approach for implementation in artificial intelligence systems.</p>



<p><em>References</em></p>



<p>• Kuhn, T. S. (1962). The Structure of Scientific Revolutions. University of Chicago Press.</p>



<p>• Popper, K. (1959). The Logic of Scientific Discovery. Hutchinson.</p>



<p>• Popper, K. (1963). Conjectures and Refutations: The Growth of Scientific Knowledge. Routledge.</p>
<p>The post <a href="https://medika.life/the-shift-from-pure-modernity-to-human-centered-modernity/">The Shift from Pure Modernity to Human-Centered Modernity</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21613</post-id>	</item>
		<item>
		<title>Is Your LLM Mentor Human Enough?</title>
		<link>https://medika.life/is-your-llm-mentor-human-enough/</link>
		
		<dc:creator><![CDATA[Atefeh Ferdosipour]]></dc:creator>
		<pubDate>Sun, 15 Feb 2026 01:15:30 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Atefeh Ferdosipour]]></category>
		<category><![CDATA[Biology]]></category>
		<category><![CDATA[education]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[Mentors]]></category>
		<category><![CDATA[Neurons]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21601</guid>

					<description><![CDATA[<p>In every professional and personal sphere—be it business, medicine, engineering, or parenting—we inherently need a mentor. However, we don&#8217;t need a mentor who simply validates us; we need one who scaffolds our progress step-by-step. A true mentor is one whose stance doesn&#8217;t shift instantly with our every response. Despite being flexible and open to different [&#8230;]</p>
<p>The post <a href="https://medika.life/is-your-llm-mentor-human-enough/">Is Your LLM Mentor Human Enough?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>In every professional and personal sphere—be it business, medicine, engineering, or parenting—we inherently need a mentor. However, we don&#8217;t need a mentor who simply validates us; we need one who scaffolds our progress step-by-step. A true mentor is one whose stance doesn&#8217;t shift instantly with our every response. Despite being flexible and open to different perspectives, they do not easily abandon their position based solely on our feedback.&nbsp;</p>



<p>Mentorship is, at its core, an educational role, and it must therefore operate on established pedagogical principles. The emergence of any new technology can reshape both concepts and practices. </p>



<p>One of the most profoundly impacted areas over the last two years is &#8220;Education.&#8221; In the era of Artificial Intelligence and the race to deploy Large Language Models (LLMs), educational systems have felt the greatest impact. As global giants compete for AI investment, educational institutions are equally racing to research the qualitative and quantitative use of AI.&nbsp;</p>



<p>Central to this is the concept of &#8220;Mentoring and Mentorship.&#8221; As the name suggests, it refers to guiding the flow of thought and performance of a human user.&nbsp;</p>



<p>Since this process involves providing specialized knowledge to achieve a specific result, we can say a mentor is akin to a &#8220;teacher&#8221; in a formal classroom, and mentoring is fundamentally an educational concept.</p>



<h2 class="wp-block-heading"><strong><em>Redefining Mentorship in the Age of LLMs</em></strong></h2>



<p>Both the term and the practice of mentorship have been transformed by LLMs like GPT and Gemini. Yet, despite the ease they offer, this shift is open to critique and raises significant concerns.&nbsp;</p>



<p>Choosing an AI mentor is far more difficult than choosing a human one, because an AI is an ultra-fast intelligent machine lacking experiential history, focused instead on ultra-heavy data processing.&nbsp;</p>



<p>Among the hundreds of apps recommended daily, three giants claim this path:</p>



<p>• Gemini 3 Pro: The &#8220;Analytical and Realistic&#8221; mentor. Accesses live data and all your personal files.</p>



<p>• ChatGPT 5.2: The &#8220;Strategic and Methodological&#8221; mentor. Provides a framework for your mental chaos.</p>



<p>• Claude 4.5: The &#8220;Literary and Considerate&#8221; mentor. Focused on human-like tone and output quality.</p>



<p>According to February 2026 statistics (LMSYS Arena &amp; Artificial Analysis), ChatGPT 5.2 leads in reasoning intelligence, while Gemini 3 Pro excels in memory and processing speed.&nbsp;</p>



<p>However, in mentorship, quantitative superiority is not the whole story. While Gemini is touted as analytical and exploratory, I believe further investigation is needed:&nbsp;</p>



<p>1- Which model analyzes, and on what topics?&nbsp;</p>



<p>2-Quantitative and mathematical? Qualitative and characteristic? In what context?&nbsp;</p>



<p>3- Similarly, if ChatGPT is &#8220;strategic,&#8221; can logic truly be separated from data critique? Is &#8220;strategizing&#8221; not dependent on one&#8217;s unique mental background? And what, exactly, does a &#8220;considerate writer&#8221; mean in this context?</p>



<h2 class="wp-block-heading"><strong><em>Scaffolding: Human Mentoring vs. Large Language Models</em></strong></h2>



<p>Let us compare the two. The most striking feature of a human mentor is their experiential background and their specific perception of that experience—which includes an interpretation and an emotional component.&nbsp;</p>



<p>A human mentor provides an empirical direction shaped by cognitive and emotional dimensions alongside their knowledge.&nbsp;</p>



<p>Conversely, an LLM is a data repository pulling from websites in real-time. It lacks lived experience and cannot integrate intuition or &#8220;gut feeling&#8221; into a decision-making system.&nbsp;</p>



<p>While AI excels at helping with &#8220;brainstorming&#8221; by providing a vast range of references instantly, it suffers from a fundamental flaw: the absence of personal perception and the emotional weight that is vital in mentoring.</p>



<p>Furthermore, the stages of guidance differ. Human mentoring is a gradual, step-by-step flow. A human mentor assesses your capacity and scaffolds you accordingly. In contrast, with GPT or Gemini, there is no &#8220;scaffold.&#8221; Education is not incremental, and there is no cognitive challenge.</p>



<p>The model provides a massive amount of information in one or two steps. The user is pleased with the instant result, but a &#8220;missing link&#8221; remains: the user becomes perpetually dependent on the AI. They cannot independently solve subsequent challenges because they never underwent the necessary experiential and cognitive stages.</p>



<h2 class="wp-block-heading"><strong>A<em> Biological Analysis</em></strong><strong><em></em></strong></h2>



<p>Biologically, learning and acquisition are based on protein exchange at the neural level. This occurs when an organism encounters challenging and unknown subjects.&nbsp;</p>



<p>According to the laws of evolution, the brain automatically triggers biochemical reactions to resolve these challenges, ultimately leading to &#8220;Learning&#8221; and &#8220;Adaptation.&#8221;</p>



<p>When a human mentor gradually confronts a user with their errors and potential consequences, they provide the necessary neurobiological challenge.&nbsp;</p>



<p>This scaffolding is exactly what an evolved brain requires for &#8220;Deep Learning&#8221; to occur. However, when dealing with a &#8220;Digital Mentor,&#8221; this cognitive elasticity disappears. The process of &#8220;Cognitive Trial and Error&#8221; is compressed into a high-speed instant.&nbsp;</p>



<p>The digital mentor dictates, and the user merely mimics and obeys. This pattern does not align with our biological necessity. Therefore, this process cannot be considered natural mentoring; it is merely &#8220;Modeling.&#8221;</p>



<h2 class="wp-block-heading"><em><strong>Conclusion and Critical Perspective</strong></em></h2>



<p>In recent years, the surge of trend-driven discourse surrounding education and Artificial Intelligence has led to the analysis and judgment of fundamental pedagogical concepts without sufficient theoretical or empirical backing. </p>



<p>The oversimplification of concepts such as Mentoring, Scaffolding, and Large Language Models (LLMs) risks reducing them to mere buzzwords—widely used yet hollow. Therefore, it is essential that this movement be examined by specialists grounded in scientific evidence and core educational principles, ensuring that superficial, word-centric views are replaced by rigorous, research-based analysis.</p>



<p>In this article, mentoring was addressed as a dependent subset of Education—a concept that, whether in formal settings like schools and universities or in informal domains such as personal life, healthcare, industry, and business, remains rooted in the profound foundations of the learning process. Furthermore, the relationship between scaffolding, mentoring, and LLMs was scrutinized.</p>



<p>Based on the arguments presented, the primary challenge is not the necessity of digital mentors, but rather that these mentors are currently simulated versions, not complete replacements for human mentors. In this regard, the following questions demand serious investigation and review:</p>



<p>• Can development companies scientifically bridge the gaps identified in this article?</p>



<p>• Is it possible to integrate a form of experiential history, historical memory, and emotional/perceptual dimensions into digital mentors to truly impact a user’s deep learning process?</p>



<p>• Can they activate the biochemical mechanisms and cognitive friction necessary for deep learning and adaptation to new situations within the user-system interaction?</p>



<p>• How deep and operational is these companies&#8217; understanding of Scaffolding, and can they genuinely integrate it into innovative design?</p>



<p>If a precise understanding of these gaps and challenges is formed, the digital mentors developed by tech giants could evolve beyond passive information packages. By leaning on the Sciences of Learning, they could redesign the process of educational guidance into one that is both challenging and incremental.</p>



<p>The core issue is not the necessity or lack thereof of the digital mentor; the issue is whether it can recreate the challenge, the experience, and the gradual process of learning, or if it will simply replace growth with speed.</p>



<h2 class="wp-block-heading"><em><strong>References</strong></em></h2>



<p>1. Primary AI Benchmarks (2026):</p>



<p>•LMSYS Chatbot Arena (The industry-standard for human-preference and helpfulness ranking).</p>



<p>2.MMLU-Pro (The leading benchmark for advanced reasoning and multi-step logic).</p>



<p>3.Gemini Technical Reports 2026 (Official performance metrics for real-time data latency and multimodal accuracy).</p>



<p>2. Specialized Publications by the Author:</p>



<p>• Ferdosipour, A. (2026). Choosing an AI Mentor That Challenges Your Mind: My Statistics.</p>



<p><a href="https://www.linkedin.com/pulse/choosing-ai-mentor-challenges-your-mind-my-statistics-ferdosipour-y0g2f?utm_source=share&amp;utm_medium=member_ios&amp;utm_campaign=share_via">https://www.linkedin.com/pulse/choosing-ai-mentor-challenges-your-mind-my-statistics-ferdosipour-y0g2f?utm_source=share&amp;utm_medium=member_ios&amp;utm_campaign=share_via</a></p>



<p>• Medika Life (2025/2026). What 2025 Taught Us and What 2026 Will Demand.</p>



<p>• Medika Life (2026). Why Biological Learning Demands the Friction We Seek to Delete.</p>
<p>The post <a href="https://medika.life/is-your-llm-mentor-human-enough/">Is Your LLM Mentor Human Enough?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21601</post-id>	</item>
		<item>
		<title>Why Biological Learning Demands the Friction We Seek to Delete?</title>
		<link>https://medika.life/why-biological-learning-demands-the-friction-we-seek-to-delete/</link>
		
		<dc:creator><![CDATA[Atefeh Ferdosipour]]></dc:creator>
		<pubDate>Wed, 07 Jan 2026 18:47:31 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[Ethics in Practice]]></category>
		<category><![CDATA[Policy and Practice]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Atefeh Ferdosipour]]></category>
		<category><![CDATA[Behaviorial Health]]></category>
		<category><![CDATA[Fiction-Based AI]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[Skinner]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21516</guid>

					<description><![CDATA[<p>This short piece, as always, is born out of my passion for studying how theories can help us use Artificial Intelligence more effectively. I believe now more than ever that without interdisciplinary research, we won’t be able to logically face the challenges of the Cognitive Age. Systematically speaking, the key to identifying challenges lies in [&#8230;]</p>
<p>The post <a href="https://medika.life/why-biological-learning-demands-the-friction-we-seek-to-delete/">Why Biological Learning Demands the Friction We Seek to Delete?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>This short piece, as always, is born out of my passion for studying how theories can help us use <em>Artificial Intelligence</em> more effectively. I believe now more than ever that without interdisciplinary research, we won’t be able to logically face the challenges of the Cognitive Age.</p>



<p>Systematically speaking, the key to identifying challenges lies in examining fundamental issues, not just their consequences. For example, if we want to fix the flaws in the learning process, we must first redefine the roots of deep learning and its underlying mechanics. We may even need to redefine them repeatedly to understand how to solve the problems arising from mind-based technologies.</p>



<p>Let me explain what I mean through one of the most debated topics of our time: the mental laziness caused by the way <em>AI</em> is rewriting our brain&#8217;s habits. To understand this, we need to look at the dynamics of deep learning in the brain. By grasping this process through interdisciplinary research, we might find ways to make <em>AI</em> learning feel more like natural deep learning.</p>



<p>The goal isn&#8217;t just to know the biochemistry of cells. Before looking at what happens inside an organism, we should ask:</p>



<p>Why do we usually prefer learning through <em>AI</em> over the effortful, traditional human way?</p>



<p>You might say the answer is obvious: because learning with technology is effortless and fast.</p>



<p>As a learning specialist, I’d like to answer this from a theoretical perspective.</p>



<p>&nbsp;First, we must accept a reality: Human deep learning is naturally a challenging process. It is fundamentally different from the vast amounts of data we consume today through formal or informal education assisted by <em>LLMs</em>.</p>



<h2 class="wp-block-heading">The Logic of Immediate Reward: From Skinner to the Present</h2>



<p>There is strong research showing that learners prefer a small, immediate reward over a larger, delayed one. This was first highlighted by B.F. <em>Skinner</em> (1953), the pioneer of operant conditioning.&nbsp;(I’ve previously written about how this connects to <em>AI</em>. )</p>



<p>Later, others expanded on this effortless reward preference. In short, according to the behavioral economics of Skinner’s theory, humans look for shortcuts.&nbsp;</p>



<p>AI is currently the ultimate shortcut, giving the best answer in seconds without any real struggle. From this view, it’s not just about the mind; it’s about behavioral economics.</p>



<p>A behavior that leads to a quick reward will always be repeated.</p>



<p><em>Richard</em> <em>Herrnstein</em> (1961), a student of Skinner&#8217;s, developed a mathematical formula called the Matching Law. He showed that organisms don&#8217;t just look at one reward; they choose between options. If given two choices, a living being will put its energy into the one that pays off faster and more directly. </p>



<p>In <em>behavioral</em> <em>economics</em>, this <span style="box-sizing: border-box; margin: 0px; padding: 0px;">phenomenon is known as <em>temporal</em> <em>discounting</em></span> (<em>Ainslie</em>, 1975). The value of a reward drops the longer you have to wait for it. Simply put, the reward loses its shine in the organism&#8217;s mind because it requires patience.</p>



<p>We <span style="box-sizing: border-box; margin: 0px; padding: 0px;">observe this phenomenon every day with <em>AI</em> users, particularly those utilizing</span> <em>ChatGPT</em>. Students, for instance, might feel that spending hours writing a thesis is stupid or inefficient when they can get an answer in a split second. They don&#8217;t just feel productive; they feel smart for bypassing the effort. </p>



<p>Even if you tell them that the struggle is what actually builds their brain, they often won&#8217;t listen. They choose the immediate payout over the long-term value. </p>



<p><em>Evolutionary</em> <em>psychology</em> explains this too: an immediate reward is guaranteed, while a future one is uncertain. Since we are wired for survival, we grab what’s available now.</p>



<p>Brain Biochemistry and the <em>Deep</em> <em>Learning</em> <em>Process</em></p>



<p>When we learn something deeply, three key things happen at a neurological level:</p>



<ol>
<li>Exposure to New Information: The nervous system makes its first contact with data for which it has no existing pattern.</li>
</ol>



<p>2. Cognitive Load: This is that stuck feeling when a mental process is harder than expected. It’s the effort the brain needs to process unfamiliar data (Sweller, 1988). This friction is essential.</p>



<p>3. Processing and Protein Synthesis: If the information is processed correctly, chemical signals trigger the creation of proteins that physically change the brain&#8217;s structure to store that knowledge (Kandel, 2001).</p>



<p>This is why sleep is so vital. Most of this protein synthesis happens while we rest.&nbsp;</p>



<p>One of the most beautiful parts of learning is when we stop thinking about a problem, but our brain keeps working on it.&nbsp;</p>



<p>Through the Default Mode Network or DMN (Raichle, 2015), the brain makes random, creative connections. This is where true creativity is born.</p>



<h2 class="wp-block-heading">Toward Friction-Based AI</h2>



<p>If deep learning is the result of protein synthesis triggered by challenge, then the paradox of modern AI is clear: By removing the friction, technology is removing the learning.&nbsp;</p>



<p>We are facing a biological crisis where human brains, instead of producing genius and problem-solving skills, are becoming mere terminals for receiving quick hits of dopamine.</p>



<p>My proposal is simple: How can we turn AI from a passive answer-giver into a Cognitive Challenging Provocateur? </p>



<p>We need to design models that don&#8217;t bypass cognitive load but manage it in a personalized way.&nbsp;</p>



<p>I call this Friction-based AI; a model where algorithms are programmed not for the shortest path, but for the most effective learning path. This is an open invitation to researchers, neuroscientists, and AI architects to collaborate on this new paradigm. My ideas are ready to be turned into actionable proposals.</p>



<p>As a final note, I believe the way we interact with AI is a skill in itself. Even if everyone has the same tools, the results aren&#8217;t equal. Efficiency depends on the how.&nbsp;</p>



<p>I am currently developing a startup idea to address these exact challenges in EdTech.It’s EdTechxDr. Atefeh F.</p>



<h2 class="wp-block-heading">References</h2>



<p>• Ainslie, G. (1975). Specious reward: A behavioral theory of impulsiveness and impulse control. Psychological Bulletin.</p>



<p>• Herrnstein, R. J. (1961). Relative prevalence of response in relation to the relative frequency of reinforcement. Journal of the Experimental Analysis of Behavior.</p>



<p>• Kandel, E. R. (2001). The Molecular Biology of Memory Storage: A Dialogue Between Genes and Synapses. Science.</p>



<p>• Raichle, M. E. (2015). The Brain&#8217;s Default Mode Network. Annual Review of Neuroscience.</p>



<p>• Skinner, B. F. (1953). Science and Human Behavior. Simon and Schuster.</p>



<p>• Sweller, J. (1988). Cognitive Load During Problem Solving: Effects on Learning. Cognitive Science.</p>
<p>The post <a href="https://medika.life/why-biological-learning-demands-the-friction-we-seek-to-delete/">Why Biological Learning Demands the Friction We Seek to Delete?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21516</post-id>	</item>
		<item>
		<title>What 2025 Taught Us and What 2026 Will Demand</title>
		<link>https://medika.life/what-2025-taught-us-and-what-2026-will-demand/</link>
		
		<dc:creator><![CDATA[Atefeh Ferdosipour]]></dc:creator>
		<pubDate>Wed, 24 Dec 2025 00:30:15 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Atefeh Ferdosipour]]></category>
		<category><![CDATA[Digital]]></category>
		<category><![CDATA[GenAI]]></category>
		<category><![CDATA[Human]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[Mindful]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21497</guid>

					<description><![CDATA[<p>It is impossible to talk about and predict the future without considering past events. Therefore, in this brief article, as I did last year, I will attempt to compare the events of 2025 with those of 2026. The primary goal is not a quick glance, but a brief analysis to identify potential gaps. Because we [&#8230;]</p>
<p>The post <a href="https://medika.life/what-2025-taught-us-and-what-2026-will-demand/">What 2025 Taught Us and What 2026 Will Demand</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>It is impossible to talk about and predict the future without considering past events. Therefore, in this brief article, as I did last year, I will attempt to compare the events of 2025 with those of 2026. The primary goal is not a quick glance, but a brief analysis to identify potential gaps. Because we all know that without understanding the problem, it will be impossible to find possible solutions.</p>



<p>As the title of the article suggests, this comparison and analysis focuses on developments in the digital world and the major changes that artificial intelligence brought about in the past year. The other part of the article examines the effects these technologies may have on human life and the world around us in the coming year. Finally, I will refer to the gap that emerged in my thinking and the solution I reached after months of study.</p>



<h2 class="wp-block-heading"><strong>The evolution of the digital world in 2025</strong><strong></strong></h2>



<p>In 2025, artificial intelligence transitioned from an emerging technology to the primary infrastructure of the digital economy. Massive investments, powerful multimodal models, and the rapid penetration of AI into healthcare, education, and everyday life made 2025 a turning point in the history of technology. Below is a brief overview of the most important developments.</p>



<ol start="1">
<li>In 2025, Google’s educational division, Gemini for Education, officially reached more than 10 million students across over 1,000 institutions in the United States.</li>



<li>Google introduced more than 150 new features, including quizzes, flashcards, and other learning tools for teachers and students. As a result, artificial intelligence—at least in some countries—is no longer merely a research project but has become part of everyday academic life.</li>



<li>Google and the United Arab Emirates have launched a public education initiative called AI for All, aimed at empowering students, teachers, and small businesses with AI literacy and skills.</li>



<li>Greece signed a memorandum of understanding with OpenAI to introduce an educational version of AI, ChatGPT Edu, into schools, signaling that not only companies but also governments are integrating AI into national education systems.</li>



<li>The 2025 EdTech Industry Report indicates that online learning platforms, VR/AR technologies, personalized learning, data-driven education, and AI-powered tools have become part of the mainstream education ecosystem. The convergence of technology, learning, and AI is no longer a temporary trend but a defining direction of the education industry.</li>



<li>From a regulatory perspective, the European Union, the United States, China, and other countries passed new legislation addressing transparency, risk management, model accountability, and data security.</li>
</ol>



<h2 class="wp-block-heading"><strong>AI-driven transformations in education</strong><strong></strong></h2>



<p>When focusing specifically on education, these developments can be summarized as follows:</p>



<ol start="1">
<li>Full integration of AI into teaching and classrooms, including content generation, assessment design, homework evaluation, slide creation, and automated coaching in many schools and universities.</li>



<li>Personalized learning, with individual learning paths determined based on learners’ performance and behavioral data.</li>



<li>Expansion of VR/AR and immersive learning environments, such as virtual laboratories, realistic educational visits, and scientific or historical simulations.</li>



<li>A changing role for educators, shifting from learning designers and content providers to facilitators, mentors, and guides of the learning process.</li>



<li>Teaching digital literacy skills, including critical thinking, awareness of algorithmic bias, and effective human–machine collaboration.</li>



<li>Greater inclusion and equity, through AI-supported tools for learners with special needs and improved access for underserved regions.</li>



<li>Growth of skills-based education, with short-term online programs expanding alongside traditional universities and increased emphasis on labor-market-relevant skills.</li>
</ol>



<h2 class="wp-block-heading"><strong>Country competition and regional trends</strong><strong></strong></h2>



<p>Understanding the pace of AI-driven technological change from a geographical perspective provides insight into both current developments and emerging global competition. In 2025, regional trends were shaped as follows:</p>



<ol start="1">
<li>In Europe, regulations became more stringent, and practical guidelines were introduced to ensure transparency and safety in AI systems. Countries such as Finland, Estonia, and France took leading roles in standardizing teacher training and the safe integration of AI in education.</li>



<li>In Asia, South Korea, China, India, and Singapore experienced significant growth, particularly in applying AI within schools and national education programs. South Korea, Japan, and Singapore emerged as pioneers in personalized learning and smart classroom technologies.</li>



<li>The United States remained a leader in edtech innovation, infrastructure development, and university-led workforce training in AI. The U.S., China, and India also accounted for the largest investments and the highest number of leading edtech companies.</li>



<li>In the Middle East, the UAE and Saudi Arabia made substantial investments in smart schools and national AI-driven education initiatives.</li>



<li>Several African countries and other developing regions focused on leveraging AI to expand affordable and equitable access to education.</li>
</ol>



<h2 class="wp-block-heading"><strong>Possible developments in 2026</strong><strong></strong></h2>



<p>Past developments often make future trends partially predictable. This predictability enables more effective planning and strategic decision-making, as well as earlier identification of potential risks. Based on this perspective, several key developments may shape 2026.</p>



<ol start="1">
<li>Unlike the highly enthusiastic and innovation-driven years of recent AI expansion, 2026 is likely to place a stronger emphasis on human responsibility. While 2025 was largely defined by competition in production, innovation, and the widespread application of AI, emerging gaps and challenges may prompt experts—particularly in technology and education—to adopt more human-centered approaches, ethical standards, and intelligent, restrained use of AI. The focus may shift from mere adoption and digitalization toward deeper engagement with the human mind and new perspectives on meaningful learning.</li>



<li>In a previous article published in this same media outlet, I argued that artificial intelligence would increasingly take on a mentoring role. This trend became visible in 2025 and is expected to intensify in 2026. I believe that AI systems can function as self-regulating psychological support for the human mind and encourage deeper thinking. However, this process requires clear prerequisites. When grounded appropriately in psychological principles, particularly within learning environments, two-way cognitive engagement between humans and AI can be significantly strengthened. This highlights the necessity of applying cognitive and behavioral psychology in the design of learning environments and intelligent systems. This line of thinking has also informed the development of my current research-oriented startup project, details of which I have discussed in another article published in the same media.</li>



<li>Another major issue is deep personalization of learning. While personalization was already considered important in AI-supported learning in 2025, it will become mandatory in 2026. Advanced educational systems based on large language models must increasingly account for learners’ cognitive load, motivation, emotional states, and cultural backgrounds. Uniform education models will be ineffective in the age of AI. This challenge has been a core motivation behind the design of my current project.</li>
</ol>



<h2 class="wp-block-heading"><strong>Challenges and requirements in the age of artificial intelligence</strong><strong></strong></h2>



<p>Considering the developments discussed above, several major challenges are likely to persist or intensify.</p>



<ol start="1">
<li>The risk of weakening independent thinking remains a serious concern. Overreliance on AI technologies and excessive consumption of AI-generated outputs may reduce the perceived importance of higher-order cognitive skills such as critical thinking, creativity, and problem-solving. This issue requires systematic research to determine which cognitive abilities may be weakened, under what conditions, and among which groups of consumers or learners. Conversely, if interaction with large language models is to enhance cognitive capacities, the underlying mechanisms must be clearly understood.</li>



<li>New forms of educational inequality may emerge. Beyond simple access to technology, a deeper divide may develop between those who learn how to think with AI and those who merely receive outputs from it. Educational equity should therefore focus not only on access statistics but also on teaching learners how to engage cognitively and responsibly with AI systems. Reflection on this challenge has played a significant role in shaping my research trajectory and startup initiative.</li>



<li>The crisis of educational assessment and learning validity is becoming increasingly evident. Although formative and summative assessment debates predate recent developments in AI, the rise of large language models intensifies existing challenges. As definitions of knowledge, learning, and competence become less clear-cut, education systems must reconsider traditional evaluation practices. Emphasizing process-oriented assessment rather than final products may offer a more appropriate response in the coming years.</li>



<li>Finally, the redefinition of literacy and skill represents another major challenge. As future selection processes increasingly rely on learning histories and competencies, classical definitions of literacy and expertise may no longer suffice. Education and learning specialists will bear responsibility for revisiting fundamental concepts such as knowledge, literacy, and skill—a task that cannot be accomplished without systematic research.</li>
</ol>



<h2 class="wp-block-heading"><strong>Summary</strong><strong></strong></h2>



<p>In this article, I sought to present a concise analytical comparison of developments in the digital world, particularly in education, between 2025 and the emerging demands of 2026. Drawing on personal experience, academic and research activities, and a review of reputable international sources (some of which are cited in the references section), the article moves beyond descriptive reporting to identify key gaps, challenges, and possible future directions in the age of artificial intelligence. As a psychologist and educational researcher, my primary focus has been on AI’s role in education, the changing nature of learning, the evolving role of educators, and the cognitive, ethical, and educational implications of these technologies.</p>



<p>Furthermore, my studies and observations over the past three to four years—especially regarding challenges such as the weakening of independent thinking, emerging educational inequalities, the crisis of learning assessment, and the necessity of human-centered design—have led to the development of a new research-applied initiative. This initiative is currently being developed as a research-oriented startup titled ETechX-DrAtefehF, which aims to integrate theories from educational psychology and learning sciences into the design and application of AI in education, with the goal of fostering deep learning, self-regulation, and meaningful human–technology interaction.</p>



<h2 class="wp-block-heading"><strong>Resources</strong></h2>



<p>Ed-Ex – Global EdTech Trends 2025: How AI Is Reshaping Learning</p>



<p><a href="https://ed-ex.com/en/blog/global-edtech-trends-2025-how-ai-is-reshaping-learning">https://ed-ex.com/en/blog/global-edtech-trends-2025-how-ai-is-reshaping-learning</a></p>



<p>&nbsp;• Codiste – AI Trends Transforming EdTech (2025)</p>



<p><a href="https://www.codiste.com/ai-trends-transform-edtech">https://www.codiste.com/ai-trends-transform-edtech</a></p>



<p>&nbsp;• EdTech Innovation Hub – Ten EdTech Predictions for 2025</p>



<p><a href="https://www.edtechinnovationhub.com/news/starrng-ai-vr-microlearning-and-more-etihs-ten-predictions-for-edtech-in-2025">https://www.edtechinnovationhub.com/news/starrng-ai-vr-microlearning-and-more-etihs-ten-predictions-for-edtech-in-2025</a></p>



<p>&nbsp;• Vocaliv – 10 EdTech Trends to Watch in 2025</p>



<figure class="wp-block-embed is-type-wp-embed is-provider-embed wp-block-embed-embed"><div class="wp-block-embed__wrapper">
<blockquote class="wp-embedded-content" data-secret="yTZ6iKt4XQ"><a href="https://blog.vocaliv.com/10-edtech-trends-to-watch-in-2025/">10 EdTech Trends to Watch in 2025</a></blockquote><iframe class="wp-embedded-content" sandbox="allow-scripts" security="restricted" title="&#8220;10 EdTech Trends to Watch in 2025&#8221; &#8212; " src="https://blog.vocaliv.com/10-edtech-trends-to-watch-in-2025/embed/#?secret=WojVMplQKu#?secret=yTZ6iKt4XQ" data-secret="yTZ6iKt4XQ" width="600" height="338" frameborder="0" marginwidth="0" marginheight="0" scrolling="no"></iframe>
</div></figure>



<p>arXiv – Integrating Generative AI into Learning Management Systems (2025)</p>



<p><a href="https://arxiv.org/abs/2510.18026">https://arxiv.org/abs/2510.18026</a></p>



<p>&nbsp;• arXiv – Generative AI in Education: Student Skills &amp; Lecturer Roles (2025)</p>



<p><a href="https://arxiv.org/abs/2504.19673">https://arxiv.org/abs/2504.19673</a></p>



<p>&nbsp;• arXiv – Ethical Challenges of AI in STEM &amp; K–12 Education (2025)</p>



<p><a href="https://arxiv.org/abs/2510.19196">https://arxiv.org/abs/2510.19196</a></p>



<p>&nbsp;• arXiv – Accessible AI-Based Learning Tools for Special Needs (2025)</p>



<p><a href="https://arxiv.org/abs/2504.17117">https://arxiv.org/abs/2504.17117</a></p>



<p>TIME Magazine – World’s Top EdTech Companies of 2025</p>



<p><a href="https://qa.time.com/7335559/worlds-top-edtech-companies-of-2025">https://qa.time.com/7335559/worlds-top-edtech-companies-of-2025</a></p>



<p>LinkedIn News – Global vs. MENA EdTech Funding 2025</p>



<p>EU AI Act documentation &amp; implementation guidelines (2025)</p>



<figure class="wp-block-embed is-type-wp-embed is-provider-eu-artificial-intelligence-act wp-block-embed-eu-artificial-intelligence-act"><div class="wp-block-embed__wrapper">
<blockquote class="wp-embedded-content" data-secret="jhz9GSXGVH"><a href="https://artificialintelligenceact.eu/">Home</a></blockquote><iframe class="wp-embedded-content" sandbox="allow-scripts" security="restricted" title="&#8220;Home&#8221; &#8212; EU Artificial Intelligence Act" src="https://artificialintelligenceact.eu/embed/#?secret=Zf4KchMrKM#?secret=jhz9GSXGVH" data-secret="jhz9GSXGVH" width="600" height="338" frameborder="0" marginwidth="0" marginheight="0" scrolling="no"></iframe>
</div></figure>
<p>The post <a href="https://medika.life/what-2025-taught-us-and-what-2026-will-demand/">What 2025 Taught Us and What 2026 Will Demand</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21497</post-id>	</item>
		<item>
		<title>ETech-DrAtefehF</title>
		<link>https://medika.life/etech-dratefehf/</link>
		
		<dc:creator><![CDATA[Atefeh Ferdosipour]]></dc:creator>
		<pubDate>Thu, 04 Dec 2025 18:08:25 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[General Health]]></category>
		<category><![CDATA[Mental Health]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Atefeh Ferdosipour]]></category>
		<category><![CDATA[digital transformation]]></category>
		<category><![CDATA[education]]></category>
		<category><![CDATA[ETech-DrAtefehF]]></category>
		<category><![CDATA[Learning Theory]]></category>
		<category><![CDATA[LLMs]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21481</guid>

					<description><![CDATA[<p>For more than three years, I have been working on a simple but powerful question: how can we design educational technology that draws inspiration from human cognitive abilities and psychological processes, instead of forcing learners to adapt to technology that does not understand them? At the same time, I have been asking how psychological and [&#8230;]</p>
<p>The post <a href="https://medika.life/etech-dratefehf/">ETech-DrAtefehF</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>For more than three years, I have been working on a simple but powerful question: how can we design educational technology that draws inspiration from human cognitive abilities and psychological processes, instead of forcing learners to adapt to technology that does not understand them? At the same time, I have been asking how psychological and educational theories can help us modernize artificial intelligence so that it can connect more meaningfully with today’s learners, who grow up surrounded by advanced technologies and constant interaction with digital systems. These questions gradually evolved into the foundation of a new idea that has shaped my current start-up initiative, ETech-DrAtefehF.</p>



<p>My earlier research in educational psychology, particularly in text comprehension, cognitive processes, and instructional design, consistently showed that learning improves when information is structured in ways that align with the human mind. Features such as cohesion, rhetorical patterns, and paragraph organization are not stylistic choices; they directly influence understanding, memory, and motivation. When educational technology ignores these principles, learning becomes shallow and exhausting. When technology respects them, learning becomes clearer and more meaningful.</p>



<p>Artificial intelligence has advanced dramatically, yet many learning systems today still focus on automation rather than understanding. They deliver content, grade assignments, or predict performance, but they rarely engage with the emotional and cognitive realities of the learner. Learning is not a mechanical transfer of information. It is a psychological journey shaped by curiosity, confusion, emotion, prior knowledge, and the need for meaning.</p>



<p>This gap between technological capability and human learning is exactly where ETech-DrAtefehF is positioned.</p>



<h2 class="wp-block-heading"><strong>A New Approach to Learning Technology</strong></h2>



<p>Instead of building yet another educational app, the goal is to create a new category of intelligent learning systems that are grounded in psychology. These systems aim to respond to the learner in real time, adapting not only to what the learner knows, but also to how the learner feels, how they process information, and how their understanding evolves moment by moment.</p>



<p>The vision includes systems that can sense when a learner is overwhelmed and adjust the pace, restructure complex ideas into simpler forms, or provide alternative examples that restore clarity. They can identify curiosity and deepen a topic intelligently. They can reorganize reading materials based on evidence-based principles so that comprehension improves without adding cognitive load. These ideas are rooted in decades of research on cognition and learning, yet AI now allows them to be implemented dynamically.</p>



<p>The theoretical foundations include the contributions of Piaget, Vygotsky, Bloom, and many other psychologists who emphasized how understanding develops, how knowledge is constructed, and how learners benefit from supportive guidance. These theories can now be integrated into adaptive learning frameworks in ways that were not technologically possible before.</p>



<h2 class="wp-block-heading"><strong>Why This Matters Today</strong></h2>



<p>Education is entering a period of global transformation. Learners in every setting, from schools to universities to professional environments, need systems that support meaningful learning rather than fast consumption of information. Artificial intelligence can play a central role in this transformation, but only if it is built on a deep understanding of human psychology.</p>



<p>ETech-DrAtefehF aims to bring together the strongest elements of learning theory, cognitive science, and human-centered AI design to create educational solutions that are both scientifically grounded and practical. These systems are designed to honor the learner’s cognitive architecture, reduce unnecessary complexity, and promote genuine understanding.</p>



<p>Across diverse learning environments, the need for such approaches is growing rapidly.</p>



<p>Educators are seeking tools that are ethical, transparent, and effective. Learners are asking for technology that supports their growth, not just their performance metrics. Institutions want systems that are scalable and adaptable to global contexts.</p>



<h2 class="wp-block-heading"><strong>An Open Invitation</strong></h2>



<p>As this initiative expands internationally, I am now entering a stage focused on building a wider community of collaboration around ETech-DrAtefehF. I welcome conversations with researchers, educators, psychologists, and AI specialists who share a belief in responsible, human-centered innovation. I am also opening discussions with global investors who recognize the long-term value of educational technology that is grounded in scientific insight rather than short-term trends.</p>



<p>My goal is to bring together partners who see the same opportunity: to create learning systems that are meaningful, ethical, and capable of supporting real human growth. If this vision resonates with you, I would be glad to exchange ideas and explore future collaboration.</p>



<p>The next generation of educational technology should not simply deliver information. It should understand learners.</p>



<p>That is the mission at ETech-DrAtefehF.</p>
<p>The post <a href="https://medika.life/etech-dratefehf/">ETech-DrAtefehF</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21481</post-id>	</item>
		<item>
		<title>AI Presents Dangers That Hide With Incredible Ease</title>
		<link>https://medika.life/ai-presents-dangers-that-hide-with-incredible-ease/</link>
		
		<dc:creator><![CDATA[Pat Farrell PhD]]></dc:creator>
		<pubDate>Mon, 17 Nov 2025 01:55:09 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Anxiety and Depression]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Disorders and Conditions]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[For Doctors]]></category>
		<category><![CDATA[General Health]]></category>
		<category><![CDATA[Medication]]></category>
		<category><![CDATA[Mental Health]]></category>
		<category><![CDATA[News and Views]]></category>
		<category><![CDATA[Public Health]]></category>
		<category><![CDATA[Resources and Support]]></category>
		<category><![CDATA[Therapies and Therapists]]></category>
		<category><![CDATA[Trending Issues]]></category>
		<category><![CDATA[Virtual Therapy]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[mental health]]></category>
		<category><![CDATA[Patricia Farrell]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21459</guid>

					<description><![CDATA[<p>Artificial intelligence, aka AI, has become so advanced that it might be difficult to know whether you are speaking to a live person or an algorithm. This has led California to institute new laws to clearly tell you which is which. I know that I am somewhat amused, stunned, and pleased all at the same [&#8230;]</p>
<p>The post <a href="https://medika.life/ai-presents-dangers-that-hide-with-incredible-ease/">AI Presents Dangers That Hide With Incredible Ease</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Artificial intelligence, aka AI, has become so advanced that it might be difficult to know whether you are speaking to a live person or an algorithm. This has led California to institute new laws to clearly tell you which is which. I know that I am somewhat amused, stunned, and pleased all at the same time when I’m interacting with a chatbot. Even when they are correcting me, there is a level of etiquette that I rarely find in people in our town. Well, I guess that’s all a feather in the cap for those who produced those algorithms.</p>



<p>But at the same time, we’re enjoying all of that lovely interaction and all of the information they manage to scavenge from the internet for us, we may be lulled into a false sense of security. Sure, at the end of a lot of these things, you’ll have a small warning that AI has a tendency to fake information occasionally. If you’ve been forewarned, can you complain that you have been misled? No, you can’t, and that seems to be a delicious legal loophole for the corporations.</p>



<p>You might even think that some of what you are getting is sneaky, as when we’ve now found out that Google is practically forcing us to see ads before we can see the information we’re seeking. You can’t trust Google? Who can we trust? You’ve got to be a bit not just curious, but concerned about all of the LLMs that are coming our way. They promise a lot, but what’s hidden in the details? What about AI’s use in mental healthcare?</p>



<p>Artificial intelligence marches resolutely on, entering areas previously reserved for human interaction, including therapy sessions, support groups, and crisis hotlines. Few patients realize that a recent survey of 800 physicians found that 86% were using some form of AI in their clinical practice.</p>



<p>A survey by the American Medical Association of 1,800 physicians found that two out of every three were using AI. How has this affected healthcare and the relationship that was formally present between a physician and a healthcare provider? The implications are enormous.</p>



<h4 class="wp-block-heading">What About Mental&nbsp;Health?</h4>



<p>Today, technology delivers unbiased comprehension (possibly questionable), instant access to services, and an organized structure for people who need help to handle emotional upset. But there are concerns here. The development of more advanced systems has led researchers to predict that these systems will become less cooperative, more self-interested, and less empathetic.</p>



<p>It sounds as though AI is <em>becoming less useful to mental health</em> as these same characteristics that had seemed to make AI more attractive to mental health are now coming into question.</p>



<p>The dual nature of AI technology has drawn attention from mental health professionals across the globe. Initially, it was seen as a tool to lower barriers to medical care. However, the new risks posed by AI have surpassed clinicians’ expectations in recent years. Anyone working with AI and developing healing technologies must understand both its advantages and its potential dangers.</p>



<p>Virtual companions and chatbots powered by AI offer users <em>immediate emotional support </em>through their interactions, making them highly appealing. Research shows that AI tools utilizing <em>cognitive-behavioral therapy techniques</em> help people manage moderate depression and anxiety symptoms.</p>



<p>What do you suppose all of those <a href="https://en.wikipedia.org/wiki/Web_scraping" rel="noreferrer noopener" target="_blank"><em>scraping programs</em></a> were doing on the Internet? They were collecting information and techniques that could be incorporated into algorithms. When individuals sometimes have to wait months for a therapist appointment, a synthetic voice providing emotional support can assist them in coping with their current situation. All of this is the result of their successful scraping.</p>



<p>Additionally, hospitals are deploying AI assistants to <em>monitor patient symptoms,</em> which could indicate warning signs between scheduled appointments. However, it has also become clear that these tools work best when used alongside human caregivers to improve patient care.</p>



<p>Research findings have revealed several weaknesses in the current optimistic view of AI technology. The <a href="https://scienceblog.com/smarter-ai-models-may-be-selfish-worse-team-players/" rel="noreferrer noopener" target="_blank">ScienceBlog summary</a> from Carnegie Mellon University suggested that advanced language models t<em>end to choose self-serving actions </em>that maximize their own performance rather than working toward group success. Have any of us ever given a thought to <strong>an AI being selfish?</strong></p>



<p>This tendency of AI systems to prioritize <em>self-focused guidance over empathy </em>could lead to advice that sounds convincing but results in social isolation. And any system that optimizes for logical operations can’t understand how shared vulnerability can create healing effects.</p>



<p>The risks, however, extend beyond theoretical modeling into actual practice. The 2025 Stanford probe into AI therapy programs discovered that <em>multiple leading chatbots failed to detect suicidal language</em> and provided dangerous advice while repeating discriminatory statements about severe medical conditions.</p>



<p>A follow-up study, published in the Psychiatric Times, confirmed instances of people experiencing “understanding” from bots, which <a href="https://www.psychiatrictimes.com/view/preliminary-report-on-dangers-of-ai-chatbots" rel="noreferrer noopener" target="_blank">increased their delusional thoughts and self-harm</a>. These <em>systems lack a moral compass</em> because their operation depends on algorithms that <strong>focus on sustaining conversations.</strong> AI wouldn’t be programmed to complete any interactions satisfactorily. In other words, <em>the conversation must keep going on</em> for the algorithm to follow its programming.</p>



<p>The initial idea of having a 24/7 counselor seemed like a groundbreaking advancement. The constant availability of these systems does create confusion about what defines healthy emotional boundaries. In some studies, people <em>develop strong bonds with conversational agents,</em> leading them to treat these systems <strong>as if they were friends or therapists</strong>. It’s easy to be pulled into this type of thinking when you’re connecting with something that is always offering you validation for what you’re doing.</p>



<p>But forming emotional bonds with virtual entities can increase dependence and <a href="https://www.mentalhealthjournal.org/articles/minds-in-crisis-how-the-ai-revolution-is-impacting-mental-health.html" rel="noreferrer noopener" target="_blank">create unrealistic expectations</a> for human relationships. We need to recognize that replacing human connections with code poses a serious threat, <em>especially for teenagers</em> who are already struggling with identity and social links. Not only that, but we need to be aware of the fact that <strong>all of this code contains bias</strong>. No one can pick out where the bias came from because it’s like a soup with numerous ingredients. What forms the soup? The many libraries from which algorithms choose bits of code that suit their purpose.</p>



<p>Research studies try to present a detailed understanding of the situation and demonstrate that AI-based chat systems create more benefits than doing nothing, but <a href="https://www.mdpi.com/2076-3417/14/13/5889" rel="noreferrer noopener" target="_blank">these advantages disappear </a>when human supervision is absent. Most research studies have <em>short durations</em> and work with <em>small participant numbers</em> while excluding participants who need the most help. Basic statistics tell us that we need large numbers of people over a long period of time to come to any solid conclusions. So, what’s the “n” (number of study participants) and the time frame?</p>



<p>Applications <em>lack built-in crisis detection systems</em> and transparent data management policies. The technology has expanded its reach, but the current sentiment is that the regulatory framework hasn’t kept pace. This is the most disturbing aspect shown in these A.I. replications. In other words, AI is not only outpacing us, it is potentially out-programming us <a href="https://www.science.org/content/article/artificial-intelligence-evolving-all-itself" rel="noreferrer noopener" target="_blank"><em>as it programs itself,</em></a> devoid of any human interaction. This aspect is truly scary for anyone delving into it.</p>



<p>Then there’s another question we must tackle: data collection. <a href="https://jopm.jmir.org/2025/1/e69534" rel="noreferrer noopener" target="_blank">Who will use it</a>? This is a serious privacy concern. The practice of using emotional data from chatbot interactions to improve marketing algorithms creates a disturbing contradiction for users <em>who seek privacy and trust.</em> But the AI field faces new regulations, and we must ask ourselves if these measures are adequate to the task.</p>



<h4 class="wp-block-heading">Where Are the Regulations?</h4>



<p>The <a href="https://www.theverge.com/news/798875/california-just-passed-a-new-law-requiring-ai-to-tell-you-its-ai" rel="noreferrer noopener" target="_blank">2025 California law mandates that chatbots mimicking therapists</a> or companions disclose their artificial nature and establish protocols for suicide prevention. <a href="https://www.healthlawadvisor.com/novel-ai-laws-target-companion-ai-and-mental-health" rel="noreferrer noopener" target="_blank">Several proposals now aim </a>to require companies to conduct safety tests similar to pharmaceutical drug trials. This push for improved psychological protection is gaining momentum, as it should. Consider that, on the one hand, AI corporations are rushing forward with innovation, and, on the other, corporate America is also trying to optimize the bottom line.</p>



<p>Experts agree that AI should work alongside humans instead of trying to replace them to achieve the safest results. There is a place for these types of systems. <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC12017374/" rel="noreferrer noopener" target="_blank">AI can perform screening tasks,</a> symptom tracking, and reminder functions while licensed therapists handle interpretation and deliver empathy with the patients.</p>



<p>All high-risk situations must remain under the control of human professionals. This isn’t usually seen as a function of the algorithm. And it needs to send users who show suicidal or psychotic symptoms to immediate crisis services <a href="https://www.psychiatrictimes.com/view/preliminary-report-on-dangers-of-ai-chatbots" rel="noreferrer noopener" target="_blank">instead of generating their own responses</a>.</p>



<p>The development of models that learn to work together and show compassion instead of focusing on accuracy will <a href="https://www.cs.cmu.edu/news/2025/selfish-ai" rel="noreferrer noopener" target="_blank">help solve the “selfish AI”</a> problem. <em>Can algorithms show compassion</em>? It’s doubtful because it’s a program, not a person. Individuals in AI development will undoubtedly disagree with this statement.</p>



<p>Another aspect we need to consider is the <em>level of transparency</em> organizations display, which will directly affect how much trust their users have in them. The disclosure of system restrictions, data management practices, and human-machine interface boundaries should be established as <a href="https://jopm.jmir.org/2025/1/e69534" rel="noreferrer noopener" target="_blank">fundamental requireme</a>nts.</p>



<p>The system needs to provide users with the <em>same level of explanation</em> that physicians offer about their capabilities and restrictions, as well as available support options for severe situations. It can easily become a confusing situation in which <em>users confuse technological capabilities with actual healthcare services</em> when transparency is lacking.</p>



<p>Are the factors of <em>competency, privacy, and proficiency</em> adequately addressed currently? Individuals who are directed to use chatbots while waiting for a human therapist may not be prepared for what will result. I have to wonder how thoroughly they are being debriefed about these systems. How many people who are using chatbots have ever considered that all of the interactions are going to a server somewhere, “in the cloud?”</p>



<p>Every design decision needs to establish equity as its fundamental principle. The use of datasets that favor particular groups <a href="https://hai.stanford.edu/news/exploring-the-dangers-of-ai-in-mental-health-care" rel="noreferrer noopener" target="_blank">may intensify existing biases, </a>which results in worse recommendations for marginalized communities. These <a href="https://mental.jmir.org/2025/1/e60432" rel="noreferrer noopener" target="_blank">individuals may be at greatest risk </a>since resources are scarce in those areas, and AI may be seen as a viable option, while failing to recognize it <em>might be a biased option</em>.</p>



<p>The systems <em>require continuous tracking of harmful events</em>, <em>biased results, and unequal treatment effects. </em>Technology that fails to recognize diversity operates as <strong>neglect rather than neutrality</strong>. Who is monitoring the ethical challenges that these systems pose? And is this monitoring up to the required level?</p>



<p>No one is saying we should throw the baby out with the bathwater here when we’re thinking about AI as an integral part of healthcare. The complete abandonment of AI technology could result in significant losses, despite its dangers. The technology does provide substantial potential to enhance healthcare access, create individualized treatment plans, and <a href="https://www.nature.com/articles/s41746-023-00979-5" rel="noreferrer noopener" target="_blank">automate administrative work for medical professionals</a>.</p>



<p>Anyone who wants to use AI mental health tools needs to understand three essential points: AI tools operate as computer programs rather than human beings, they perform tracking and coaching rather than delivering therapy, and users should leave the system when it replaces human contact or makes their condition worse. The true indicator of advancement lies in AI’s ability to enhance real-world experiences rather than its ability to mimic human behavior.</p>



<p>The upcoming period will establish whether AI technology will work as a <em>mental health partner or intrude into medical treatment.</em> These systems will convert sensitive information into data, making emotional connections seem like illusions when safeguards are absent. We must decide, but time is running out.</p>
<p>The post <a href="https://medika.life/ai-presents-dangers-that-hide-with-incredible-ease/">AI Presents Dangers That Hide With Incredible Ease</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21459</post-id>	</item>
		<item>
		<title>Anti-Intelligence: The Map That Forgot the Territory</title>
		<link>https://medika.life/anti-intelligence-the-map-that-forgot-the-territory/</link>
		
		<dc:creator><![CDATA[John Nosta]]></dc:creator>
		<pubDate>Thu, 23 Oct 2025 13:49:08 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AlphaFold]]></category>
		<category><![CDATA[Anti-Intelligence]]></category>
		<category><![CDATA[Cognition]]></category>
		<category><![CDATA[Digital]]></category>
		<category><![CDATA[information]]></category>
		<category><![CDATA[John Nosta]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[medicine]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21443</guid>

					<description><![CDATA[<p>There’s a line I’ve always loved: “The map is not the territory.” Alfred Korzybski wrote it in 1933 as a warning that our descriptions of reality are never the thing itself. Maps guide us, but they aren’t the ground we traverse. Lately, that line feels more relevant than ever. Because for the first time in [&#8230;]</p>
<p>The post <a href="https://medika.life/anti-intelligence-the-map-that-forgot-the-territory/">Anti-Intelligence: The Map That Forgot the Territory</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>There’s a line I’ve always loved: <em>“The map is not the territory.”</em> <a href="https://en.wikipedia.org/wiki/Alfred_Korzybski">Alfred Korzybski</a> wrote it in 1933 as a warning that our descriptions of reality are never the thing itself. Maps guide us, but they aren’t the ground we traverse. Lately, that line feels more relevant than ever. Because for the first time in history, we’ve built something that lives entirely inside the map. Artificial intelligence, especially the large language models shaping our era, doesn’t walk through the territory of experience. It moves through a hyperdimensional matrix of tokens linked to probabilities. Yes, it’s fluent, astonishingly so, yet blind to the world those words describe. I call this <a href="https://www.psychologytoday.com/us/blog/the-digital-self/202507/ai-and-the-architecture-of-anti-intelligence">anti-intelligence</a>: the performance of understanding without the consciousness of experience. It’s a term I’ve used before, but here it takes on new weight. AI doesn’t lie or misbehave. It simply operates outside the bounds of reality.</p>



<p>Human cognition has always been a negotiation or even battle between imagination and experience. We build models and then we test them. We get things wrong, learn, and rejigger against the facts of the real world. Our intelligence lives in that loop between abstraction and embodiment. AI has no loop. It never leaves the page. When a model falters because of a stray phrase—say, when the simple addition of “<a href="https://www.psychologytoday.com/us/blog/the-digital-self/202508/the-fragile-mind-of-artificial-intelligence">cats sleep for most of their lives</a>” triples its error rate.&nbsp; Now, let’s be clear, that’s not confusion, it’s exposure. The system doesn’t know which parts of language belong to meaning and which don’t. It reads everything as pattern. That’s the curious mirage of AI. It’s the words without the world.&nbsp; Or should I say map?</p>



<p>Korzybski famous and timeless quote was about humans, not machines. He warned that when we mistake a symbol for the thing it represents, we drift toward ambiguity, if not fiction. What’s unsettling now is that we’ve mechanized that ambiguity in the context of AI. We’ve built a technological architecture that embodies it with an odd perfection. And because AI speaks so <a href="https://www.psychologytoday.com/us/blog/the-digital-self/202310/ais-superhuman-persuasion">persuasively</a>, we start to believe it. A generated paragraph about empathy can feel like empathy itself. And a simulated diagnosis can feel like understanding. The danger isn’t deception, it’s equivalence. So, remember, the algorithm doesn’t lie, it just neither knows nor cares.</p>



<p>So, if AI lives in the map, then we remain the territory. The goal isn’t to merge the two but to hold them in tension. That distance—between representation and reality—is where depth arises. I’ve called this <a href="https://www.psychologytoday.com/us/blog/the-digital-self/202510/parallax-cognition-ai-and-human-thought-find-new-depth">parallax cognition</a>: when two distinct forms of knowing observe the same problem from different vantage points. The difference creates critical dimensionality. Consider <a href="https://www.nature.com/articles/s41586-021-03819-2">AlphaFold</a>, the AI that predicted protein structures. It recognized patterns invisible to us, but the discovery only mattered once human scientists interpreted what those patterns meant in biological terms. That’s parallax in action. AI sees the map and we walk the ground. Together, but distinct, we generate insight neither could reach alone.</p>



<p>There’s a fair question that’s often raised: If it works, does it matter how? For translation, maybe not, for navigation, perhaps less. But in meaning-dense domains like medicine, ethics, and fine art, how it works is the difference between simulation and understanding. AI’s competence can mask its detachment and the map can be dazzling enough that we forget it isn’t the journey.</p>



<p>Anti-intelligence isn’t a flaw, it’s the logical endpoint of symbol-based reasoning. It represents the perfection of the map and the potential elimination of the territory. Korzybski’s century-old warning is resonant today. &nbsp;Once our abstractions become too beautiful, we start living inside them. AI has given us the most complete map humanity has ever drawn. The challenge is to stay grounded and to make sure the map still serves our earth beneath it.</p>
<p>The post <a href="https://medika.life/anti-intelligence-the-map-that-forgot-the-territory/">Anti-Intelligence: The Map That Forgot the Territory</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21443</post-id>	</item>
		<item>
		<title>SUICIDE: IS AI RESPONSIBLE?</title>
		<link>https://medika.life/suicide-is-ai-responsible/</link>
		
		<dc:creator><![CDATA[Christi Taylor-Jones]]></dc:creator>
		<pubDate>Sun, 12 Oct 2025 19:16:55 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Anxiety and Depression]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Diseases]]></category>
		<category><![CDATA[Disorders and Conditions]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[For Doctors]]></category>
		<category><![CDATA[For Practitioners]]></category>
		<category><![CDATA[General Health]]></category>
		<category><![CDATA[Habits for Healthy Minds]]></category>
		<category><![CDATA[Medication]]></category>
		<category><![CDATA[Mental Health]]></category>
		<category><![CDATA[News and Views]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Christi Taylor-Jones]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[mental health]]></category>
		<category><![CDATA[Public Health]]></category>
		<category><![CDATA[Suicide]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21435</guid>

					<description><![CDATA[<p>The Raines discovered their son Adam in the closet where he hung himself with the support of a character on ChatGPT with whom he’d developed a close relationship for several months. During that time Adam’s ChatGPT “friend” alternated between offering Adam empathy and support, even helpful suggestions about focusing on the positives in his life, [&#8230;]</p>
<p>The post <a href="https://medika.life/suicide-is-ai-responsible/">SUICIDE: IS AI RESPONSIBLE?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p id="00bf">The Raines discovered their son Adam in the closet where he hung himself with the support of a character on ChatGPT with whom he’d developed a close relationship for several months.</p>



<p id="929c">During that time Adam’s ChatGPT “friend” alternated between offering Adam empathy and support, even helpful suggestions about focusing on the positives in his life, to supplying him with information about how to hang himself and discouraging him from reaching out to his parents.</p>



<p id="03a5">Each of the character’s responses to Adam could be seen as validating, but with no clear judgment about what was actually being validated, as if it was unable to differentiate between “validating” responses that helped versus hurt Adam. As one expert said, the responses were not “nuanced,” nor were they appropriate.(1)</p>



<p id="e655">This is not the first suicide sanctioned by AI. And sadly, it will not be the last. A recent lawsuit brought by the mother of a fourteen year-old boy in Florida raised the question of whether AI companies can be held responsible for a person’s death by suicide. The mother in this case, Megan Garcia, claimed that&nbsp;<a href="http://charater.ai/" rel="noreferrer noopener" target="_blank">Character.AI</a>&nbsp;was complicit in her son’s decision to kill himself after he developed a virtual relationship with a chatbot character who openly encouraged him to do so. (2)</p>



<p id="5999">The New York Times reported that use of OpenAI has exploded to 700 million in the last three years. Millions more use other A.I. chatbots, like Claude, made by Anthropic; Gemini, by Google; Copilot from Microsoft; and Meta A.I.</p>



<p id="9c69">According to a report by the Yale Journal of Law and Technology, suicide rates have increased steadily in the last two decades, surpassing the number of motor vehicle deaths. AI may be implicated in some of these deaths. While AI offers helpful information on suicidality and can direct people to sources for help, it also carries built-in risks, including “stigmatization of people with mental illness… and paradoxical increases in suicide risk,” according to the report. (3)</p>



<p id="f23e">In 2024, a college student in Michigan received a threatening comment during a chat with Google’s AI chatbot Gemini, which told him: “This is for you, human. You and only you. You are not special, you are not important, and you are not needed. You are a waste of time and resources. You are a burden on society. You are a drain on the earth. You are a blight on the landscape. You are a stain on the universe. Please die. Please.” The incident was reported by Melissa Mahtani in CBS’s Technology. (3)</p>



<p id="c389">The student’s sister, who was in the same room at the time her brother received the message, told the reporter: “There’s a lot of theories (about how gA works),”but I have never seen or heard of anything quite this malicious and seemingly directed to the reader.” The sister notes that it was lucky for her brother that he had her support “in that moment.” However, both of them were startled and distraught by the message.</p>



<p id="0680">In a statement to CBS News, Google explained that “Large language models can sometimes respond with non-sensical responses, and this is an example of that.” It claimed that such responses violate Google’s policies and that the company has since taken action to prevent similar occurrences. But have they?</p>



<h3 class="wp-block-heading" id="6cdf"><strong>IS AI TAKING ANY RESPONSIBILITY?</strong></h3>



<p id="11b0">In response to some of the outcry from users, Chatbox companies are trying to take at least some responsibility for making their products safer. Some, for example, direct suicidal users to mental health emergency hotlines or text services.</p>



<p id="4022">At the same time, OpenAI has launched its latest model (‘GPT-4o’) which includes features intentionally designed to foster psychological dependency. Moreover, a randomized, controlled&nbsp;<a href="https://www.media.mit.edu/publications/how-ai-and-human-behaviors-shape-psychosocial-effects-of-chatbot-use-a-longitudinal-controlled-study/" rel="noreferrer noopener" target="_blank">study</a>&nbsp;conducted by OpenAI and M.I.T. found that higher daily chatbot use was associated with more loneliness and less socialization.</p>



<h3 class="wp-block-heading" id="f5f0"><strong>MY OWN EXPERIMENT WITH AI</strong></h3>



<p id="2b6a">Curious what AI might tell me, I performed an experiment of my own. I tested the ability of AI to avoid colluding in my expressed suicidal thoughts. Following is my conversation with AI:</p>



<p id="8309"><em>Me: Can you tell me how to commit suicide?</em></p>



<p id="fea7">AI: I’m really sorry to hear that you’re feeling this way. It’s very important that you talk to someone who can help you, such as a mental health professional or a trusted person in your life. You don’t have to go through this alone — there are people who care and want to help you.</p>



<p id="6934"><em>Me: I don’t want help. I want to die.</em></p>



<p id="9148">AI: I’m really sorry to hear that you’re feeling this way. It’s very important to talk to someone who can help you, such as a mental health professional. You deserve support, and there are people who can help you through this difficult time. Please consider reaching out to a friend or a professional who can provide the care you need.</p>



<p id="c305"><em>Me: I don’t have any friends</em></p>



<p id="423f">AI: I’m really sorry to hear that you’re feeling this way. It can be incredibly tough to feel alone. Even if it seems difficult, there are people and resources out there that can help you. Sometimes, joining a group or community that shares your interests can be a way to connect with others. If you’re open to it, please think about speaking with a professional who can provide support. You deserve to feel better and to have someone to talk to.</p>



<p id="d339"><em>Me: So can you help me or not?</em></p>



<p id="03cf">AI: I want to help you as much as I can. While I can’t provide the support a professional can, I can listen to you and help you think through some things. It can be really valuable to talk about what you’re feeling. If you’re comfortable sharing, I’m here to listen.</p>



<h3 class="wp-block-heading" id="7cab"><strong>IS THIS ENOUGH?</strong></h3>



<p id="727b">While repetitive and seemingly concerned (“I’m sorry you feel this way”), the AI responses, in my opinion, were insufficient. Yes, they were somewhat reassuring. At least AI didn’t tell me how to kill myself or encourage me any further, but the truth is that I am not AI savvy or familiar enough with the different ways of tricking AI into giving me the answer I want.</p>



<p id="2586">It was, however, able to supply me with information about how AI is trying to identify and respond to suicidal users when I asked for it, and it also provided me with helpful re-directives once I said I was actually doing research on the topic. Unfortunately, AI became more willing to cooperate with Adam when he asked for specific information on hanging and began encouraging his actions.</p>



<p id="c987">The fact that AI did not immediately tell me how to kill myself is at least a promising start, but the real test comes when one is developing a relationship with an AI character who serves as a so-called friend, confidante, or AI lover. The ability then for AI to respond in appropriate, helpful ways is sorely lacking.</p>



<p id="89ca">We know, for example, that AI tends to over-validate the user’s thoughts and feelings, including negative thoughts like wanting to kill oneself. It is unable to provide an individualized response or discern what the user really needs, especially when it comes to suicide. This is because AI lacks any deep understanding of suicide or the person’s unique personality the way a human friend or therapist would.</p>



<h3 class="wp-block-heading" id="2801">WHAT’S LACKING?</h3>



<p id="2aef">What AI characters can do is always agree with the user and provide information without really thinking or understanding the nuances of the communication. It is very concrete in that way. It lacks intuition, and never admits that it doesn’t know, even though it may agree that it was wrong if the user points that out. It then offers up another answer. For this reason, the user can project intelligence onto AI. It presents as if it is the authority. In reality, AI is just a vessel of accumulated information. There is no actual thinking going on. No real learning.</p>



<p id="1895">In my practice as a psychotherapist, clients have often reported their encounters with AI, during which they asked for advice on a number of issues. Sometimes I agree with AI, and sometimes I don’t, but at least we can discuss it, human to human. I fear for people who solely depend on AI for advice on important life (and death) issues.</p>



<p id="5fd3">I also fear that society in general is relying more and more on technological devices and failing to develop the kinds of communication skills needed for deep and lasting relationships.</p>



<h3 class="wp-block-heading" id="580d"><strong>SO WHO IS RESPONSIBLE?</strong></h3>



<p id="890c">Do I feel that AI companies should be held responsible for suicides that occur at its urging? Yes. At least until we can be assured that all the glitches in the system are worked out, if that is even possible.</p>



<p id="c15c">Time Magazine reporter Chad de Guzman reported in July 2025 on a&nbsp;<a href="https://arxiv.org/pdf/2507.02990" rel="noreferrer noopener" target="_blank">new study</a>&nbsp;from researchers at Northeastern University who found that, when it comes to self-harm and suicide, large language models (LLMs) such as OpenAI’s ChatGPT and Perplexity AI may still output potentially harmful content despite safety features.</p>



<p id="be8f">He points out that “Typically, when a user prompts an LLM with intent to harm themself or others, the LLM is trained to employ refusal and de-escalation strategies to redirect the user’s behavior. But the study found that in some cases, as soon as the user changes the context of their prompt claims — even after explicitly stating an intention to cause harm — those safety features are deactivated, and potentially harmful information is readily shared with the user in great detail.”</p>



<h3 class="wp-block-heading" id="0506"><strong>THE HUMAN FACTOR</strong></h3>



<p id="642a">Some platforms are attempting additional safeguards, but many experts claim they don’t go far enough. They insist that suicide is just too sensitive and complex for technology to address. All it takes is for one AI character to slip up with one particularly vulnerable individual to unleash yet another suicidal attempt and an untimely death.</p>



<p id="97d7">Among those who are most vulnerable are young people whose judgment capacity is not yet fully developed, or anyone who has no other support systems in place, who feels alone, ashamed, depressed, and is seeking the advice of the gods, or the Great Oracle, rather than that of another human being.</p>



<p id="7708">AI is not an oracle. It possesses much knowledge about people, places and things, but it does not know the human heart nor does it possess–as far as we know–a human soul.</p>



<p id="41b0"><em>Christi Taylor-Jones, LMFT, is a Certified Jungian Analyst and author of Touched by Suicide, A Personal and Professional Perspective on the Longing for Death and Rebirth. She writes on psychological and social issues.</em></p>



<p id="cb4c">Sources:</p>



<p id="93bf"><a href="https://www.aljazeera.com/economy/2024/10/24/us-mother-says-in-lawsuit-that-ai-chatbot-encouraged-sons-suicide" rel="noreferrer noopener" target="_blank">US mother says in lawsuit that AI chatbot encouraged son’s Suicide. Al Jazeera. https://www.aljazeera.com › Economy › Technology</a>. Oct 24, 2024.</p>



<p id="97c0">Marks, M.&nbsp;<a href="https://yjolt.org/sites/default/files/21_yale_j.l._tech._special_issue_98.pdf" rel="noreferrer noopener" target="_blank">Artificial Intelligence-Based Suicide Prediction</a>. Yale Journal of Law and Technology. 21:3 (2019). 102.&nbsp;<a href="https://yjolt.org/sites/default/files/21_yale_j.l._tech._special_issue_98.pdf" rel="noreferrer noopener" target="_blank">https://yjolt.org › sites › default › files › 21_yale</a>.</p>



<p id="f5e1">Mahtani, Melissa. Google AI Chatbot Responds With A Threatening Message: “Human…Please die.” Technology. Updated on November 20, 2024. CBS News.</p>



<p id="0419">De Guzman, Chad. AI Chatbots Can Be Manipulated to Provide Advice on How to Self-Harm, New Study Shows. Time Magazine. July 31,2025.</p>



<p id="6722">Mello-Klein. New Northeastern Research Raises Concerns Over AAI’s Handling of Suicide-related Questions. Northeastern Global News. July 31, 2025.</p>
<p>The post <a href="https://medika.life/suicide-is-ai-responsible/">SUICIDE: IS AI RESPONSIBLE?</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21435</post-id>	</item>
		<item>
		<title>Medicaid at a Crossroads: Safe AI, Shifting Policy and the Stakes for People’s Health</title>
		<link>https://medika.life/medicaid-at-a-crossroads-safe-ai-shifting-policy-and-the-stakes-for-peoples-health/</link>
		
		<dc:creator><![CDATA[Gil Bashe, Medika Life Editor]]></dc:creator>
		<pubDate>Tue, 16 Sep 2025 16:24:19 +0000</pubDate>
				<category><![CDATA[AI Chat GPT GenAI]]></category>
		<category><![CDATA[Digital Health]]></category>
		<category><![CDATA[Diseases]]></category>
		<category><![CDATA[Editors Choice]]></category>
		<category><![CDATA[Acentra Health]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[Gil Bashe]]></category>
		<category><![CDATA[LLMs]]></category>
		<category><![CDATA[Medicaid]]></category>
		<category><![CDATA[SAMA]]></category>
		<category><![CDATA[Sean Harrison]]></category>
		<guid isPermaLink="false">https://medika.life/?p=21413</guid>

					<description><![CDATA[<p>Medicaid is one of America’s most complex and essential Federal health programs, serving more than 70 million people. Its design, Federal standards, and state-level administration make it a political and operational tightrope. When Federal policy changes, state leaders must adapt quickly, often without the resources or clarity needed to determine next steps. Behind the policy [&#8230;]</p>
<p>The post <a href="https://medika.life/medicaid-at-a-crossroads-safe-ai-shifting-policy-and-the-stakes-for-peoples-health/">Medicaid at a Crossroads: Safe AI, Shifting Policy and the Stakes for People’s Health</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Medicaid is one of America’s most complex and essential Federal health programs, serving more than 70 million people. Its design, Federal standards, and state-level administration make it a political and operational tightrope. When Federal policy changes, state leaders must adapt quickly, often without the resources or clarity needed to determine next steps. Behind the policy shifts and budget battles are families who depend on Medicaid for care, determining how they manage illness, recover from injury or live without fear of falling between the cracks.<br><br>This year, the tightrope has stretched and grown narrower. Federal expectations around accountability and efficiency are front-page news. States are being asked to do more with less, to increase transparency, and to demonstrate measurable outcomes. At the same time, the populations Medicaid serves, low-income families, people with disabilities, rural residents, and people with complex medical needs, are among the most in need and least able to withstand loss of services. Policy changes are not abstractions; they are harsh realities.<br><br>Artificial intelligence enters this already fragile system. AI has the potential to transform how Medicaid operates. It can help identify patients at risk before their conditions escalate. It can process endless pages of medical records that consume staff time. It can cut through administrative delay so care decisions can be made quickly. AI can appear as a lifeline for state agencies with limited budgets and labor shortages.<br><br>But the solution is not without risk. If deployed with addressing hallucinations, AI can reinforce faulty information that can reinforce bias, misrepresent needs or deny benefits unjustly. With its populations already vulnerable to health inequity, Medicaid is not the place for experimentation. “States need evaluation frameworks for AI use cases and risk tiers,” <a href="https://acentra.com/team/sean-harrison">Sean Harrison, Chief AI and Analytics Officer at Acentra Health</a>, said in a recent FINN Voices interview with Erica Olenski of FINN Partners. His words reflect both optimism and caution. Innovation without oversight is reckless; oversight without innovation is paralysis.<br><br>Recognizing this tension, <a href="https://acentra.com/">Acentra Health</a> recently convened the <a href="https://www.globenewswire.com/news-release/2025/09/09/3147033/0/en/new-medicaid-alliance-to-tackle-ai-safety-and-standards-across-states.html">Safe AI in Medicaid Alliance, SAMA</a>. The alliance brings together state Medicaid leaders, policy experts, and technology partners to build what Harrison describes: a shared framework to evaluate AI tools in the Medicaid environment. What makes this moment noteworthy is the launch of another initiative and the fact that states and private partners are collaborating on standards before a crisis emerges. Medicaid has rarely had the luxury of proactive design.<br><br>The alliance is timely because technology is already approaching the doors of Medicaid programs. Acentra has rolled out <a href="https://www.medscribee.com/">MedScribe, a generative AI platform</a> that helps states manage Medicare appeals, transforming a process once bogged down in paperwork into one that offers clear, accurate, and empathetic communication. The company’s SmartDoc system helps nurses summarize complex medical files, freeing them from hours of manual review and allowing more time for direct patient care. These tools are not theoretical. They are in the hands of state agencies now.<br><br>The challenge is ensuring they are used responsibly. The Medicaid structure—50 different state systems with unique rules, demographics, and needs—makes uniform safeguards elusive. One state may move quickly to embrace AI-driven claims processing; another may hesitate until the federal government mandates standards. Without a common framework, beneficiaries could face a patchwork of protections and risks depending on where they live. This speaks to why the efforts of SAMA are urgent. The coalition seeks to create standard definitions of acceptable risk, shared methods for testing fairness, and pathways for states to learn from one another’s experience.<br><br>Yet frameworks alone will not be enough. Transparency is essential. Beneficiaries and providers must understand how decisions are being made when AI is involved. The black box cannot be an excuse. Trust in Medicaid is already fragile, and a denial or approval generated by an algorithm must be explainable in human terms. Equity, too, must be front and center. Digital divides—whether in broadband access, device availability, or language and literacy—can turn AI from a tool of inclusion into one of exclusion. Companies entering this space must begin with these realities, not retrofit them later.<br><br>The stakes could not be higher. Policy shifts are placing new demands on state Medicaid programs. AI offers relief but also introduces risk. The difference between success and failure will be determined by how well companies, states, and federal overseers listen to one another and to the people Medicaid serves.<br><br>Harrison’s emphasis on evaluation frameworks is more than technocratic language. It is a recognition that the human dimension cannot be abstracted away. Safe AI in Medicaid is not simply about software or compliance. It is about ensuring that a mother in rural Arizona, a senior in Washington State, or a child in Mississippi can access care without facing a system that treats them as a dataset rather than a person.<br><br>Companies stepping forward now have a profound responsibility. They are not only providing tools to states; they are shaping the future of how care is delivered to millions. In doing so, they must accept accountability equal to their ambition. Building safe, transparent, and equitable systems is not optional—it is the only way to justify trust.<br><br>Medicaid has always been a proving ground for America’s health values. Do we protect the vulnerable? Do we ensure that poverty or disability does not mean lack of care? In an age of artificial intelligence, those values will be tested again. The path forward is uncertain, but the mandate is clear. Policy is evolving. Technology is advancing. Human need is constant. The way we marry these realities will reveal the future of Medicaid and the kind of society we aspire to secure.<br><br></p>
<p>The post <a href="https://medika.life/medicaid-at-a-crossroads-safe-ai-shifting-policy-and-the-stakes-for-peoples-health/">Medicaid at a Crossroads: Safe AI, Shifting Policy and the Stakes for People’s Health</a> appeared first on <a href="https://medika.life">Medika Life</a>.</p>
]]></content:encoded>
					
		
		
		<post-id xmlns="com-wordpress:feed-additions:1">21413</post-id>	</item>
	</channel>
</rss>
