When AI also has short-term memory like humans

Cát Tiên (THEO INDIANEXPRESS) |

The phenomenon of AI chatbot losing memory originates from the context window, when the model crosses the token limit and is forced to remove old information in the conversation.

AI chatbot such as ChatGPT, Gemini and Claude are becoming a popular tool in digital life. However, users often find a strange phenomenon that after a period of chatting, these models seem to los memory and forget what you have just said, or start repeating and answering incorrectly.

Regarding this issue, according to technology experts, it originates from a key technical concept called context window.

YouTuber and AI researcher Matt Pocock recently shared in his video that context windows are an important limitation, but also the most misunderstood in the way large language models (LLM) work. Simply put, it is the short-term memory of artificial intelligence.

What is a context window?

Every time a user submits a question and an answer model, the entire text will be divided into small units called token. Each token can represent a few characters or parts of a word. All the widgets in the conversation will create a context that the model can see at one point.

If the context window of a model is 200,000 units, it can only remember that much information. When it crosses this limit, older data will be deleted, causing AI to forget the beginning of the conversation.

For example, Claude 4.5 can remember up to 200,000 token, while the Gemini 2.5 Pro can process up to 2 million. In contrast, small models such as LLaMA or Mistral are limited to a few thousand.

Why can't AI have unlimited memory?

Increasing the capacity of context windows is not always feasible. Each additional account consumes computing resources and memory, causing operating costs to increase.

In addition, when the context is too big, the model finds it difficult to find the right details, like digging the needle at the bottom of the tank.

More importantly, each model is designed with fixed architectural limits, so it is impossible to limit the expansion of memory without exchanging performance.

Lost in the middle: When AI Forgets the middle of the Story

Pocock calls a typical phenomenon of context limitation midpoint forgetting, which is sometimes translated as missing in the middle.

AI often focuses on the beginning (instruction) and end ( Latest message) of a conversation, while the middle part receives less attention.

This mechanism comes from the way LLM models allocate attention to the token. Like humans, they prioritize what is happening more recently. As a result, the information in the middle, although important, is easily overlooked.

This is especially difficult for programmers. If a developer asked AI to fix the error in the code a few hundred lines ago, the model may not remember the exact part because it has slipped out of the picture.

Impact of programming AI tools

Tools like Claude Code or GitHub Copilot also work within the context window limitation. When the project or session is too long, they can easily forget orders, give incorrect feedback or stop responding.

Therefore, professional users often have to divide them into smaller, concise or reschedule the session to help AI stay focused.

The fact that chatbots like ChatGPT or Gemini are "forgotten" is not a system failure, but a natural limitation of current technology.

As technology companies continue to expand the cloud cloud and optimize storage, the future could see AI models that remember longer, understand more deeply, and get closer to real AI.

Cát Tiên (THEO INDIANEXPRESS)
RELATED NEWS

Junk data makes AI unreliable, easily creating dangerous false feedback

|

Cornell University research warns that AI models like ChatGPT can suffer from human-like brain degeneration due to long-term exposure to online waste.

India tightens AI regulations to prevent the increase of deepfake

|

India has proposed comprehensive new regulations to manage artificial intelligence (AI), aiming to prevent the increase of misinformation and deepfake videos.

AI feature with more than 8 billion visits per day is open for free

|

Snapchat officially offers Imagine Lens for free - the platform's first AI lens that allows users to create images with any prompt.

91-year-old woman in Hanoi receives support after Lao Dong Newspaper article

|

After the article by Lao Dong Newspaper, a 91-year-old woman in Hanoi has received a new allowance for January, February and March, with a level of 650,000 VND per month.

AI cameras support Hanoi traffic police to maintain traffic flow on the days leading up to Tet

|

Hanoi - AI camera data in real time helps traffic police forces flexibly operate traffic lights, remote traffic diversion...

Opposition party raises doubts in Thai election

|

Leaders of the Thai People's Party admitted defeat according to preliminary results, but called for an investigation into unusual signs to ensure the transparency of the election.

Truck bursts into flames on Vinh Hao - Phan Thiet expressway

|

Lam Dong - A truck traveling on the Vinh Hao - Phan Thiet expressway suddenly caught fire fiercely.

Large fire at a charity house in Bac Ninh

|

Bac Ninh - On the morning of February 10, a fire broke out at the Huong La charity house (Luong Tai commune), burning down about 80m2.

Junk data makes AI unreliable, easily creating dangerous false feedback

Cát Tiên (THEO INDIANEXPRESS) |

Cornell University research warns that AI models like ChatGPT can suffer from human-like brain degeneration due to long-term exposure to online waste.

India tightens AI regulations to prevent the increase of deepfake

NGUYỄN ĐĂNG |

India has proposed comprehensive new regulations to manage artificial intelligence (AI), aiming to prevent the increase of misinformation and deepfake videos.

AI feature with more than 8 billion visits per day is open for free

HẠO THIÊN |

Snapchat officially offers Imagine Lens for free - the platform's first AI lens that allows users to create images with any prompt.