<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Agents on Mohamed Abdelrahman</title><link>https://mkabdelrahman.github.io/tags/agents/</link><description>Recent content in Agents on Mohamed Abdelrahman</description><generator>Hugo</generator><language>en</language><lastBuildDate>Mon, 04 May 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://mkabdelrahman.github.io/tags/agents/index.xml" rel="self" type="application/rss+xml"/><item><title>An AI tutor that watches every keystroke</title><link>https://mkabdelrahman.github.io/posts/lernen-ai-tutor-watching-keystrokes/</link><pubDate>Mon, 04 May 2026 00:00:00 +0000</pubDate><guid>https://mkabdelrahman.github.io/posts/lernen-ai-tutor-watching-keystrokes/</guid><description>&lt;h2 id="a-flashcard-app-but-the-ai-is-in-the-loop">
 A flashcard app, but the AI is in the loop
 &lt;a class="heading-link" href="#a-flashcard-app-but-the-ai-is-in-the-loop">
 &lt;i class="fa-solid fa-link" aria-hidden="true" title="Link to heading">&lt;/i>
 &lt;span class="sr-only">Link to heading&lt;/span>
 &lt;/a>
&lt;/h2>
&lt;p>Most &amp;ldquo;AI study buddies&amp;rdquo; are chatbots glued to the side of a flashcard app. You study, then you ask a question, then the model answers from cold context. The model has no idea you just got the same word wrong twice and replayed the audio four times.&lt;/p>
&lt;p>I wanted to flip that. The model should see every action — reveals, replays, grades, notes — and react inside the study loop, not after it.&lt;/p></description></item><item><title>[Draft] Compaction Algorithms for Long-Running Agent Sessions</title><link>https://mkabdelrahman.github.io/posts/agent-context-compaction-strategies/</link><pubDate>Thu, 30 Apr 2026 00:00:00 +0000</pubDate><guid>https://mkabdelrahman.github.io/posts/agent-context-compaction-strategies/</guid><description>&lt;h2 id="the-problem">
 The problem
 &lt;a class="heading-link" href="#the-problem">
 &lt;i class="fa-solid fa-link" aria-hidden="true" title="Link to heading">&lt;/i>
 &lt;span class="sr-only">Link to heading&lt;/span>
 &lt;/a>
&lt;/h2>
&lt;p>A long-running agent accumulates messages over time: user prompts, model responses, reasoning traces, tool calls, tool outputs. The model&amp;rsquo;s context window is finite — from a hundred thousand tokens on the smaller end to a million or more on the largest current frontier models. Once the conversation no longer fits, continuing without intervention triggers a hard error or silent truncation by the provider.&lt;/p></description></item></channel></rss>