Jim Groth<p>Just published a small experiment called ThinContext:<br>A simple way to reduce context bloat in LLMs by summarizing previous outputs instead of feeding everything back verbatim.</p><p>message → full reply<br>context → short summary<br>Only context gets reused.</p><p>It’s lossy, but functional.<br>Maybe a small step toward more meaningful memory in AI systems.</p><p>Repo here if you’re curious, or want to break it: <a href="https://github.com/VanDerGroot/ThinContext" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">github.com/VanDerGroot/ThinCon</span><span class="invisible">text</span></a></p><p>Critique very welcome.</p><p><a href="https://mastodon.social/tags/AI" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AI</span></a> <a href="https://mastodon.social/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://mastodon.social/tags/ContextWindow" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ContextWindow</span></a> <a href="https://mastodon.social/tags/PromptEngineering" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>PromptEngineering</span></a> <a href="https://mastodon.social/tags/OpenSource" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>OpenSource</span></a></p>