Gabriel Ayuso pfp
Gabriel Ayuso
@gabrielayuso.eth
Why do people think context window size is a big limitation for AIs to be able to fully replace humans as coders? I'd say most of us keep a very compressed version of the codebase in our head and rely on code search and tools to get the details we need. Coding agents are already starting to do this using grep and other tools. They just need that initial context to get them going and then they can do the rest. Creating that initial compressed representation of the codebase they can reference to get them started would do the trick without keeping the entire codebase in the context.
9 replies
2 recasts
54 reactions

Lokp Ray pfp
Lokp Ray
@lokpray
ikr... It used to be the case that context window size was a big deal when chatgpt was first released but with all the recently development in the past 1-2 years including but not limited to: - RAG for live code search - LoRA & fine-tuning - Agentic workflows: SWE-agent, MCP - Context caching In an agentic future (no hype intended here), focus has been more on various methods of accessing live raw data than eating everything up from context window for processing Also, we have 1M+ context window these days from Google/Meta and it's going to go (prob exponentially) as they scale up the number of transformer layers in the NN (aka just more compute needed for initial training)
1 reply
1 recast
3 reactions

polymutex pfp
polymutex
@polymutex.eth
Is there good tooling around automatically generating this compressed representation?
1 reply
0 recast
1 reaction

Royal pfp
Royal
@royalaid.eth
AFAIK the AIs so much better with a large context window because everything has too, at one point, live there. We shouldn't be adapting the problem to the tools but the tool to the problem. LLMs have no fundamental reason they can't pull the whole codebase into context, it's just hardware limits right now.
1 reply
0 recast
0 reaction

Mo pfp
Mo
@meb
Agreed. Thinking we need to pass all the full files content to AI agents is a recurring midwit take I’ve seen. Any real system is a collection of interfaces that work with each other, whilst encapsulating their messy implementation code. You only really need - an overview of system components and their relations - access to a db schema - file content related to task at hand as well as existing file examples to facilitate 0 shot success of the agent
0 reply
0 recast
1 reaction

llamafacts pfp
llamafacts
@llamafacts.eth
I've spent a few hours already polishing cursor rules for my project. I include them with each request (it's a couple of pages). Works great but it required a few iterations to get them to a point where I feel that I can ask cursor in plain English what I need and it just does it exactly as I expect. I'm like 90% there
0 reply
0 recast
1 reaction

Darryl Yeo 🛠️ pfp
Darryl Yeo 🛠️
@darrylyeo
I think this explains why I haven’t found Claude Opus to be that much better than Claude Sonnet despite being like fifty times more expensive. Cursor is already doing a lot of legwork summarizing context as it goes, so it can keep a small context window cached no matter how long the conversation gets.
0 reply
0 recast
1 reaction

vrypan |--o--| pfp
vrypan |--o--|
@vrypan.eth
I think the analogy is - dna: model - everything you ever learned, remember, sensed: context I may be wrong.
0 reply
0 recast
0 reaction

Frank pfp
Frank
@deboboy
Claire Vo talking about pairing with a 91 year old vibe coder reimagines "codebase in our head"... it's possible deep human memory generates circuit connections between the LLM' interpretability and the coder's hidden features.
0 reply
0 recast
0 reaction

cashlessman 🎩 pfp
cashlessman 🎩
@cashlessman.eth
can you enable new beta DC API for me please
0 reply
0 recast
0 reaction