Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

This is impressive, how do people handle the limited context window of 64k tokens?



Same as they did it back in the "old days" when GPT-4 was 8k and LLaMA was 2k. Chunking, RAG etc, then cross your fingers and hope that it all works reasonably well.


By using o1




Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: