Hacker News
new
|
past
|
comments
|
ask
|
show
|
jobs
|
submit
login
Alifatisk
5 months ago
|
parent
|
context
|
favorite
| on:
DeepSeek-R1
This is impressive, how do people handle the limited context window of 64k tokens?
int_19h
5 months ago
|
next
[–]
Same as they did it back in the "old days" when GPT-4 was 8k and LLaMA was 2k. Chunking, RAG etc, then cross your fingers and hope that it all works reasonably well.
m3kw9
5 months ago
|
prev
[–]
By using o1
Consider applying for YC's Fall 2025 batch! Applications are open till Aug 4
Guidelines
|
FAQ
|
Lists
|
API
|
Security
|
Legal
|
Apply to YC
|
Contact
Search: