I'm curious,what is limiting the context length windows

Your model here are running at insane speed,which is really cool.

But it only has max context window of 32k,which a max output 8k.Which is even a bit unsufficient for complex single problem.As you know DeepSeek R1 is known for outputing lots of tokens while thinking,and your speed will fill 8k within seconds.

I’m wondering what is limiting that number.If you can trigger it to 128k,as the normal NNML model,then it will start to have lots of application.

Is your way of running the model having the drawback of context window?

Hi @haveyuner ,
We appreciate you bringing this to our attention and will investigate the matter further. We will get back to you with our findings.

Regards,
Shivani Moze

Hi @haveyuner ,

Thanks for your feedback on DeepSeek R1. We’ve forwarded your request to our product management team for consideration.

Also, welcome to the SambaNova Community! We’re glad to have you here and look forward to your continued participation and insights.