Documentation about Llama 3.2 11B Vision Instruct Model says 128K context window but not able to process more than 8k tokens

Maheshbabu Boggu 0 Reputation points
2025-01-23T12:07:14.96+00:00

I am writing to inquire about the context window of the Llama 3.2 11B Vision Instruct model.

The documentation states that the context window is 128K tokens. However, when using the model, I am unable to provide input exceeding 8192 tokens. I would appreciate it if you could clarify this discrepancy and provide guidance on how to utilize the full 128K context window.

Thank you for your time and assistance.

Azure OpenAI Service
Azure OpenAI Service
An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
3,582 questions
Azure AI services
Azure AI services
A group of Azure services, SDKs, and APIs designed to make apps more intelligent, engaging, and discoverable.
3,072 questions
{count} votes

Your answer

Answers can be marked as Accepted Answers by the question author, which helps users to know the answer solved the author's problem.