LLama is supposed to have 2048 tokens but i keep getting error for more than 512 tokens . I am prompting to write summary of text which has greater lengths, any solution
Increase your context size. Try passing n_ctx=4096 to the Llama()
n_ctx=4096
it works , thanks
Β· Sign up or log in to comment