Rather than having multiple models with 32K context length, I think it would be more effective to have at least one coder model with an extended context length.
For example, 'llama 3.2 3b' seems to have a good context length, but it's not particularly useful for code and math.
On the other hand, 'llama 3.1 405b' has an 'ok' context length, but it becomes slow as the context grows.
In contrast, 'qwen2.5-coder 32b' can handle up to 128K context length, but Venice is currently using 32K.
My suggestion is:
a) Increase the context length in the current 'qwen2.5-coder 32b' model.
b) And/or consider adding a smaller version with a higher context length, such as 'qwen2.5-coder 7b'.
Please authenticate to join the conversation.
Completed
Feature Requests
About 1 year ago

An Anonymous User
Get notified by email when there are changes.
Completed
Feature Requests
About 1 year ago

An Anonymous User
Get notified by email when there are changes.