Increasing Context Length for Coder Models

Rather than having multiple models with 32K context length, I think it would be more effective to have at least one coder model with an extended context length.

For example, 'llama 3.2 3b' seems to have a good context length, but it's not particularly useful for code and math.

On the other hand, 'llama 3.1 405b' has an 'ok' context length, but it becomes slow as the context grows.

In contrast, 'qwen2.5-coder 32b' can handle up to 128K context length, but Venice is currently using 32K.

My suggestion is:

a) Increase the context length in the current 'qwen2.5-coder 32b' model.

b) And/or consider adding a smaller version with a higher context length, such as 'qwen2.5-coder 7b'.

Please authenticate to join the conversation.

Upvoters
Status

Completed

Board
💡

Feature Requests

Date

About 1 year ago

Author

An Anonymous User

Subscribe to post

Get notified by email when there are changes.