Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Optimization of non-GPT 4 major model outputs #548

Closed
warlockedward opened this issue Mar 29, 2024 · 1 comment
Closed

Optimization of non-GPT 4 major model outputs #548

warlockedward opened this issue Mar 29, 2024 · 1 comment

Comments

@warlockedward
Copy link

Currently I try to use Qwen1.5-72B, deepseek-33b, and mixtral-8x7b models to drive mentat, and I find that the answers given always have some errors, there are misunderstandings and inaccurate code modifications, and I'm not sure what is causing them, and I don't know if there is any later support for the ability to target non-GPT4 models? Thank you very much.

@biobootloader
Copy link
Member

in our testing, no models other than GPT-4 and Claude 3 Opus can handle the complex edit format required for Mentat.

We do have some changes coming that might make things easier for local models though. This experiment is a step in that direction: #530

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants