Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: Self-Extend support #349

Closed
hahuyhoang411 opened this issue Jan 16, 2024 · 0 comments · Fixed by #351
Closed

feat: Self-Extend support #349

hahuyhoang411 opened this issue Jan 16, 2024 · 0 comments · Fixed by #351
Labels
P3: nice to have Nice to have feature type: feature request A new feature

Comments

@hahuyhoang411
Copy link
Contributor

Problem
This is new implementation of Self-Extend to make LLM extend to 8k or 16k without further training

Success Criteria
We can also use in Nitro

--grp-attn-n 2
--grp-attn-w 128

Additional context
The llama.cpp supports it: ggerganov/llama.cpp#4815

@hahuyhoang411 hahuyhoang411 added P3: nice to have Nice to have feature type: feature request A new feature labels Jan 16, 2024
@tikikun tikikun linked a pull request Jan 16, 2024 that will close this issue
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
P3: nice to have Nice to have feature type: feature request A new feature
Projects
None yet
Development

Successfully merging a pull request may close this issue.

1 participant