Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Expand scratch size to accommodate 2k context. #61

Merged
merged 6 commits into from
Jul 22, 2023
Merged

Expand scratch size to accommodate 2k context. #61

merged 6 commits into from
Jul 22, 2023

Conversation

li-plus
Copy link
Owner

@li-plus li-plus commented Jul 22, 2023

This should fix #29. Users shouldn't worry about scratch size. I just make it large enough for context up to 2k tokens.

@li-plus li-plus merged commit 61f6e88 into main Jul 22, 2023
@li-plus li-plus deleted the fix-scratch branch July 22, 2023 10:07
@volagold
Copy link

volagold commented Aug 1, 2023

2k context size is far too limited for industrial applications......for example, we might want to analyze large blocks of contexts with 5k~10k+ sizes from long legal documents.

@volagold
Copy link

volagold commented Aug 1, 2023

Can we make it to 32k? so that we can take advantage of this large context length provided by the ChatGLM2-6B model.

@li-plus
Copy link
Owner Author

li-plus commented Jun 21, 2024

Fixed in #305. With the latest release (v0.4.0), memory allocation is on demand. No need to preset max scratch size and memory size.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Allow configuring SCRATCH_SIZE (not enough space in the scratch memory pool)
2 participants