Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Increase Total tokens to 128K (currently 4K) #1186

Closed
Padmaapparao opened this issue Nov 23, 2024 · 6 comments
Closed

Increase Total tokens to 128K (currently 4K) #1186

Padmaapparao opened this issue Nov 23, 2024 · 6 comments
Assignees
Milestone

Comments

@Padmaapparao
Copy link

Padmaapparao commented Nov 23, 2024

For Doc Sum example as we will upload 100's of files, we need the input token length to be large and same with the output. Currently it is fixed at 4096 total, so if we upload even 1 large file, output token length for summarization will be only 32 tokens which is very very small for a summary.

need total tokens 128K, so we can get at least 16K-32K summary.
These values are hardcoded in compose.yaml. Need them to be parametrizable.

@lvliang-intel
Copy link
Collaborator

We are considering adding these parameters and making them configurable in compose.yaml to support flexible setups. A PR will be created for this, and we will update the details here once the PR is ready.

However, the models themselves currently don't support a 256K context length. And some hardware also have limitation to support large input token length and max output token length. We recommend exploring alternative approaches, such as chunking files or using recursive summarization techniques, to achieve optimal results within the current technical limitations.

@lvliang-intel lvliang-intel self-assigned this Nov 27, 2024
@Padmaapparao Padmaapparao changed the title Increase Total tokens to 256K (currently 4K) Increase Total tokens to 128K (currently 4K) Nov 27, 2024
@Padmaapparao
Copy link
Author

Padmaapparao commented Dec 19, 2024 via email

@yongfengdu
Copy link
Collaborator

@Padmaapparao Once a PR was created and referred to/mentioned this issue, it will link automatically.

@eero-t
Copy link
Contributor

eero-t commented Dec 30, 2024

(Already merged) PRs for supporting longer documents (with current small token amounts are):

@joshuayao joshuayao added this to the v1.3 milestone Feb 25, 2025
@joshuayao
Copy link
Collaborator

joshuayao commented Feb 25, 2025

Hi @Padmaapparao, as @eero-t mentioned, OPEA now offers multiple strategies to support long contexts for DocSum, including auto, stuff, truncate, map_reduce, and refine. Please refer to the section MegaService with long context of the doc for more details. Could we proceed with closing this issue if these PRs align with your requirements?

@joshuayao
Copy link
Collaborator

Closed for no active responses in the recent 30 days. Please feel free to reopen it if the PRs do not resolve the issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants