We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Describe the bug
The code in the section: https://www.deepspeed.ai/tutorials/inference-tutorial/#initializing-for-inference appears to be incorrect.
Specifically, in the first code block of that section
model = ds_engine.module output = model('Input String')
should probably be:
model = ds_engine.module pipe = pipeline("text-generation", model=model, tokenizer=tokenizer) print(f"{pipe('Hello')}")
(or something similar). The code in https://www.deepspeed.ai/tutorials/inference-tutorial/#end-to-end-gpt-neo-27b-inference is closer to what's required.
To Reproduce
I attempted to use the code sample as the basis for my own script.
Expected behavior
Code sample to work
ds_report output Please run ds_report to give us details about your setup.
ds_report
torch install path ............... ['/azureml-envs/azureml_28a03b9f9b7c5f401fdd179f0d3ee4d8/lib/python3.10/site-packages/torch'] torch version .................... 2.6.0+cu124 deepspeed install path ........... ['/azureml-envs/azureml_28a03b9f9b7c5f401fdd179f0d3ee4d8/lib/python3.10/site-packages/deepspeed'] deepspeed info ................... 0.16.4, unknown, unknown torch cuda version ............... 12.4 torch hip version ................ None nvcc version ..................... 12.1 deepspeed wheel compiled w. ...... torch 0.0, cuda 0.0 shared memory (/dev/shm) size .... 2.00 GB
Screenshots If applicable, add screenshots to help explain your problem.
System info (please complete the following information):
Docker context Are you using a specific docker image that you can share?
Additional context
Running in AzureML.
The text was updated successfully, but these errors were encountered:
c07b635
Improve inference tutorial docs (deepspeedai#7083)
7357edc
Fixes: deepspeedai#7082 --------- Signed-off-by: Logan Adams <[email protected]> Signed-off-by: Max Kovalenko <[email protected]>
5e379ad
Fixes: deepspeedai#7082 --------- Signed-off-by: Logan Adams <[email protected]> Signed-off-by: gyou2021 <[email protected]>
Improve inference tutorial docs (#7083)
697050e
Fixes: #7082 --------- Signed-off-by: Logan Adams <[email protected]> Signed-off-by: Masahiro Tanaka <[email protected]>
loadams
Successfully merging a pull request may close this issue.
Describe the bug
The code in the section:
https://www.deepspeed.ai/tutorials/inference-tutorial/#initializing-for-inference
appears to be incorrect.
Specifically, in the first code block of that section
should probably be:
(or something similar). The code in
https://www.deepspeed.ai/tutorials/inference-tutorial/#end-to-end-gpt-neo-27b-inference
is closer to what's required.
To Reproduce
I attempted to use the code sample as the basis for my own script.
Expected behavior
Code sample to work
ds_report output
Please run
ds_report
to give us details about your setup.Screenshots
If applicable, add screenshots to help explain your problem.
System info (please complete the following information):
Docker context
Are you using a specific docker image that you can share?
Additional context
Running in AzureML.
The text was updated successfully, but these errors were encountered: