Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Missing keys in state_dict #2

Open
kuraga opened this issue Sep 7, 2021 · 3 comments
Open

Missing keys in state_dict #2

kuraga opened this issue Sep 7, 2021 · 3 comments

Comments

@kuraga
Copy link

kuraga commented Sep 7, 2021

Hello, @Alenush !

According to error, some weights are missing in .pth. (Missing keys in state_dict: ...) Could model change?

Thanks!

@kuraga
Copy link
Author

kuraga commented Sep 12, 2021

Well applying ai-forever/ru-gpts#42 (comment) let me load the model.

But on model apply I get ai-forever/ru-gpts#42 (comment) itself:

<...>

/usr/local/lib/python3.8/dist-packages/deepspeed/ops/sparse_attention/matmul.py in _sdd_matmul(a, b, trans_a, trans_b, trans_c, spdims, block, luts, num_locks, widths, packs, bench, time)                                                   
    202             total = 0 if bench else None                                                                                                                                                                                              
    203             for off_width in range(0, width, max_width):                                                                                                                                                                              
--> 204                 current = kernel(                                                                                                                                                                                                     
    205                     a.data_ptr(),                                                                                                                                                                                                     
    206                     b.data_ptr(),                                                                                                                                                                                                     
                                                                                                                                                                                                                                              
/usr/local/lib/python3.8/dist-packages/triton/kernel.py in __call__(self, grid, *args)                                                                                                                                                        
    114         # pack parameters into a byte buffer                                                                                                                                                                                          
    115         params = struct.pack(self.tys, *args)                                                                                                                                                                                         
--> 116         kernel = self.fn.autotune(params, grid, self.stream)                                                                                                                                                                          
    117         # run kernel
    118         grid = grid(kernel.opt)

RuntimeError: CUDA: Error- invalid ptx

@Alenush
Copy link
Owner

Alenush commented Sep 27, 2021

Hi!

My only guess that during uploading of the checkpoint smth went wrong.
I uploaded again checkpoint in another source. Please, try here:
https://sc.link/Dk6Y

@kuraga
Copy link
Author

kuraga commented Oct 13, 2021

I'm facing ai-forever/ru-gpts#68. After applying ai-forever/ru-gpts#68 (comment) error dissapears but I'm facing out-of-memory error.

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants