Skip to content

Select device: GPU for model eval #297

@kn0wn-cyber

Description

@kn0wn-cyber

has the error:

torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU 0 has a total capacty of 23.69 GiB of which 2.53 GiB is free. Process 3238833 has 3.43 GiB memory in use. Process 3238831 has 2.53 GiB memory in use. Process 3238830 has 2.75 GiB memory in use. Process 3238832 has 2.61 GiB memory in use. Including non-PyTorch memory, this process has 2.41 GiB memory in use. Process 3509629 has 2.41 GiB memory in use. Process 3509627 has 2.38 GiB memory in use. Process 3509630 has 2.41 GiB memory in use. Of the allocated memory 2.16 GiB is allocated by PyTorch, and 1.92 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF

How can I eval the model edumunozsala/llama-2-7b-int4-python-code-20k with cuda:0 , beacause that the cuda:0 has been occupied by other tasks. Shall I change the config by (accelerate config) ?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions