I used docker image "flexflow/flexflow-cuda-12.1:latest" to run flexflow on a 24GB RTX 3090,but it generated a out of memory error:
import flexflow.serve as ff
ff.init(num_gpus=1, memory_per_gpu=11000, zero_copy_memory_per_node=11000)
[0 - 7fe1339374c0] 0.000000 {5}{gpu}: /usr/FlexFlow/deps/legion/runtime/realm/cuda/cuda_module.cc(4745):CUDA_DRIVER_FNPTR(cuIpcGetMemHandle)(&alloc.ipc_handle, alloc.dev_ptr) = 2(CUDA_ERROR_OUT_OF_MEMORY): out of memory
Aborted
Was it because I used the wrong code? How can I fix it?
I used docker image "flexflow/flexflow-cuda-12.1:latest" to run flexflow on a 24GB RTX 3090,but it generated a out of memory error:
Was it because I used the wrong code? How can I fix it?