Tensor Fusion Deployment for VM/BareMetal
Files
https://r2-upload.code2life.top/CUDA-Fusion.tar
Client
Environment: Ubuntu 24.04 or Rocky Linux 9.0
Dependencies:
- Install CUDA 12.1 official package Download Link During installation, select to install CUDA Toolkit only, do not install the NVIDIA driver;
- Install Tensor Fusion client part Extract CUDA-Fusion.tar, copy all files from
CUDA-Fusion/client/libs
to the CUDA installation directory/usr/local/cuda-12.1/lib64/
- Install PyTorch, Download Link This version uses PyTorch code tag-v2.1.2 and dynamically links to the CUDA runtime.
Startup Command:
sh
./vcuda-client [Node ID] [Server IP] [Protocol Type] [Send Port] [Receive Port] [Register Port] [GPU ID]
Example:
sh
./vcuda-client 0 192.168.2.129 native 9998 9997 9999 0
Parameter Explanation:
- Node ID: Numeric ID defined for each server
- Protocol Type:
native
/light
transmission protocol, must be consistent between client and server - GPU ID: The ID of the GPU used by this configuration on the server
Note:
- The send port corresponds to the server's receive port, the receive port corresponds to the server's send port, and the register port must be consistent between client and server.
- After the
vcuda-client
successfully establishes a connection with the server, thevcuda-client
can take over all CUDA requests of the current system.
Server
Environment: Ubuntu 24.04 or Rocky Linux 9.0
Dependencies:
- Install the official NVIDIA driver:sh
apt install nvidia-driver-535
- Install CUDA 12.1 official package Download Link During installation, select to install CUDA Toolkit only, do not install the NVIDIA driver.
- Install CUDNN, Download Link Extract and place it in the
/usr/local/
directory.
Startup Command:
sh
LD_LIBRARY_PATH=/usr/local/cudnn-linux-x86_64-8.8.1.3_cuda12-archive/lib/ ./vcuda -n native -s 9997 -r 9998 -p 9999 -a 0x1129 -k 0x298
Parameter Explanation:
-n
:native
/light
transmission protocol, must be consistent between client and server-s
: Send port-r
: Receive port-p
: Register port-a
: Memory tag-k
: Kernel parameter offset (related to the server's NVIDIA driver and CUDA version)