Files
insightface/cpp-package/inspireface/docker/Dockerfile.cuda12_ubuntu22
2025-05-22 16:07:26 +08:00

48 lines
1.6 KiB
Docker

# Use Ubuntu 18.04 as the base image
FROM nvidia/cuda:12.0.0-devel-ubuntu22.04
ARG https_proxy
ARG http_proxy
ARG all_proxy
RUN apt-get update
# Update the package list and install basic development tools
RUN apt-get install -y --no-install-recommends build-essential
RUN apt-get install -y --no-install-recommends software-properties-common
RUN apt-get install -y --no-install-recommends curl
RUN apt-get install -y --no-install-recommends wget
# Install CMake
RUN apt-get install -y --no-install-recommends
# Install CMake 3.20.6
RUN cd /opt && \
wget https://github.com/Kitware/CMake/releases/download/v3.20.6/cmake-3.20.6-linux-x86_64.tar.gz && \
tar -zxvf cmake-3.20.6-linux-x86_64.tar.gz && \
ln -s /opt/cmake-3.20.6-linux-x86_64/bin/* /usr/local/bin/ && \
rm cmake-3.20.6-linux-x86_64.tar.gz
# Clean temporary files to reduce image size
RUN apt-get clean && \
rm -rf /var/lib/apt/lists/*
# Download TensorRT-10.8
WORKDIR /
RUN curl -L https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.8.0/tars/TensorRT-10.8.0.43.Linux.x86_64-gnu.cuda-12.8.tar.gz -o TensorRT.tar.gz && \
tar -zxvf TensorRT.tar.gz && \
rm TensorRT.tar.gz
# Set environment variables
RUN echo 'export TENSORRT_ROOT=/TensorRT-10.8.0.43/' >> /root/.bashrc && \
echo 'export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$TENSORRT_ROOT/lib' >> /root/.bashrc
# Set environment variables
ENV TENSORRT_ROOT=/TensorRT-10.8.0.43/
ENV LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$TENSORRT_ROOT/lib
# Set the working directory
WORKDIR /workspace
# Default to running Bash
CMD ["/bin/bash"]