diff --git a/docs/source/getting_started/installation.rst b/docs/source/getting_started/installation.rst index 5dfb3208..e7826114 100644 --- a/docs/source/getting_started/installation.rst +++ b/docs/source/getting_started/installation.rst @@ -85,13 +85,3 @@ You can also build and install vLLM from source: $ nvcc --version # verify that nvcc is in your PATH $ ${CUDA_HOME}/bin/nvcc --version # verify that nvcc is in your CUDA_HOME - -.. note:: - If you are developing the C++ backend of vLLM, consider building vLLM with - - .. code-block:: console - - $ python setup.py develop - - since it will give you incremental builds. The downside is that this method - is `deprecated by setuptools `_.