[Bugfix] fix lora_dtype value type in arg_utils.py - part 2 (#5428)
This commit is contained in:
parent
3ac50b47d0
commit
dad961ef5c
@ -5,6 +5,8 @@ from dataclasses import dataclass
|
|||||||
from typing import (TYPE_CHECKING, Dict, List, Mapping, Optional, Tuple, Type,
|
from typing import (TYPE_CHECKING, Dict, List, Mapping, Optional, Tuple, Type,
|
||||||
Union)
|
Union)
|
||||||
|
|
||||||
|
import torch
|
||||||
|
|
||||||
import vllm.envs as envs
|
import vllm.envs as envs
|
||||||
from vllm.config import (CacheConfig, DecodingConfig, DeviceConfig,
|
from vllm.config import (CacheConfig, DecodingConfig, DeviceConfig,
|
||||||
EngineConfig, LoadConfig, LoRAConfig, ModelConfig,
|
EngineConfig, LoadConfig, LoRAConfig, ModelConfig,
|
||||||
@ -113,7 +115,7 @@ class EngineArgs:
|
|||||||
fully_sharded_loras: bool = False
|
fully_sharded_loras: bool = False
|
||||||
lora_extra_vocab_size: int = 256
|
lora_extra_vocab_size: int = 256
|
||||||
long_lora_scaling_factors: Optional[Tuple[float]] = None
|
long_lora_scaling_factors: Optional[Tuple[float]] = None
|
||||||
lora_dtype: str = 'auto'
|
lora_dtype: Optional[Union[str, torch.dtype]] = 'auto'
|
||||||
max_cpu_loras: Optional[int] = None
|
max_cpu_loras: Optional[int] = None
|
||||||
device: str = 'auto'
|
device: str = 'auto'
|
||||||
num_scheduler_steps: int = 1
|
num_scheduler_steps: int = 1
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user