diff --git a/fastdeploy/engine/engine.py b/fastdeploy/engine/engine.py index 1525b48d785..9ad677fb9e1 100644 --- a/fastdeploy/engine/engine.py +++ b/fastdeploy/engine/engine.py @@ -503,7 +503,6 @@ def _start_worker_service(self): f" --tensor_parallel_size {self.cfg.parallel_config.tensor_parallel_size}" f" --engine_worker_queue_port {ports}" f" --pod_ip {self.cfg.master_ip}" - f" --total_block_num {self.cfg.cache_config.total_block_num}" f" --block_size {self.cfg.cache_config.block_size}" f" --enc_dec_block_num {self.cfg.cache_config.enc_dec_block_num}" f" --eos_tokens_lens {self.engine.data_processor.eos_token_id_len}" @@ -536,7 +535,7 @@ def _start_worker_service(self): f" --logprobs_mode {self.cfg.model_config.logprobs_mode}" ) - worker_append_flag = { + worker_store_true_flag = { "enable_expert_parallel": self.cfg.parallel_config.enable_expert_parallel, "enable_prefix_caching": self.cfg.cache_config.enable_prefix_caching, "enable_chunked_prefill": self.cfg.cache_config.enable_chunked_prefill, @@ -547,9 +546,17 @@ def _start_worker_service(self): "enable_logprob": self.cfg.model_config.enable_logprob, "lm_head_fp32": self.cfg.model_config.lm_head_fp32, } - for worker_flag, value in worker_append_flag.items(): + for worker_flag, value in worker_store_true_flag.items(): if value: arguments = arguments + f" --{worker_flag}" + + worker_default_none_flag = { + "num_gpu_blocks_override": self.cfg.cache_config.num_gpu_blocks_override, + } + for worker_flag, value in worker_default_none_flag.items(): + if value: + arguments = arguments + f" --{worker_flag} {value}" + if self.cfg.nnode > 1: pd_cmd = pd_cmd + f" --ips {ips} --nnodes {len(self.cfg.ips)}" pd_cmd = pd_cmd + arguments + f" 2>{log_dir}/launch_worker.log" diff --git a/fastdeploy/worker/worker_process.py b/fastdeploy/worker/worker_process.py index 9ea08542bc3..3279622526b 100644 --- a/fastdeploy/worker/worker_process.py +++ b/fastdeploy/worker/worker_process.py @@ -480,7 +480,7 @@ def parse_args(): help="model dir", ) parser.add_argument("-mbs", "--max_num_seqs", type=int, default=34, help="max batch size") - parser.add_argument("--total_block_num", type=int, default=2000) + parser.add_argument("--num_gpu_blocks_override", type=int, default=None) parser.add_argument("--block_size", type=int, default=64) parser.add_argument("--pod_ip", type=str, default="127.0.0.1") parser.add_argument("--engine_worker_queue_port", type=str, default="9923") @@ -707,6 +707,7 @@ def initialize_fd_config(args, ranks: int = 1, local_rank: int = 0) -> FDConfig: parallel_config = ParallelConfig(vars(args)) cache_config = CacheConfig(vars(args)) scheduler_config = SchedulerConfig(vars(args)) + parallel_config.tensor_parallel_rank = local_rank % parallel_config.tensor_parallel_size parallel_config.data_parallel_rank = local_rank // parallel_config.tensor_parallel_size # config for EP