Skip to content

Commit

Permalink
[BugFix] Use args.trust_remote_code (vllm-project#7121)
Browse files Browse the repository at this point in the history
  • Loading branch information
VastoLorde95 authored Aug 5, 2024
1 parent 18fec04 commit c6088f2
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions vllm/entrypoints/openai/api_server.py
Original file line number Diff line number Diff line change
Expand Up @@ -60,11 +60,11 @@
_running_tasks: Set[asyncio.Task] = set()


def model_is_embedding(model_name: str) -> bool:
def model_is_embedding(model_name: str, trust_remote_code: bool) -> bool:
return ModelConfig(model=model_name,
tokenizer=model_name,
tokenizer_mode="auto",
trust_remote_code=False,
trust_remote_code=trust_remote_code,
seed=0,
dtype="float16").embedding_mode

Expand Down Expand Up @@ -97,7 +97,7 @@ async def build_async_engine_client(args) -> AsyncIterator[AsyncEngineClient]:

# If manually triggered or embedding model, use AsyncLLMEngine in process.
# TODO: support embedding model via RPC.
if (model_is_embedding(args.model)
if (model_is_embedding(args.model, args.trust_remote_code)
or args.disable_frontend_multiprocessing):
async_engine_client = AsyncLLMEngine.from_engine_args(
engine_args, usage_context=UsageContext.OPENAI_API_SERVER)
Expand Down

0 comments on commit c6088f2

Please sign in to comment.