Skip to content
This repository was archived by the owner on Sep 23, 2025. It is now read-only.

Commit cc356f6

Browse files
committed
update
1 parent 62ab1bf commit cc356f6

File tree

4 files changed

+3
-13
lines changed

4 files changed

+3
-13
lines changed

examples/inference/api_server_openai/query_http_requests_tool.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -73,15 +73,14 @@
7373

7474
messages = [
7575
[
76-
{"role": "user", "content": "You are a helpful assistant"},
7776
{"role": "user", "content": "What's the weather like in Boston today?"},
7877
],
7978
]
8079

8180
proxies = {"http": None, "https": None}
8281

8382
for message in messages:
84-
print(f"User: {message[1]['content']}")
83+
print(f"User: {message[0]['content']}")
8584
print("Assistant:", end=" ", flush=True)
8685

8786
body = {

llm_on_ray/inference/models/gpt-j-6b.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,4 +14,4 @@ ipex:
1414
model_description:
1515
model_id_or_path: EleutherAI/gpt-j-6b
1616
tokenizer_name_or_path: EleutherAI/gpt-j-6b
17-
gpt_base_model: true
17+
gpt_base_model: true

llm_on_ray/inference/predictor_deployment.py

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -310,9 +310,6 @@ def preprocess_prompts(self, input: Union[str, List], tools=None, tool_choice=No
310310
Raises:
311311
HTTPException: If the input prompt format is invalid or not supported.
312312
"""
313-
logger.info("preprocess_prompts")
314-
logger.info(input)
315-
logger.info(type(input))
316313

317314
if isinstance(input, str):
318315
return input
@@ -357,7 +354,6 @@ def preprocess_prompts(self, input: Union[str, List], tools=None, tool_choice=No
357354
raise HTTPException(400, "Invalid prompt format.")
358355

359356
async def __call__(self, http_request: Request) -> Union[StreamingResponse, JSONResponse, str]:
360-
logger.info("PredictorDeployment call")
361357
self.use_openai = False
362358

363359
try:
@@ -376,7 +372,6 @@ async def __call__(self, http_request: Request) -> Union[StreamingResponse, JSON
376372
content="Empty prompt is not supported.",
377373
)
378374
config = json_request["config"] if "config" in json_request else {}
379-
logger.info(input)
380375
# return prompt or list of prompts preprocessed
381376
prompts = self.preprocess_prompts(input)
382377

llm_on_ray/ui/start_ui.py

Lines changed: 1 addition & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -31,11 +31,7 @@
3131
from ray.util import queue
3232
from llm_on_ray.inference.inference_config import all_models, ModelDescription, Prompt
3333
from llm_on_ray.inference.inference_config import InferenceConfig as FinetunedConfig
34-
from llm_on_ray.inference.chat_template_process import (
35-
ChatModelGptJ,
36-
ChatModelLLama,
37-
ChatModelwithImage,
38-
)
34+
3935
from llm_on_ray.inference.predictor_deployment import PredictorDeployment
4036
from llm_on_ray.ui.html_format import cpu_memory_html, ray_status_html, custom_css
4137
from langchain.vectorstores import FAISS

0 commit comments

Comments
 (0)