diff --git a/llama_cli/llama_cli/api/__init__.py b/llama_cli/llama_cli/api/__init__.py index 5af4719..4194bbe 100644 --- a/llama_cli/llama_cli/api/__init__.py +++ b/llama_cli/llama_cli/api/__init__.py @@ -60,6 +60,7 @@ def text_cb(feedback) -> None: goal = GenerateResponse.Goal() goal.prompt = prompt goal.sampling_config.temp = temp - llama_client.generate_response(goal, text_cb) - print("") + response = llama_client.generate_response(goal, text_cb)[0].response.text + if not response.endswith("\n"): + print("") rclpy.shutdown() diff --git a/llama_ros/llama_cpp b/llama_ros/llama_cpp index 20fc380..d7fd29f 160000 --- a/llama_ros/llama_cpp +++ b/llama_ros/llama_cpp @@ -1 +1 @@ -Subproject commit 20fc3804bfb727074bc270b6eacb60af8d0bf7d4 +Subproject commit d7fd29fff16456ce9c3a23fd2d09a66256b05aff