remove unused cmd-line files
This commit is contained in:
parent
456ac51634
commit
34bd6eabc9
|
@ -3,9 +3,8 @@ from ..base_device import Device
|
||||||
device = Device()
|
device = Device()
|
||||||
|
|
||||||
|
|
||||||
def main(server_url, tts_service):
|
def main(server_url):
|
||||||
device.server_url = server_url
|
device.server_url = server_url
|
||||||
device.tts_service = tts_service
|
|
||||||
device.start()
|
device.start()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -3,9 +3,8 @@ from ..base_device import Device
|
||||||
device = Device()
|
device = Device()
|
||||||
|
|
||||||
|
|
||||||
def main(server_url, tts_service):
|
def main(server_url):
|
||||||
device.server_url = server_url
|
device.server_url = server_url
|
||||||
device.tts_service = tts_service
|
|
||||||
device.start()
|
device.start()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -3,9 +3,8 @@ from ..base_device import Device
|
||||||
device = Device()
|
device = Device()
|
||||||
|
|
||||||
|
|
||||||
def main(server_url, tts_service):
|
def main(server_url):
|
||||||
device.server_url = server_url
|
device.server_url = server_url
|
||||||
device.tts_service = tts_service
|
|
||||||
device.start()
|
device.start()
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -7,9 +7,6 @@ import importlib
|
||||||
from source.server.tunnel import create_tunnel
|
from source.server.tunnel import create_tunnel
|
||||||
from source.server.async_server import main
|
from source.server.async_server import main
|
||||||
|
|
||||||
# from source.server.server import main
|
|
||||||
from source.server.utils.local_mode import select_local_model
|
|
||||||
|
|
||||||
import signal
|
import signal
|
||||||
|
|
||||||
app = typer.Typer()
|
app = typer.Typer()
|
||||||
|
@ -41,64 +38,20 @@ def run(
|
||||||
client_type: str = typer.Option(
|
client_type: str = typer.Option(
|
||||||
"auto", "--client-type", help="Specify the client type"
|
"auto", "--client-type", help="Specify the client type"
|
||||||
),
|
),
|
||||||
llm_service: str = typer.Option(
|
qr: bool = typer.Option(
|
||||||
"litellm", "--llm-service", help="Specify the LLM service"
|
False, "--qr", help="Display QR code to scan to connect to the server"
|
||||||
),
|
|
||||||
model: str = typer.Option("gpt-4", "--model", help="Specify the model"),
|
|
||||||
llm_supports_vision: bool = typer.Option(
|
|
||||||
False,
|
|
||||||
"--llm-supports-vision",
|
|
||||||
help="Specify if the LLM service supports vision",
|
|
||||||
),
|
|
||||||
llm_supports_functions: bool = typer.Option(
|
|
||||||
False,
|
|
||||||
"--llm-supports-functions",
|
|
||||||
help="Specify if the LLM service supports functions",
|
|
||||||
),
|
|
||||||
context_window: int = typer.Option(
|
|
||||||
2048, "--context-window", help="Specify the context window size"
|
|
||||||
),
|
|
||||||
max_tokens: int = typer.Option(
|
|
||||||
4096, "--max-tokens", help="Specify the maximum number of tokens"
|
|
||||||
),
|
|
||||||
temperature: float = typer.Option(
|
|
||||||
0.8, "--temperature", help="Specify the temperature for generation"
|
|
||||||
),
|
|
||||||
tts_service: str = typer.Option(
|
|
||||||
"elevenlabs", "--tts-service", help="Specify the TTS service"
|
|
||||||
),
|
|
||||||
stt_service: str = typer.Option(
|
|
||||||
"openai", "--stt-service", help="Specify the STT service"
|
|
||||||
),
|
|
||||||
local: bool = typer.Option(
|
|
||||||
False, "--local", help="Use recommended local services for LLM, STT, and TTS"
|
|
||||||
),
|
|
||||||
qr: bool = typer.Option(False, "--qr", help="Print the QR code for the server URL"),
|
|
||||||
mobile: bool = typer.Option(
|
|
||||||
False, "--mobile", help="Toggle server to support mobile app"
|
|
||||||
),
|
),
|
||||||
):
|
):
|
||||||
_run(
|
_run(
|
||||||
server=server or mobile,
|
server=server,
|
||||||
server_host=server_host,
|
server_host=server_host,
|
||||||
server_port=server_port,
|
server_port=server_port,
|
||||||
tunnel_service=tunnel_service,
|
tunnel_service=tunnel_service,
|
||||||
expose=expose or mobile,
|
expose=expose,
|
||||||
client=client,
|
client=client,
|
||||||
server_url=server_url,
|
server_url=server_url,
|
||||||
client_type=client_type,
|
client_type=client_type,
|
||||||
llm_service=llm_service,
|
qr=qr,
|
||||||
model=model,
|
|
||||||
llm_supports_vision=llm_supports_vision,
|
|
||||||
llm_supports_functions=llm_supports_functions,
|
|
||||||
context_window=context_window,
|
|
||||||
max_tokens=max_tokens,
|
|
||||||
temperature=temperature,
|
|
||||||
tts_service=tts_service,
|
|
||||||
stt_service=stt_service,
|
|
||||||
local=local,
|
|
||||||
qr=qr or mobile,
|
|
||||||
mobile=mobile,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@ -111,24 +64,8 @@ def _run(
|
||||||
client: bool = False,
|
client: bool = False,
|
||||||
server_url: str = None,
|
server_url: str = None,
|
||||||
client_type: str = "auto",
|
client_type: str = "auto",
|
||||||
llm_service: str = "litellm",
|
|
||||||
model: str = "gpt-4",
|
|
||||||
llm_supports_vision: bool = False,
|
|
||||||
llm_supports_functions: bool = False,
|
|
||||||
context_window: int = 2048,
|
|
||||||
max_tokens: int = 4096,
|
|
||||||
temperature: float = 0.8,
|
|
||||||
tts_service: str = "elevenlabs",
|
|
||||||
stt_service: str = "openai",
|
|
||||||
local: bool = False,
|
|
||||||
qr: bool = False,
|
qr: bool = False,
|
||||||
mobile: bool = False,
|
|
||||||
):
|
):
|
||||||
if local:
|
|
||||||
tts_service = "coqui"
|
|
||||||
# llm_service = "llamafile"
|
|
||||||
stt_service = "local-whisper"
|
|
||||||
select_local_model()
|
|
||||||
|
|
||||||
system_type = platform.system()
|
system_type = platform.system()
|
||||||
if system_type == "Windows":
|
if system_type == "Windows":
|
||||||
|
@ -156,21 +93,10 @@ def _run(
|
||||||
main(
|
main(
|
||||||
server_host,
|
server_host,
|
||||||
server_port,
|
server_port,
|
||||||
tts_service,
|
|
||||||
# llm_service,
|
|
||||||
# model,
|
|
||||||
# llm_supports_vision,
|
|
||||||
# llm_supports_functions,
|
|
||||||
# context_window,
|
|
||||||
# max_tokens,
|
|
||||||
# temperature,
|
|
||||||
# stt_service,
|
|
||||||
# mobile,
|
|
||||||
),
|
),
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
server_thread.start()
|
server_thread.start()
|
||||||
print("server thread started")
|
|
||||||
|
|
||||||
if expose:
|
if expose:
|
||||||
tunnel_thread = threading.Thread(
|
tunnel_thread = threading.Thread(
|
||||||
|
@ -199,11 +125,8 @@ def _run(
|
||||||
f".clients.{client_type}.device", package="source"
|
f".clients.{client_type}.device", package="source"
|
||||||
)
|
)
|
||||||
|
|
||||||
client_thread = threading.Thread(
|
client_thread = threading.Thread(target=module.main, args=[server_url])
|
||||||
target=module.main, args=[server_url, tts_service]
|
|
||||||
)
|
|
||||||
client_thread.start()
|
client_thread.start()
|
||||||
print("client thread started")
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
if server:
|
if server:
|
||||||
|
|
Loading…
Reference in New Issue