From 525efb264ebca6a48d7efad17adcccd2e0232289 Mon Sep 17 00:00:00 2001 From: mudler Date: Sat, 12 Apr 2025 17:37:19 +0200 Subject: [PATCH] use openthinker, it's smaller --- Makefile | 2 +- core/agent/agent.go | 4 +++- docker-compose.gpu.intel.yaml | 4 ++-- docker-compose.yaml | 5 +++-- 4 files changed, 9 insertions(+), 6 deletions(-) diff --git a/Makefile b/Makefile index 97c8ec8..0049827 100644 --- a/Makefile +++ b/Makefile @@ -9,7 +9,7 @@ cleanup-tests: docker compose down tests: prepare-tests - LOCALAGI_MODEL="gemma-3-12b-it" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... + LOCALAGI_MODEL="openthinker-7b" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... run-nokb: $(MAKE) run KBDISABLEINDEX=true diff --git a/core/agent/agent.go b/core/agent/agent.go index 39b9fa3..e34ca44 100644 --- a/core/agent/agent.go +++ b/core/agent/agent.go @@ -249,7 +249,7 @@ func (a *Agent) runAction(ctx context.Context, chosenAction types.Action, params } } - xlog.Info("Running action", "action", chosenAction.Definition().Name, "agent", a.Character.Name) + xlog.Info("[runAction] Running action", "action", chosenAction.Definition().Name, "agent", a.Character.Name, "params", params.String()) if chosenAction.Definition().Name.Is(action.StateActionName) { // We need to store the result in the state @@ -270,6 +270,8 @@ func (a *Agent) runAction(ctx context.Context, chosenAction types.Action, params } } + xlog.Debug("[runAction] Action result", "action", chosenAction.Definition().Name, "params", params.String(), "result", result.Result) + return result, nil } diff --git a/docker-compose.gpu.intel.yaml b/docker-compose.gpu.intel.yaml index 307bde4..e447f3a 100644 --- a/docker-compose.gpu.intel.yaml +++ b/docker-compose.gpu.intel.yaml @@ -8,7 +8,7 @@ services: image: localai/localai:master-sycl-f32-ffmpeg-core command: # - rombo-org_rombo-llm-v3.0-qwen-32b # minimum suggested model - - gemma-3-12b-it # (smaller) + - openthinker-7b # (smaller) - granite-embedding-107m-multilingual healthcheck: test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] @@ -62,7 +62,7 @@ services: - 8080:3000 image: quay.io/mudler/localagi:master environment: - - LOCALAGI_MODEL=gemma-3-12b-it + - LOCALAGI_MODEL=openthinker-7b - LOCALAGI_LLM_API_URL=http://localai:8080 #- LOCALAGI_LLM_API_KEY=sk-1234567890 - LOCALAGI_LOCALRAG_URL=http://localrecall:8080 diff --git a/docker-compose.yaml b/docker-compose.yaml index a6598c4..60963ba 100644 --- a/docker-compose.yaml +++ b/docker-compose.yaml @@ -7,7 +7,8 @@ services: # Image list (dockerhub): https://hub.docker.com/r/localai/localai image: localai/localai:master-ffmpeg-core command: - - gemma-3-12b-it + # - gemma-3-12b-it + - ${MODEL_NAME:-openthinker-7b} - granite-embedding-107m-multilingual healthcheck: test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] @@ -65,7 +66,7 @@ services: - 8080:3000 #image: quay.io/mudler/localagi:master environment: - - LOCALAGI_MODEL=gemma-3-12b-it + - LOCALAGI_MODEL=${MODEL_NAME:-openthinker-7b} - LOCALAGI_LLM_API_URL=http://localai:8080 #- LOCALAGI_LLM_API_KEY=sk-1234567890 - LOCALAGI_LOCALRAG_URL=http://localrecall:8080