From 7fca3620f63f6c6a11a50a0dd14b3930c44bcfcb Mon Sep 17 00:00:00 2001 From: mudler Date: Sat, 12 Apr 2025 18:46:17 +0200 Subject: [PATCH] Back at arcee-agent as default Signed-off-by: mudler --- Makefile | 2 +- README.md | 6 +++--- docker-compose.yaml | 12 ++++++------ 3 files changed, 10 insertions(+), 10 deletions(-) diff --git a/Makefile b/Makefile index 33ee861..cb71f78 100644 --- a/Makefile +++ b/Makefile @@ -9,7 +9,7 @@ cleanup-tests: docker compose down tests: prepare-tests - LOCALAGI_MODEL="gemma-3-4b-it" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... + LOCALAGI_MODEL="arcee-agent" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... run-nokb: $(MAKE) run KBDISABLEINDEX=true diff --git a/README.md b/README.md index 60084d3..6af349e 100644 --- a/README.md +++ b/README.md @@ -83,7 +83,7 @@ LocalAGI supports multiple hardware configurations through Docker Compose profil - Supports text, multimodal, and image generation models - Run with: `docker compose --profile nvidia up` - Default models: - - Text: `gemma-3-4b-it` + - Text: `arcee-agent` - Multimodal: `minicpm-v-2_6` - Image: `flux.1-dev` - Environment variables: @@ -99,7 +99,7 @@ LocalAGI supports multiple hardware configurations through Docker Compose profil - Supports text, multimodal, and image generation models - Run with: `docker compose --profile intel up` - Default models: - - Text: `gemma-3-4b-it` + - Text: `arcee-agent` - Multimodal: `minicpm-v-2_6` - Image: `sd-1.5-ggml` - Environment variables: @@ -130,7 +130,7 @@ docker compose --profile intel up ``` If no models are specified, it will use the defaults: -- Text model: `gemma-3-4b-it` +- Text model: `arcee-agent` - Multimodal model: `minicpm-v-2_6` - Image model: `flux.1-dev` (NVIDIA) or `sd-1.5-ggml` (Intel) diff --git a/docker-compose.yaml b/docker-compose.yaml index c37670c..2778c4e 100644 --- a/docker-compose.yaml +++ b/docker-compose.yaml @@ -8,7 +8,7 @@ services: image: localai/localai:master-ffmpeg-core command: # - gemma-3-12b-it - - ${MODEL_NAME:-gemma-3-4b-it} + - ${MODEL_NAME:-arcee-agent} - granite-embedding-107m-multilingual healthcheck: test: ["CMD", "curl", "-f", "http://localhost:8080/readyz"] @@ -39,7 +39,7 @@ services: count: 1 capabilities: [gpu] command: - - ${MODEL_NAME:-gemma-3-4b-it} + - ${MODEL_NAME:-arcee-agent} - ${MULTIMODAL_MODEL:-minicpm-v-2_6} - ${IMAGE_MODEL:-flux.1-dev} - granite-embedding-107m-multilingual @@ -57,7 +57,7 @@ services: - /dev/dri/card1 - /dev/dri/renderD129 command: - - ${MODEL_NAME:-gemma-3-4b-it} + - ${MODEL_NAME:-arcee-agent} - ${MULTIMODAL_MODEL:-minicpm-v-2_6} - ${IMAGE_MODEL:-sd-1.5-ggml} - granite-embedding-107m-multilingual @@ -96,7 +96,7 @@ services: - 8080:3000 #image: quay.io/mudler/localagi:master environment: - - LOCALAGI_MODEL=${MODEL_NAME:-gemma-3-4b-it} + - LOCALAGI_MODEL=${MODEL_NAME:-arcee-agent} - LOCALAGI_LLM_API_URL=http://localai:8080 #- LOCALAGI_LLM_API_KEY=sk-1234567890 - LOCALAGI_LOCALRAG_URL=http://localrecall:8080 @@ -113,7 +113,7 @@ services: extends: service: localagi environment: - - LOCALAGI_MODEL=${MODEL_NAME:-gemma-3-4b-it} + - LOCALAGI_MODEL=${MODEL_NAME:-arcee-agent} - LOCALAGI_MULTIMODAL_MODEL=${MULTIMODAL_MODEL:-minicpm-v-2_6} - LOCALAGI_IMAGE_MODEL=${IMAGE_MODEL:-flux.1-dev} - LOCALAGI_LLM_API_URL=http://localai:8080 @@ -127,7 +127,7 @@ services: extends: service: localagi environment: - - LOCALAGI_MODEL=${MODEL_NAME:-gemma-3-4b-it} + - LOCALAGI_MODEL=${MODEL_NAME:-arcee-agent} - LOCALAGI_MULTIMODAL_MODEL=${MULTIMODAL_MODEL:-minicpm-v-2_6} - LOCALAGI_IMAGE_MODEL=${IMAGE_MODEL:-sd-1.5-ggml} - LOCALAGI_LLM_API_URL=http://localai:8080