From 45969d31876e4f155be9aa0ca725277b9822f4a1 Mon Sep 17 00:00:00 2001 From: mudler Date: Tue, 29 Apr 2025 16:25:09 +0200 Subject: [PATCH] chore: use qwen3 as default Signed-off-by: mudler --- Makefile | 2 +- README.md | 6 +++--- docker-compose.yaml | 4 ++-- 3 files changed, 6 insertions(+), 6 deletions(-) diff --git a/Makefile b/Makefile index 9c978c7..45c30bb 100644 --- a/Makefile +++ b/Makefile @@ -11,7 +11,7 @@ cleanup-tests: docker compose down tests: prepare-tests - LOCALAGI_MCPBOX_URL="http://localhost:9090" LOCALAGI_MODEL="gemma-3-12b-it-qat" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... + LOCALAGI_MCPBOX_URL="http://localhost:9090" LOCALAGI_MODEL="qwen3-8b" LOCALAI_API_URL="http://localhost:8081" LOCALAGI_API_URL="http://localhost:8080" $(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --fail-fast -v -r ./... run-nokb: $(MAKE) run KBDISABLEINDEX=true diff --git a/README.md b/README.md index e562dc9..007dcfd 100644 --- a/README.md +++ b/README.md @@ -120,7 +120,7 @@ LocalAGI supports multiple hardware configurations through Docker Compose profil - Supports text, multimodal, and image generation models - Run with: `docker compose -f docker-compose.nvidia.yaml up` - Default models: - - Text: `gemma-3-12b-it-qat` + - Text: `qwen3-8b` - Multimodal: `minicpm-v-2_6` - Image: `sd-1.5-ggml` - Environment variables: @@ -136,7 +136,7 @@ LocalAGI supports multiple hardware configurations through Docker Compose profil - Supports text, multimodal, and image generation models - Run with: `docker compose -f docker-compose.intel.yaml up` - Default models: - - Text: `gemma-3-12b-it-qat` + - Text: `qwen3-8b` - Multimodal: `minicpm-v-2_6` - Image: `sd-1.5-ggml` - Environment variables: @@ -167,7 +167,7 @@ docker compose -f docker-compose.intel.yaml up ``` If no models are specified, it will use the defaults: -- Text model: `gemma-3-12b-it-qat` +- Text model: `qwen3-8b` - Multimodal model: `minicpm-v-2_6` - Image model: `sd-1.5-ggml` diff --git a/docker-compose.yaml b/docker-compose.yaml index 83ab13b..13e213a 100644 --- a/docker-compose.yaml +++ b/docker-compose.yaml @@ -7,7 +7,7 @@ services: # Image list (dockerhub): https://hub.docker.com/r/localai/localai image: localai/localai:master-ffmpeg-core command: - - ${MODEL_NAME:-gemma-3-12b-it-qat} + - ${MODEL_NAME:-qwen3-8b} - ${MULTIMODAL_MODEL:-minicpm-v-2_6} - ${IMAGE_MODEL:-sd-1.5-ggml} - granite-embedding-107m-multilingual @@ -91,7 +91,7 @@ services: - 8080:3000 #image: quay.io/mudler/localagi:master environment: - - LOCALAGI_MODEL=${MODEL_NAME:-gemma-3-12b-it-qat} + - LOCALAGI_MODEL=${MODEL_NAME:-qwen3-8b} - LOCALAGI_MULTIMODAL_MODEL=${MULTIMODAL_MODEL:-minicpm-v-2_6} - LOCALAGI_IMAGE_MODEL=${IMAGE_MODEL:-sd-1.5-ggml} - LOCALAGI_LLM_API_URL=http://localai:8080