chore(tests): Mock LLM in tests for PRs
This saves time when testing on CPU which is the only sensible thing to do on GitHub CI for PRs. For releases or once the commit is merged we could use an external runner with GPU or just wait. Signed-off-by: Richard Palethorpe <io@richiejp.com>
This commit is contained in:
@@ -8,7 +8,6 @@ import (
|
||||
"github.com/mudler/LocalAGI/core/types"
|
||||
"github.com/mudler/LocalAGI/pkg/config"
|
||||
"github.com/mudler/LocalAGI/pkg/llm"
|
||||
"github.com/sashabaranov/go-openai"
|
||||
"github.com/sashabaranov/go-openai/jsonschema"
|
||||
)
|
||||
|
||||
@@ -16,7 +15,7 @@ const FilterClassifier = "classifier"
|
||||
|
||||
type ClassifierFilter struct {
|
||||
name string
|
||||
client *openai.Client
|
||||
client llm.LLMClient
|
||||
model string
|
||||
description string
|
||||
allowOnMatch bool
|
||||
|
||||
Reference in New Issue
Block a user