Compare commits
68 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| 7cc283a850 | |||
|
|
2368a39d11 | ||
|
|
0be9ad030a | ||
|
|
febc9bd5b5 | ||
|
|
2d5ae034c9 | ||
|
|
1bc11de465 | ||
|
|
575e16f2fa | ||
|
|
615b05c8d6 | ||
|
|
d1cca04e76 | ||
|
|
90fd0e46f7 | ||
|
|
14a309d7d6 | ||
|
|
8dbb2286dc | ||
|
|
b6bd53b01a | ||
|
|
986b1949cd | ||
|
|
1e81e4db53 | ||
|
|
23aecd372e | ||
|
|
db53f27a1a | ||
|
|
c83e9a859b | ||
|
|
02fd70726b | ||
|
|
9d50395dc5 | ||
|
|
9d125a87d9 | ||
|
|
61e930bf8a | ||
|
|
4db60b6a6f | ||
|
|
69e9c7de55 | ||
|
|
e96fa163cd | ||
|
|
cfef80e1e5 | ||
|
|
9b74a4354b | ||
|
|
fca193b5b2 | ||
|
|
cc9eede856 | ||
|
|
f0ff3d5e5a | ||
|
|
81d6dea7da | ||
|
|
1328bd1306 | ||
|
|
6fa88be433 | ||
|
|
2892f24030 | ||
|
|
1e3442db14 | ||
|
|
f74154d96f | ||
|
|
36d83e0a0e | ||
|
|
33defac76c | ||
|
|
4306a6866f | ||
|
|
039f6890a7 | ||
|
|
4fff318ea9 | ||
|
|
ea6efd553d | ||
|
|
d45ef5c622 | ||
|
|
9358f83229 | ||
|
|
e49d31d725 | ||
|
|
13a27e1d00 | ||
|
|
3e7f3920b2 | ||
|
|
8f8e3bd85e | ||
|
|
7e7f83e985 | ||
|
|
c42f981f55 | ||
|
|
00cd0a5b5a | ||
|
|
4e9ebbbc2c | ||
|
|
eefbf790c3 | ||
|
|
942c175b90 | ||
|
|
10e895bb94 | ||
|
|
a1cc54f01f | ||
|
|
e3256682ba | ||
|
|
7635cce15a | ||
|
|
53a041921b | ||
|
|
af3399515a | ||
|
|
01991c0060 | ||
|
|
3f8d67b145 | ||
|
|
ab8b597843 | ||
|
|
ddf9070a64 | ||
|
|
b9727981cc | ||
|
|
e1db799b1d | ||
|
|
f5c01ad83a | ||
|
|
190915214d |
@@ -1 +0,0 @@
|
||||
NODE_ENV=development\nOPENAI_API_KEY=your_openai_api_key_here\nHASS_HOST=http://homeassistant.local:8123\nHASS_TOKEN=your_hass_token_here\nPORT=3000\nHASS_SOCKET_URL=ws://homeassistant.local:8123/api/websocket\nLOG_LEVEL=debug\nMCP_SERVER=http://localhost:3000\nOPENAI_MODEL=deepseek-v3\nMAX_RETRIES=3\nANALYSIS_TIMEOUT=30000\n\n# Home Assistant specific settings\nAUTOMATION_PATH=./config/automations.yaml\nBLUEPRINT_REPO=https://blueprints.home-assistant.io/\nENERGY_DASHBOARD=true\n\n# Available models: gpt-4o, gpt-4-turbo, gpt-4, gpt-4-o1, gpt-4-o3, gpt-3.5-turbo, gpt-3.5-turbo-16k, deepseek-v3, deepseek-r1\n\n# For DeepSeek models\nDEEPSEEK_API_KEY=your_deepseek_api_key_here\nDEEPSEEK_BASE_URL=https://api.deepseek.com/v1\n\n# Model specifications:\n# - gpt-4-o1: 128k context, general purpose\n# - gpt-4-o3: 1M context, large-scale analysis\n\n# Add processor type specification\nPROCESSOR_TYPE=claude # Change to openai when using OpenAI
|
||||
135
.env.example
135
.env.example
@@ -1,43 +1,17 @@
|
||||
# Server Configuration
|
||||
NODE_ENV=development
|
||||
PORT=7123
|
||||
DEBUG=false
|
||||
LOG_LEVEL=info
|
||||
MCP_SERVER=http://localhost:7123
|
||||
USE_STDIO_TRANSPORT=true
|
||||
|
||||
# Home Assistant Configuration
|
||||
# The URL of your Home Assistant instance
|
||||
HASS_HOST=http://homeassistant.local:8123
|
||||
|
||||
# Long-lived access token from Home Assistant
|
||||
# Generate from Profile -> Long-Lived Access Tokens
|
||||
HASS_TOKEN=your_home_assistant_token
|
||||
|
||||
# WebSocket URL for real-time updates
|
||||
HASS_TOKEN=your_long_lived_token
|
||||
HASS_SOCKET_URL=ws://homeassistant.local:8123/api/websocket
|
||||
|
||||
# Server Configuration
|
||||
# Port for the MCP server (default: 3000)
|
||||
PORT=3000
|
||||
|
||||
# Environment (development/production/test)
|
||||
NODE_ENV=development
|
||||
|
||||
# Debug mode (true/false)
|
||||
DEBUG=false
|
||||
|
||||
# Logging level (debug/info/warn/error)
|
||||
LOG_LEVEL=info
|
||||
|
||||
# AI Configuration
|
||||
# Natural Language Processor type (claude/gpt4/custom)
|
||||
PROCESSOR_TYPE=claude
|
||||
|
||||
# OpenAI API Key (required for GPT-4 analysis)
|
||||
OPENAI_API_KEY=your_openai_api_key
|
||||
|
||||
# Rate Limiting
|
||||
# Requests per minute per IP for regular endpoints
|
||||
RATE_LIMIT_REGULAR=100
|
||||
|
||||
# Requests per minute per IP for WebSocket connections
|
||||
RATE_LIMIT_WEBSOCKET=1000
|
||||
|
||||
# Security Configuration
|
||||
# JWT Configuration
|
||||
JWT_SECRET=your_jwt_secret_key_min_32_chars
|
||||
JWT_EXPIRY=86400000
|
||||
JWT_MAX_AGE=2592000000
|
||||
@@ -46,31 +20,18 @@ JWT_ALGORITHM=HS256
|
||||
# Rate Limiting
|
||||
RATE_LIMIT_WINDOW=900000
|
||||
RATE_LIMIT_MAX_REQUESTS=100
|
||||
|
||||
# Token Security
|
||||
TOKEN_MIN_LENGTH=32
|
||||
MAX_FAILED_ATTEMPTS=5
|
||||
LOCKOUT_DURATION=900000
|
||||
RATE_LIMIT_MAX_AUTH_REQUESTS=5
|
||||
RATE_LIMIT_REGULAR=100
|
||||
RATE_LIMIT_WEBSOCKET=1000
|
||||
|
||||
# CORS Configuration
|
||||
CORS_ORIGINS=http://localhost:3000,http://localhost:8123
|
||||
CORS_ORIGINS=http://localhost:3000,http://localhost:8123,http://homeassistant.local:8123
|
||||
CORS_METHODS=GET,POST,PUT,DELETE,OPTIONS
|
||||
CORS_ALLOWED_HEADERS=Content-Type,Authorization,X-Requested-With
|
||||
CORS_EXPOSED_HEADERS=
|
||||
CORS_CREDENTIALS=true
|
||||
CORS_MAX_AGE=86400
|
||||
|
||||
# Content Security Policy
|
||||
CSP_ENABLED=true
|
||||
CSP_REPORT_ONLY=false
|
||||
CSP_REPORT_URI=
|
||||
|
||||
# SSL/TLS Configuration
|
||||
REQUIRE_HTTPS=true
|
||||
HSTS_MAX_AGE=31536000
|
||||
HSTS_INCLUDE_SUBDOMAINS=true
|
||||
HSTS_PRELOAD=true
|
||||
|
||||
# Cookie Security
|
||||
COOKIE_SECRET=your_cookie_secret_key_min_32_chars
|
||||
COOKIE_SECURE=true
|
||||
@@ -81,31 +42,57 @@ COOKIE_SAME_SITE=Strict
|
||||
MAX_REQUEST_SIZE=1048576
|
||||
MAX_REQUEST_FIELDS=1000
|
||||
|
||||
# SSE Configuration
|
||||
SSE_MAX_CLIENTS=1000
|
||||
SSE_PING_INTERVAL=30000
|
||||
# AI Configuration
|
||||
PROCESSOR_TYPE=openai
|
||||
OPENAI_API_KEY=your_openai_api_key
|
||||
OPENAI_MODEL=gpt-3.5-turbo
|
||||
MAX_RETRIES=3
|
||||
ANALYSIS_TIMEOUT=30000
|
||||
|
||||
# Logging Configuration
|
||||
LOG_LEVEL=info
|
||||
LOG_DIR=logs
|
||||
LOG_MAX_SIZE=20m
|
||||
LOG_MAX_DAYS=14d
|
||||
LOG_COMPRESS=true
|
||||
LOG_REQUESTS=true
|
||||
# Speech Features Configuration
|
||||
ENABLE_SPEECH_FEATURES=false
|
||||
ENABLE_WAKE_WORD=false
|
||||
ENABLE_SPEECH_TO_TEXT=false
|
||||
WHISPER_MODEL_PATH=/models
|
||||
WHISPER_MODEL_TYPE=base
|
||||
|
||||
# Audio Configuration
|
||||
NOISE_THRESHOLD=0.05
|
||||
MIN_SPEECH_DURATION=1.0
|
||||
SILENCE_DURATION=0.5
|
||||
SAMPLE_RATE=16000
|
||||
CHANNELS=1
|
||||
CHUNK_SIZE=1024
|
||||
PULSE_SERVER=unix:/run/user/1000/pulse/native
|
||||
|
||||
# Whisper Configuration
|
||||
ASR_MODEL=base
|
||||
ASR_ENGINE=faster_whisper
|
||||
WHISPER_BEAM_SIZE=5
|
||||
COMPUTE_TYPE=float32
|
||||
LANGUAGE=en
|
||||
|
||||
# SSE Configuration
|
||||
SSE_MAX_CLIENTS=50
|
||||
SSE_RECONNECT_TIMEOUT=5000
|
||||
|
||||
# Development Flags
|
||||
HOT_RELOAD=true
|
||||
|
||||
# Test Configuration (only needed for running tests)
|
||||
TEST_HASS_HOST=http://homeassistant.local:8123
|
||||
TEST_HASS_TOKEN=test_token
|
||||
TEST_HASS_SOCKET_URL=ws://homeassistant.local:8123/api/websocket
|
||||
TEST_PORT=3001
|
||||
|
||||
# Version
|
||||
VERSION=0.1.0
|
||||
|
||||
# Test Configuration
|
||||
# Only needed if running tests
|
||||
TEST_HASS_HOST=http://localhost:8123
|
||||
TEST_HASS_TOKEN=test_token
|
||||
TEST_HASS_SOCKET_URL=ws://localhost:8123/api/websocket
|
||||
TEST_PORT=3001
|
||||
# Docker Configuration
|
||||
COMPOSE_PROJECT_NAME=mcp
|
||||
|
||||
# Speech Features Configuration
|
||||
ENABLE_SPEECH_FEATURES=false
|
||||
ENABLE_WAKE_WORD=true
|
||||
ENABLE_SPEECH_TO_TEXT=true
|
||||
WHISPER_MODEL_PATH=/models
|
||||
WHISPER_MODEL_TYPE=base
|
||||
# Resource Limits
|
||||
FAST_WHISPER_CPU_LIMIT=4.0
|
||||
FAST_WHISPER_MEMORY_LIMIT=2G
|
||||
MCP_CPU_LIMIT=1.0
|
||||
MCP_MEMORY_LIMIT=512M
|
||||
52
.github/workflows/deploy-docs.yml
vendored
52
.github/workflows/deploy-docs.yml
vendored
@@ -1,4 +1,4 @@
|
||||
name: Deploy Documentation to GitHub Pages
|
||||
name: Deploy Documentation
|
||||
|
||||
on:
|
||||
push:
|
||||
@@ -6,57 +6,69 @@ on:
|
||||
- main
|
||||
paths:
|
||||
- 'docs/**'
|
||||
- '.github/workflows/deploy-docs.yml'
|
||||
- 'mkdocs.yml'
|
||||
# Allow manual trigger
|
||||
workflow_dispatch:
|
||||
|
||||
# Sets permissions of the GITHUB_TOKEN to allow deployment to GitHub Pages
|
||||
permissions:
|
||||
contents: read
|
||||
pages: write
|
||||
id-token: write
|
||||
|
||||
# Allow only one concurrent deployment
|
||||
# Allow only one concurrent deployment, skipping runs queued between the run in-progress and latest queued.
|
||||
concurrency:
|
||||
group: "pages"
|
||||
cancel-in-progress: true
|
||||
cancel-in-progress: false
|
||||
|
||||
jobs:
|
||||
build:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkout
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Setup Ruby
|
||||
uses: ruby/setup-ruby@v1
|
||||
with:
|
||||
ruby-version: '3.2'
|
||||
bundler-cache: true
|
||||
cache-version: 0
|
||||
fetch-depth: 0
|
||||
|
||||
- name: Setup Python
|
||||
uses: actions/setup-python@v5
|
||||
with:
|
||||
python-version: '3.x'
|
||||
cache: 'pip'
|
||||
|
||||
- name: Setup Pages
|
||||
uses: actions/configure-pages@v4
|
||||
|
||||
- name: Install dependencies
|
||||
run: |
|
||||
cd docs
|
||||
bundle install
|
||||
python -m pip install --upgrade pip
|
||||
pip install -r docs/requirements.txt
|
||||
|
||||
- name: Build site
|
||||
- name: List mkdocs configuration
|
||||
run: |
|
||||
cd docs
|
||||
bundle exec jekyll build
|
||||
env:
|
||||
JEKYLL_ENV: production
|
||||
echo "Current directory contents:"
|
||||
ls -la
|
||||
echo "MkDocs version:"
|
||||
mkdocs --version
|
||||
echo "MkDocs configuration:"
|
||||
cat mkdocs.yml
|
||||
|
||||
- name: Build documentation
|
||||
run: |
|
||||
mkdocs build --strict
|
||||
echo "Build output contents:"
|
||||
ls -la site/advanced-homeassistant-mcp
|
||||
|
||||
- name: Upload artifact
|
||||
uses: actions/upload-pages-artifact@v3
|
||||
with:
|
||||
path: docs/_site
|
||||
path: ./site/advanced-homeassistant-mcp
|
||||
|
||||
deploy:
|
||||
needs: build
|
||||
environment:
|
||||
name: github-pages
|
||||
url: ${{ steps.deployment.outputs.page_url }}
|
||||
needs: build
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Deploy to GitHub Pages
|
||||
|
||||
32
.github/workflows/docs-deploy.yml
vendored
32
.github/workflows/docs-deploy.yml
vendored
@@ -1,32 +0,0 @@
|
||||
name: Deploy Documentation
|
||||
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- main
|
||||
paths:
|
||||
- 'docs/**'
|
||||
- 'mkdocs.yml'
|
||||
|
||||
permissions:
|
||||
contents: write
|
||||
|
||||
jobs:
|
||||
deploy-docs:
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Checkout repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Python
|
||||
uses: actions/setup-python@v4
|
||||
with:
|
||||
python-version: 3.x
|
||||
|
||||
- name: Install dependencies
|
||||
run: |
|
||||
pip install mkdocs-material
|
||||
pip install mkdocs
|
||||
|
||||
- name: Deploy documentation
|
||||
run: mkdocs gh-deploy --force
|
||||
15
.gitignore
vendored
15
.gitignore
vendored
@@ -31,7 +31,7 @@ wheels/
|
||||
venv/
|
||||
ENV/
|
||||
env/
|
||||
|
||||
.venv/
|
||||
# Logs
|
||||
logs
|
||||
*.log
|
||||
@@ -71,7 +71,7 @@ coverage/
|
||||
# Environment files
|
||||
.env
|
||||
.env.*
|
||||
!.env.*.template
|
||||
!.env.example
|
||||
|
||||
.cursor/
|
||||
.cursor/*
|
||||
@@ -88,3 +88,14 @@ site/
|
||||
__pycache__/
|
||||
*.py[cod]
|
||||
*$py.class
|
||||
|
||||
models/
|
||||
|
||||
*.code-workspace
|
||||
*.ttf
|
||||
*.otf
|
||||
*.woff
|
||||
*.woff2
|
||||
*.eot
|
||||
*.svg
|
||||
*.png
|
||||
89
Dockerfile
89
Dockerfile
@@ -4,26 +4,31 @@ FROM node:20-slim as builder
|
||||
# Set working directory
|
||||
WORKDIR /app
|
||||
|
||||
# Install bun
|
||||
RUN npm install -g bun@1.0.25
|
||||
# Install bun with the latest version
|
||||
RUN npm install -g bun@1.0.35
|
||||
|
||||
# Install only the minimal dependencies needed and clean up in the same layer
|
||||
# Install Python and other dependencies
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
ca-certificates \
|
||||
curl \
|
||||
&& rm -rf /var/lib/apt/lists/* \
|
||||
&& apt-get clean \
|
||||
&& rm -rf /var/cache/apt/*
|
||||
python3 \
|
||||
python3-pip \
|
||||
python3-venv \
|
||||
build-essential \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Set build-time environment variables
|
||||
ENV NODE_ENV=production \
|
||||
NODE_OPTIONS="--max-old-space-size=2048"
|
||||
# Create and activate virtual environment
|
||||
RUN python3 -m venv /opt/venv
|
||||
ENV PATH="/opt/venv/bin:$PATH"
|
||||
ENV VIRTUAL_ENV="/opt/venv"
|
||||
|
||||
# Copy only package files first
|
||||
# Upgrade pip in virtual environment
|
||||
RUN /opt/venv/bin/python -m pip install --upgrade pip
|
||||
|
||||
# Install Python packages in virtual environment
|
||||
RUN /opt/venv/bin/python -m pip install --no-cache-dir numpy scipy
|
||||
|
||||
# Copy package.json and install dependencies
|
||||
COPY package.json ./
|
||||
|
||||
# Install dependencies without lockfile
|
||||
RUN bun install --no-cache
|
||||
RUN bun install --frozen-lockfile || bun install
|
||||
|
||||
# Copy source files and build
|
||||
COPY src ./src
|
||||
@@ -33,23 +38,55 @@ RUN bun build ./src/index.ts --target=bun --minify --outdir=./dist
|
||||
# Create a smaller production image
|
||||
FROM node:20-slim as runner
|
||||
|
||||
# Install bun in production image
|
||||
RUN npm install -g bun@1.0.25
|
||||
# Install bun in production image with the latest version
|
||||
RUN npm install -g bun@1.0.35
|
||||
|
||||
# Set production environment variables
|
||||
ENV NODE_ENV=production \
|
||||
NODE_OPTIONS="--max-old-space-size=1024"
|
||||
# Install system dependencies
|
||||
RUN apt-get update && apt-get install -y --no-install-recommends \
|
||||
curl \
|
||||
python3 \
|
||||
python3-pip \
|
||||
python3-venv \
|
||||
alsa-utils \
|
||||
pulseaudio \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Create a non-root user
|
||||
# Configure ALSA
|
||||
COPY docker/speech/asound.conf /etc/asound.conf
|
||||
|
||||
# Create and activate virtual environment
|
||||
RUN python3 -m venv /opt/venv
|
||||
ENV PATH="/opt/venv/bin:$PATH"
|
||||
ENV VIRTUAL_ENV="/opt/venv"
|
||||
|
||||
# Upgrade pip in virtual environment
|
||||
RUN /opt/venv/bin/python -m pip install --upgrade pip
|
||||
|
||||
# Install Python packages in virtual environment
|
||||
RUN /opt/venv/bin/python -m pip install --no-cache-dir numpy scipy
|
||||
|
||||
# Create a non-root user and add to audio group
|
||||
RUN addgroup --system --gid 1001 nodejs && \
|
||||
adduser --system --uid 1001 bunjs
|
||||
adduser --system --uid 1001 --gid 1001 bunjs && \
|
||||
adduser bunjs audio
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
# Copy Python virtual environment from builder
|
||||
COPY --from=builder --chown=bunjs:nodejs /opt/venv /opt/venv
|
||||
|
||||
# Copy source files
|
||||
COPY --chown=bunjs:nodejs . .
|
||||
|
||||
# Copy only the necessary files from builder
|
||||
COPY --from=builder --chown=bunjs:nodejs /app/dist ./dist
|
||||
COPY --from=builder --chown=bunjs:nodejs /app/node_modules ./node_modules
|
||||
COPY --chown=bunjs:nodejs package.json ./
|
||||
|
||||
# Ensure audio setup script is executable
|
||||
RUN chmod +x /app/docker/speech/setup-audio.sh
|
||||
|
||||
# Create logs and audio directories with proper permissions
|
||||
RUN mkdir -p /app/logs /app/audio && chown -R bunjs:nodejs /app/logs /app/audio
|
||||
|
||||
# Switch to non-root user
|
||||
USER bunjs
|
||||
@@ -59,7 +96,7 @@ HEALTHCHECK --interval=30s --timeout=10s --start-period=5s --retries=3 \
|
||||
CMD curl -f http://localhost:4000/health || exit 1
|
||||
|
||||
# Expose port
|
||||
EXPOSE 4000
|
||||
EXPOSE ${PORT:-4000}
|
||||
|
||||
# Start the application with optimized flags
|
||||
CMD ["bun", "--smol", "run", "start"]
|
||||
# Start the application with audio setup
|
||||
CMD ["/bin/bash", "-c", "/app/docker/speech/setup-audio.sh || echo 'Audio setup failed, continuing anyway' && bun --smol run fix-env.js"]
|
||||
96
PUBLISHING.md
Normal file
96
PUBLISHING.md
Normal file
@@ -0,0 +1,96 @@
|
||||
# Publishing to npm
|
||||
|
||||
This document outlines the steps to publish the Home Assistant MCP server to npm.
|
||||
|
||||
## Prerequisites
|
||||
|
||||
1. You need an npm account. Create one at [npmjs.com](https://www.npmjs.com/signup) if you don't have one.
|
||||
2. You need to be logged in to npm on your local machine:
|
||||
```bash
|
||||
npm login
|
||||
```
|
||||
3. You need to have all the necessary dependencies installed:
|
||||
```bash
|
||||
npm install
|
||||
```
|
||||
|
||||
## Before Publishing
|
||||
|
||||
1. Make sure all tests pass:
|
||||
```bash
|
||||
npm test
|
||||
```
|
||||
|
||||
2. Build all the necessary files:
|
||||
```bash
|
||||
npm run build # Build for Bun
|
||||
npm run build:node # Build for Node.js
|
||||
npm run build:stdio # Build the stdio server
|
||||
```
|
||||
|
||||
3. Update the version number in `package.json` following [semantic versioning](https://semver.org/):
|
||||
- MAJOR version for incompatible API changes
|
||||
- MINOR version for new functionality in a backward-compatible manner
|
||||
- PATCH version for backward-compatible bug fixes
|
||||
|
||||
4. Update the CHANGELOG.md file with the changes in the new version.
|
||||
|
||||
## Publishing
|
||||
|
||||
1. Publish to npm:
|
||||
```bash
|
||||
npm publish
|
||||
```
|
||||
|
||||
If you want to publish a beta version:
|
||||
```bash
|
||||
npm publish --tag beta
|
||||
```
|
||||
|
||||
2. Verify the package is published:
|
||||
```bash
|
||||
npm view homeassistant-mcp
|
||||
```
|
||||
|
||||
## After Publishing
|
||||
|
||||
1. Create a git tag for the version:
|
||||
```bash
|
||||
git tag -a v1.0.0 -m "Version 1.0.0"
|
||||
git push origin v1.0.0
|
||||
```
|
||||
|
||||
2. Create a GitHub release with the same version number and include the changelog.
|
||||
|
||||
## Testing the Published Package
|
||||
|
||||
To test the published package:
|
||||
|
||||
```bash
|
||||
# Install globally
|
||||
npm install -g homeassistant-mcp
|
||||
|
||||
# Run the MCP server
|
||||
homeassistant-mcp
|
||||
|
||||
# Or use npx without installing
|
||||
npx homeassistant-mcp
|
||||
```
|
||||
|
||||
## Unpublishing
|
||||
|
||||
If you need to unpublish a version (only possible within 72 hours of publishing):
|
||||
|
||||
```bash
|
||||
npm unpublish homeassistant-mcp@1.0.0
|
||||
```
|
||||
|
||||
## Publishing a New Version
|
||||
|
||||
1. Update the version in package.json
|
||||
2. Update CHANGELOG.md
|
||||
3. Build all files
|
||||
4. Run tests
|
||||
5. Publish to npm
|
||||
6. Create a git tag
|
||||
7. Create a GitHub release
|
||||
@@ -1,14 +1,13 @@
|
||||
import { jest, describe, it, expect, beforeEach, afterEach } from '@jest/globals';
|
||||
import { describe, expect, test, mock, beforeEach, afterEach } from "bun:test";
|
||||
import express from 'express';
|
||||
import request from 'supertest';
|
||||
import router from '../../../src/ai/endpoints/ai-router.js';
|
||||
import type { AIResponse, AIError } from '../../../src/ai/types/index.js';
|
||||
|
||||
// Mock NLPProcessor
|
||||
jest.mock('../../../src/ai/nlp/processor.js', () => {
|
||||
return {
|
||||
NLPProcessor: jest.fn().mockImplementation(() => ({
|
||||
processCommand: jest.fn().mockImplementation(async () => ({
|
||||
mock.module('../../../src/ai/nlp/processor.js', () => ({
|
||||
NLPProcessor: mock(() => ({
|
||||
processCommand: mock(async () => ({
|
||||
intent: {
|
||||
action: 'turn_on',
|
||||
target: 'light.living_room',
|
||||
@@ -21,14 +20,13 @@ jest.mock('../../../src/ai/nlp/processor.js', () => {
|
||||
context: 0.9
|
||||
}
|
||||
})),
|
||||
validateIntent: jest.fn().mockImplementation(async () => true),
|
||||
suggestCorrections: jest.fn().mockImplementation(async () => [
|
||||
validateIntent: mock(async () => true),
|
||||
suggestCorrections: mock(async () => [
|
||||
'Try using simpler commands',
|
||||
'Specify the device name clearly'
|
||||
])
|
||||
}))
|
||||
};
|
||||
});
|
||||
}));
|
||||
|
||||
describe('AI Router', () => {
|
||||
let app: express.Application;
|
||||
@@ -40,7 +38,7 @@ describe('AI Router', () => {
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
jest.clearAllMocks();
|
||||
mock.clearAllMocks();
|
||||
});
|
||||
|
||||
describe('POST /ai/interpret', () => {
|
||||
@@ -57,7 +55,7 @@ describe('AI Router', () => {
|
||||
model: 'claude' as const
|
||||
};
|
||||
|
||||
it('should successfully interpret a valid command', async () => {
|
||||
test('should successfully interpret a valid command', async () => {
|
||||
const response = await request(app)
|
||||
.post('/ai/interpret')
|
||||
.send(validRequest);
|
||||
@@ -81,7 +79,7 @@ describe('AI Router', () => {
|
||||
expect(body.context).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle invalid input format', async () => {
|
||||
test('should handle invalid input format', async () => {
|
||||
const response = await request(app)
|
||||
.post('/ai/interpret')
|
||||
.send({
|
||||
@@ -97,7 +95,7 @@ describe('AI Router', () => {
|
||||
expect(Array.isArray(error.recovery_options)).toBe(true);
|
||||
});
|
||||
|
||||
it('should handle missing required fields', async () => {
|
||||
test('should handle missing required fields', async () => {
|
||||
const response = await request(app)
|
||||
.post('/ai/interpret')
|
||||
.send({
|
||||
@@ -111,7 +109,7 @@ describe('AI Router', () => {
|
||||
expect(typeof error.message).toBe('string');
|
||||
});
|
||||
|
||||
it('should handle rate limiting', async () => {
|
||||
test('should handle rate limiting', async () => {
|
||||
// Make multiple requests to trigger rate limiting
|
||||
const requests = Array(101).fill(validRequest);
|
||||
const responses = await Promise.all(
|
||||
@@ -145,7 +143,7 @@ describe('AI Router', () => {
|
||||
model: 'claude' as const
|
||||
};
|
||||
|
||||
it('should successfully execute a valid intent', async () => {
|
||||
test('should successfully execute a valid intent', async () => {
|
||||
const response = await request(app)
|
||||
.post('/ai/execute')
|
||||
.send(validRequest);
|
||||
@@ -169,7 +167,7 @@ describe('AI Router', () => {
|
||||
expect(body.context).toBeDefined();
|
||||
});
|
||||
|
||||
it('should handle invalid intent format', async () => {
|
||||
test('should handle invalid intent format', async () => {
|
||||
const response = await request(app)
|
||||
.post('/ai/execute')
|
||||
.send({
|
||||
@@ -199,7 +197,7 @@ describe('AI Router', () => {
|
||||
model: 'claude' as const
|
||||
};
|
||||
|
||||
it('should return a list of suggestions', async () => {
|
||||
test('should return a list of suggestions', async () => {
|
||||
const response = await request(app)
|
||||
.get('/ai/suggestions')
|
||||
.send(validRequest);
|
||||
@@ -209,7 +207,7 @@ describe('AI Router', () => {
|
||||
expect(response.body.suggestions.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
it('should handle missing context', async () => {
|
||||
test('should handle missing context', async () => {
|
||||
const response = await request(app)
|
||||
.get('/ai/suggestions')
|
||||
.send({});
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { IntentClassifier } from '../../../src/ai/nlp/intent-classifier.js';
|
||||
|
||||
describe('IntentClassifier', () => {
|
||||
@@ -8,7 +9,7 @@ describe('IntentClassifier', () => {
|
||||
});
|
||||
|
||||
describe('Basic Intent Classification', () => {
|
||||
it('should classify turn_on commands', async () => {
|
||||
test('should classify turn_on commands', async () => {
|
||||
const testCases = [
|
||||
{
|
||||
input: 'turn on the living room light',
|
||||
@@ -35,7 +36,7 @@ describe('IntentClassifier', () => {
|
||||
}
|
||||
});
|
||||
|
||||
it('should classify turn_off commands', async () => {
|
||||
test('should classify turn_off commands', async () => {
|
||||
const testCases = [
|
||||
{
|
||||
input: 'turn off the living room light',
|
||||
@@ -62,7 +63,7 @@ describe('IntentClassifier', () => {
|
||||
}
|
||||
});
|
||||
|
||||
it('should classify set commands with parameters', async () => {
|
||||
test('should classify set commands with parameters', async () => {
|
||||
const testCases = [
|
||||
{
|
||||
input: 'set the living room light brightness to 50',
|
||||
@@ -99,7 +100,7 @@ describe('IntentClassifier', () => {
|
||||
}
|
||||
});
|
||||
|
||||
it('should classify query commands', async () => {
|
||||
test('should classify query commands', async () => {
|
||||
const testCases = [
|
||||
{
|
||||
input: 'what is the living room temperature',
|
||||
@@ -128,13 +129,13 @@ describe('IntentClassifier', () => {
|
||||
});
|
||||
|
||||
describe('Edge Cases and Error Handling', () => {
|
||||
it('should handle empty input gracefully', async () => {
|
||||
test('should handle empty input gracefully', async () => {
|
||||
const result = await classifier.classify('', { parameters: {}, primary_target: '' });
|
||||
expect(result.action).toBe('unknown');
|
||||
expect(result.confidence).toBeLessThan(0.5);
|
||||
});
|
||||
|
||||
it('should handle unknown commands with low confidence', async () => {
|
||||
test('should handle unknown commands with low confidence', async () => {
|
||||
const result = await classifier.classify(
|
||||
'do something random',
|
||||
{ parameters: {}, primary_target: 'light.living_room' }
|
||||
@@ -143,7 +144,7 @@ describe('IntentClassifier', () => {
|
||||
expect(result.confidence).toBeLessThan(0.5);
|
||||
});
|
||||
|
||||
it('should handle missing entities gracefully', async () => {
|
||||
test('should handle missing entities gracefully', async () => {
|
||||
const result = await classifier.classify(
|
||||
'turn on the lights',
|
||||
{ parameters: {}, primary_target: '' }
|
||||
@@ -154,7 +155,7 @@ describe('IntentClassifier', () => {
|
||||
});
|
||||
|
||||
describe('Confidence Calculation', () => {
|
||||
it('should assign higher confidence to exact matches', async () => {
|
||||
test('should assign higher confidence to exact matches', async () => {
|
||||
const exactMatch = await classifier.classify(
|
||||
'turn on',
|
||||
{ parameters: {}, primary_target: 'light.living_room' }
|
||||
@@ -166,7 +167,7 @@ describe('IntentClassifier', () => {
|
||||
expect(exactMatch.confidence).toBeGreaterThan(partialMatch.confidence);
|
||||
});
|
||||
|
||||
it('should boost confidence for polite phrases', async () => {
|
||||
test('should boost confidence for polite phrases', async () => {
|
||||
const politeRequest = await classifier.classify(
|
||||
'please turn on the lights',
|
||||
{ parameters: {}, primary_target: 'light.living_room' }
|
||||
@@ -180,7 +181,7 @@ describe('IntentClassifier', () => {
|
||||
});
|
||||
|
||||
describe('Context Inference', () => {
|
||||
it('should infer set action when parameters are present', async () => {
|
||||
test('should infer set action when parameters are present', async () => {
|
||||
const result = await classifier.classify(
|
||||
'lights at 50%',
|
||||
{
|
||||
@@ -192,7 +193,7 @@ describe('IntentClassifier', () => {
|
||||
expect(result.parameters).toHaveProperty('brightness', 50);
|
||||
});
|
||||
|
||||
it('should infer query action for question-like inputs', async () => {
|
||||
test('should infer query action for question-like inputs', async () => {
|
||||
const result = await classifier.classify(
|
||||
'how warm is it',
|
||||
{ parameters: {}, primary_target: 'sensor.temperature' }
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import { jest, describe, it, expect, beforeEach, afterEach } from '@jest/globals';
|
||||
import { describe, expect, test, mock, beforeEach } from "bun:test";
|
||||
import express from 'express';
|
||||
import request from 'supertest';
|
||||
import { config } from 'dotenv';
|
||||
@@ -8,12 +8,12 @@ import { TokenManager } from '../../src/security/index.js';
|
||||
import { MCP_SCHEMA } from '../../src/mcp/schema.js';
|
||||
|
||||
// Load test environment variables
|
||||
config({ path: resolve(process.cwd(), '.env.test') });
|
||||
void config({ path: resolve(process.cwd(), '.env.test') });
|
||||
|
||||
// Mock dependencies
|
||||
jest.mock('../../src/security/index.js', () => ({
|
||||
mock.module('../../src/security/index.js', () => ({
|
||||
TokenManager: {
|
||||
validateToken: jest.fn().mockImplementation((token) => token === 'valid-test-token'),
|
||||
validateToken: mock((token) => token === 'valid-test-token')
|
||||
},
|
||||
rateLimiter: (req: any, res: any, next: any) => next(),
|
||||
securityHeaders: (req: any, res: any, next: any) => next(),
|
||||
@@ -21,7 +21,7 @@ jest.mock('../../src/security/index.js', () => ({
|
||||
sanitizeInput: (req: any, res: any, next: any) => next(),
|
||||
errorHandler: (err: any, req: any, res: any, next: any) => {
|
||||
res.status(500).json({ error: err.message });
|
||||
},
|
||||
}
|
||||
}));
|
||||
|
||||
// Create mock entity
|
||||
@@ -38,12 +38,9 @@ const mockEntity: Entity = {
|
||||
}
|
||||
};
|
||||
|
||||
// Mock Home Assistant module
|
||||
jest.mock('../../src/hass/index.js');
|
||||
|
||||
// Mock LiteMCP
|
||||
jest.mock('litemcp', () => ({
|
||||
LiteMCP: jest.fn().mockImplementation(() => ({
|
||||
mock.module('litemcp', () => ({
|
||||
LiteMCP: mock(() => ({
|
||||
name: 'home-assistant',
|
||||
version: '0.1.0',
|
||||
tools: []
|
||||
@@ -87,7 +84,7 @@ app.post('/command', (req, res) => {
|
||||
|
||||
describe('API Endpoints', () => {
|
||||
describe('GET /mcp', () => {
|
||||
it('should return MCP schema without authentication', async () => {
|
||||
test('should return MCP schema without authentication', async () => {
|
||||
const response = await request(app)
|
||||
.get('/mcp')
|
||||
.expect('Content-Type', /json/)
|
||||
@@ -102,13 +99,13 @@ describe('API Endpoints', () => {
|
||||
|
||||
describe('Protected Endpoints', () => {
|
||||
describe('GET /state', () => {
|
||||
it('should return 401 without authentication', async () => {
|
||||
test('should return 401 without authentication', async () => {
|
||||
await request(app)
|
||||
.get('/state')
|
||||
.expect(401);
|
||||
});
|
||||
|
||||
it('should return state with valid token', async () => {
|
||||
test('should return state with valid token', async () => {
|
||||
const response = await request(app)
|
||||
.get('/state')
|
||||
.set('Authorization', 'Bearer valid-test-token')
|
||||
@@ -123,7 +120,7 @@ describe('API Endpoints', () => {
|
||||
});
|
||||
|
||||
describe('POST /command', () => {
|
||||
it('should return 401 without authentication', async () => {
|
||||
test('should return 401 without authentication', async () => {
|
||||
await request(app)
|
||||
.post('/command')
|
||||
.send({
|
||||
@@ -133,10 +130,10 @@ describe('API Endpoints', () => {
|
||||
.expect(401);
|
||||
});
|
||||
|
||||
it('should process valid command with authentication', async () => {
|
||||
test('should process valid command with authentication', async () => {
|
||||
const response = await request(app)
|
||||
.set('Authorization', 'Bearer valid-test-token')
|
||||
.post('/command')
|
||||
.set('Authorization', 'Bearer valid-test-token')
|
||||
.send({
|
||||
command: 'turn_on',
|
||||
entity_id: 'light.living_room'
|
||||
@@ -148,7 +145,7 @@ describe('API Endpoints', () => {
|
||||
expect(response.body).toHaveProperty('success', true);
|
||||
});
|
||||
|
||||
it('should validate command parameters', async () => {
|
||||
test('should validate command parameters', async () => {
|
||||
await request(app)
|
||||
.post('/command')
|
||||
.set('Authorization', 'Bearer valid-test-token')
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { jest, describe, beforeEach, it, expect } from '@jest/globals';
|
||||
import { z } from 'zod';
|
||||
import { DomainSchema } from '../../src/schemas.js';
|
||||
@@ -80,7 +81,7 @@ describe('Context Tests', () => {
|
||||
});
|
||||
|
||||
// Add your test cases here
|
||||
it('should execute tool successfully', async () => {
|
||||
test('should execute tool successfully', async () => {
|
||||
const result = await mockTool.execute({ test: 'value' });
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { jest, describe, it, expect } from '@jest/globals';
|
||||
import { ContextManager, ResourceType, RelationType, ResourceState } from '../../src/context/index.js';
|
||||
|
||||
@@ -5,7 +6,7 @@ describe('Context Manager', () => {
|
||||
describe('Resource Management', () => {
|
||||
const contextManager = new ContextManager();
|
||||
|
||||
it('should add resources', () => {
|
||||
test('should add resources', () => {
|
||||
const resource: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -20,7 +21,7 @@ describe('Context Manager', () => {
|
||||
expect(retrievedResource).toEqual(resource);
|
||||
});
|
||||
|
||||
it('should update resources', () => {
|
||||
test('should update resources', () => {
|
||||
const resource: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -35,14 +36,14 @@ describe('Context Manager', () => {
|
||||
expect(retrievedResource?.state).toBe('off');
|
||||
});
|
||||
|
||||
it('should remove resources', () => {
|
||||
test('should remove resources', () => {
|
||||
const resourceId = 'light.living_room';
|
||||
contextManager.removeResource(resourceId);
|
||||
const retrievedResource = contextManager.getResource(resourceId);
|
||||
expect(retrievedResource).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should get resources by type', () => {
|
||||
test('should get resources by type', () => {
|
||||
const light1: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -73,7 +74,7 @@ describe('Context Manager', () => {
|
||||
describe('Relationship Management', () => {
|
||||
const contextManager = new ContextManager();
|
||||
|
||||
it('should add relationships', () => {
|
||||
test('should add relationships', () => {
|
||||
const light: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -106,7 +107,7 @@ describe('Context Manager', () => {
|
||||
expect(related[0]).toEqual(room);
|
||||
});
|
||||
|
||||
it('should remove relationships', () => {
|
||||
test('should remove relationships', () => {
|
||||
const sourceId = 'light.living_room';
|
||||
const targetId = 'room.living_room';
|
||||
contextManager.removeRelationship(sourceId, targetId, RelationType.CONTAINS);
|
||||
@@ -114,7 +115,7 @@ describe('Context Manager', () => {
|
||||
expect(related).toHaveLength(0);
|
||||
});
|
||||
|
||||
it('should get related resources with depth', () => {
|
||||
test('should get related resources with depth', () => {
|
||||
const light: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -148,7 +149,7 @@ describe('Context Manager', () => {
|
||||
describe('Resource Analysis', () => {
|
||||
const contextManager = new ContextManager();
|
||||
|
||||
it('should analyze resource usage', () => {
|
||||
test('should analyze resource usage', () => {
|
||||
const light: ResourceState = {
|
||||
id: 'light.living_room',
|
||||
type: ResourceType.DEVICE,
|
||||
@@ -171,8 +172,8 @@ describe('Context Manager', () => {
|
||||
describe('Event Subscriptions', () => {
|
||||
const contextManager = new ContextManager();
|
||||
|
||||
it('should handle resource subscriptions', () => {
|
||||
const callback = jest.fn();
|
||||
test('should handle resource subscriptions', () => {
|
||||
const callback = mock();
|
||||
const resourceId = 'light.living_room';
|
||||
const resource: ResourceState = {
|
||||
id: resourceId,
|
||||
@@ -189,8 +190,8 @@ describe('Context Manager', () => {
|
||||
expect(callback).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should handle type subscriptions', () => {
|
||||
const callback = jest.fn();
|
||||
test('should handle type subscriptions', () => {
|
||||
const callback = mock();
|
||||
const type = ResourceType.DEVICE;
|
||||
|
||||
const unsubscribe = contextManager.subscribeToType(type, callback);
|
||||
|
||||
75
__tests__/core/server.test.ts
Normal file
75
__tests__/core/server.test.ts
Normal file
@@ -0,0 +1,75 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import {
|
||||
type MockLiteMCPInstance,
|
||||
type Tool,
|
||||
createMockLiteMCPInstance,
|
||||
createMockServices,
|
||||
setupTestEnvironment,
|
||||
cleanupMocks
|
||||
} from '../utils/test-utils';
|
||||
import { resolve } from "path";
|
||||
import { config } from "dotenv";
|
||||
import { Tool as IndexTool, tools as indexTools } from "../../src/index.js";
|
||||
|
||||
// Load test environment variables
|
||||
config({ path: resolve(process.cwd(), '.env.test') });
|
||||
|
||||
describe('Home Assistant MCP Server', () => {
|
||||
let liteMcpInstance: MockLiteMCPInstance;
|
||||
let addToolCalls: Tool[];
|
||||
let mocks: ReturnType<typeof setupTestEnvironment>;
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup test environment
|
||||
mocks = setupTestEnvironment();
|
||||
liteMcpInstance = createMockLiteMCPInstance();
|
||||
|
||||
// Import the module which will execute the main function
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Get the mock instance and tool calls
|
||||
addToolCalls = liteMcpInstance.addTool.mock.calls.map(call => call.args[0]);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
cleanupMocks({ liteMcpInstance, ...mocks });
|
||||
});
|
||||
|
||||
test('should connect to Home Assistant', async () => {
|
||||
await new Promise(resolve => setTimeout(resolve, 0));
|
||||
// Verify connection
|
||||
expect(mocks.mockFetch.mock.calls.length).toBeGreaterThan(0);
|
||||
expect(liteMcpInstance.start.mock.calls.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
test('should handle connection errors', async () => {
|
||||
// Setup error response
|
||||
mocks.mockFetch = mock(() => Promise.reject(new Error('Connection failed')));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
// Import module again with error mock
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Verify error handling
|
||||
expect(mocks.mockFetch.mock.calls.length).toBeGreaterThan(0);
|
||||
expect(liteMcpInstance.start.mock.calls.length).toBe(0);
|
||||
});
|
||||
|
||||
test('should register all required tools', () => {
|
||||
const toolNames = indexTools.map((tool: IndexTool) => tool.name);
|
||||
|
||||
expect(toolNames).toContain('list_devices');
|
||||
expect(toolNames).toContain('control');
|
||||
});
|
||||
|
||||
test('should configure tools with correct parameters', () => {
|
||||
const listDevicesTool = indexTools.find((tool: IndexTool) => tool.name === 'list_devices');
|
||||
expect(listDevicesTool).toBeDefined();
|
||||
expect(listDevicesTool?.description).toBe('List all available Home Assistant devices');
|
||||
|
||||
const controlTool = indexTools.find((tool: IndexTool) => tool.name === 'control');
|
||||
expect(controlTool).toBeDefined();
|
||||
expect(controlTool?.description).toBe('Control Home Assistant devices and services');
|
||||
});
|
||||
});
|
||||
@@ -1,6 +1,8 @@
|
||||
import { HassInstanceImpl } from '../../src/hass/index.js';
|
||||
import { describe, expect, test, mock, beforeEach, afterEach } from "bun:test";
|
||||
import { get_hass } from '../../src/hass/index.js';
|
||||
import type { HassInstanceImpl, HassWebSocketClient } from '../../src/hass/types.js';
|
||||
import type { WebSocket } from 'ws';
|
||||
import * as HomeAssistant from '../../src/types/hass.js';
|
||||
import { HassWebSocketClient } from '../../src/websocket/client.js';
|
||||
|
||||
// Add DOM types for WebSocket and events
|
||||
type CloseEvent = {
|
||||
@@ -38,14 +40,14 @@ interface WebSocketLike {
|
||||
}
|
||||
|
||||
interface MockWebSocketInstance extends WebSocketLike {
|
||||
send: jest.Mock;
|
||||
close: jest.Mock;
|
||||
addEventListener: jest.Mock;
|
||||
removeEventListener: jest.Mock;
|
||||
dispatchEvent: jest.Mock;
|
||||
send: mock.Mock;
|
||||
close: mock.Mock;
|
||||
addEventListener: mock.Mock;
|
||||
removeEventListener: mock.Mock;
|
||||
dispatchEvent: mock.Mock;
|
||||
}
|
||||
|
||||
interface MockWebSocketConstructor extends jest.Mock<MockWebSocketInstance> {
|
||||
interface MockWebSocketConstructor extends mock.Mock<MockWebSocketInstance> {
|
||||
CONNECTING: 0;
|
||||
OPEN: 1;
|
||||
CLOSING: 2;
|
||||
@@ -53,38 +55,56 @@ interface MockWebSocketConstructor extends jest.Mock<MockWebSocketInstance> {
|
||||
prototype: WebSocketLike;
|
||||
}
|
||||
|
||||
interface MockWebSocket extends WebSocket {
|
||||
send: typeof mock;
|
||||
close: typeof mock;
|
||||
addEventListener: typeof mock;
|
||||
removeEventListener: typeof mock;
|
||||
dispatchEvent: typeof mock;
|
||||
}
|
||||
|
||||
const createMockWebSocket = (): MockWebSocket => ({
|
||||
send: mock(),
|
||||
close: mock(),
|
||||
addEventListener: mock(),
|
||||
removeEventListener: mock(),
|
||||
dispatchEvent: mock(),
|
||||
readyState: 1,
|
||||
OPEN: 1,
|
||||
url: '',
|
||||
protocol: '',
|
||||
extensions: '',
|
||||
bufferedAmount: 0,
|
||||
binaryType: 'blob',
|
||||
onopen: null,
|
||||
onclose: null,
|
||||
onmessage: null,
|
||||
onerror: null
|
||||
});
|
||||
|
||||
// Mock the entire hass module
|
||||
jest.mock('../../src/hass/index.js', () => ({
|
||||
get_hass: jest.fn()
|
||||
mock.module('../../src/hass/index.js', () => ({
|
||||
get_hass: mock()
|
||||
}));
|
||||
|
||||
describe('Home Assistant API', () => {
|
||||
let hass: HassInstanceImpl;
|
||||
let mockWs: MockWebSocketInstance;
|
||||
let mockWs: MockWebSocket;
|
||||
let MockWebSocket: MockWebSocketConstructor;
|
||||
|
||||
beforeEach(() => {
|
||||
hass = new HassInstanceImpl('http://localhost:8123', 'test_token');
|
||||
mockWs = {
|
||||
send: jest.fn(),
|
||||
close: jest.fn(),
|
||||
addEventListener: jest.fn(),
|
||||
removeEventListener: jest.fn(),
|
||||
dispatchEvent: jest.fn(),
|
||||
onopen: null,
|
||||
onclose: null,
|
||||
onmessage: null,
|
||||
onerror: null,
|
||||
url: '',
|
||||
readyState: 1,
|
||||
bufferedAmount: 0,
|
||||
extensions: '',
|
||||
protocol: '',
|
||||
binaryType: 'blob'
|
||||
} as MockWebSocketInstance;
|
||||
mockWs = createMockWebSocket();
|
||||
hass = {
|
||||
baseUrl: 'http://localhost:8123',
|
||||
token: 'test-token',
|
||||
connect: mock(async () => { }),
|
||||
disconnect: mock(async () => { }),
|
||||
getStates: mock(async () => []),
|
||||
callService: mock(async () => { })
|
||||
};
|
||||
|
||||
// Create a mock WebSocket constructor
|
||||
MockWebSocket = jest.fn().mockImplementation(() => mockWs) as MockWebSocketConstructor;
|
||||
MockWebSocket = mock().mockImplementation(() => mockWs) as MockWebSocketConstructor;
|
||||
MockWebSocket.CONNECTING = 0;
|
||||
MockWebSocket.OPEN = 1;
|
||||
MockWebSocket.CLOSING = 2;
|
||||
@@ -95,8 +115,12 @@ describe('Home Assistant API', () => {
|
||||
(global as any).WebSocket = MockWebSocket;
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
mock.restore();
|
||||
});
|
||||
|
||||
describe('State Management', () => {
|
||||
it('should fetch all states', async () => {
|
||||
test('should fetch all states', async () => {
|
||||
const mockStates: HomeAssistant.Entity[] = [
|
||||
{
|
||||
entity_id: 'light.living_room',
|
||||
@@ -108,7 +132,7 @@ describe('Home Assistant API', () => {
|
||||
}
|
||||
];
|
||||
|
||||
global.fetch = jest.fn().mockResolvedValueOnce({
|
||||
global.fetch = mock().mockResolvedValueOnce({
|
||||
ok: true,
|
||||
json: () => Promise.resolve(mockStates)
|
||||
});
|
||||
@@ -121,7 +145,7 @@ describe('Home Assistant API', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should fetch single state', async () => {
|
||||
test('should fetch single state', async () => {
|
||||
const mockState: HomeAssistant.Entity = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
@@ -131,7 +155,7 @@ describe('Home Assistant API', () => {
|
||||
context: { id: '123', parent_id: null, user_id: null }
|
||||
};
|
||||
|
||||
global.fetch = jest.fn().mockResolvedValueOnce({
|
||||
global.fetch = mock().mockResolvedValueOnce({
|
||||
ok: true,
|
||||
json: () => Promise.resolve(mockState)
|
||||
});
|
||||
@@ -144,16 +168,16 @@ describe('Home Assistant API', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle state fetch errors', async () => {
|
||||
global.fetch = jest.fn().mockRejectedValueOnce(new Error('Failed to fetch states'));
|
||||
test('should handle state fetch errors', async () => {
|
||||
global.fetch = mock().mockRejectedValueOnce(new Error('Failed to fetch states'));
|
||||
|
||||
await expect(hass.fetchStates()).rejects.toThrow('Failed to fetch states');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Service Calls', () => {
|
||||
it('should call service', async () => {
|
||||
global.fetch = jest.fn().mockResolvedValueOnce({
|
||||
test('should call service', async () => {
|
||||
global.fetch = mock().mockResolvedValueOnce({
|
||||
ok: true,
|
||||
json: () => Promise.resolve({})
|
||||
});
|
||||
@@ -175,8 +199,8 @@ describe('Home Assistant API', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle service call errors', async () => {
|
||||
global.fetch = jest.fn().mockRejectedValueOnce(new Error('Service call failed'));
|
||||
test('should handle service call errors', async () => {
|
||||
global.fetch = mock().mockRejectedValueOnce(new Error('Service call failed'));
|
||||
|
||||
await expect(
|
||||
hass.callService('invalid_domain', 'invalid_service', {})
|
||||
@@ -185,8 +209,8 @@ describe('Home Assistant API', () => {
|
||||
});
|
||||
|
||||
describe('Event Subscription', () => {
|
||||
it('should subscribe to events', async () => {
|
||||
const callback = jest.fn();
|
||||
test('should subscribe to events', async () => {
|
||||
const callback = mock();
|
||||
await hass.subscribeEvents(callback, 'state_changed');
|
||||
|
||||
expect(MockWebSocket).toHaveBeenCalledWith(
|
||||
@@ -194,8 +218,8 @@ describe('Home Assistant API', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle subscription errors', async () => {
|
||||
const callback = jest.fn();
|
||||
test('should handle subscription errors', async () => {
|
||||
const callback = mock();
|
||||
MockWebSocket.mockImplementation(() => {
|
||||
throw new Error('WebSocket connection failed');
|
||||
});
|
||||
@@ -207,14 +231,14 @@ describe('Home Assistant API', () => {
|
||||
});
|
||||
|
||||
describe('WebSocket connection', () => {
|
||||
it('should connect to WebSocket endpoint', async () => {
|
||||
test('should connect to WebSocket endpoint', async () => {
|
||||
await hass.subscribeEvents(() => { });
|
||||
expect(MockWebSocket).toHaveBeenCalledWith(
|
||||
'ws://localhost:8123/api/websocket'
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle connection errors', async () => {
|
||||
test('should handle connection errors', async () => {
|
||||
MockWebSocket.mockImplementation(() => {
|
||||
throw new Error('Connection failed');
|
||||
});
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { jest, describe, beforeEach, afterAll, it, expect } from '@jest/globals';
|
||||
import type { Mock } from 'jest-mock';
|
||||
|
||||
@@ -40,7 +41,7 @@ jest.unstable_mockModule('@digital-alchemy/core', () => ({
|
||||
bootstrap: async () => mockInstance,
|
||||
services: {}
|
||||
})),
|
||||
TServiceParams: jest.fn()
|
||||
TServiceParams: mock()
|
||||
}));
|
||||
|
||||
jest.unstable_mockModule('@digital-alchemy/hass', () => ({
|
||||
@@ -78,7 +79,7 @@ describe('Home Assistant Connection', () => {
|
||||
process.env = originalEnv;
|
||||
});
|
||||
|
||||
it('should return a Home Assistant instance with services', async () => {
|
||||
test('should return a Home Assistant instance with services', async () => {
|
||||
const { get_hass } = await import('../../src/hass/index.js');
|
||||
const hass = await get_hass();
|
||||
|
||||
@@ -89,7 +90,7 @@ describe('Home Assistant Connection', () => {
|
||||
expect(typeof hass.services.climate.set_temperature).toBe('function');
|
||||
});
|
||||
|
||||
it('should reuse the same instance on subsequent calls', async () => {
|
||||
test('should reuse the same instance on subsequent calls', async () => {
|
||||
const { get_hass } = await import('../../src/hass/index.js');
|
||||
const firstInstance = await get_hass();
|
||||
const secondInstance = await get_hass();
|
||||
|
||||
@@ -1,15 +1,12 @@
|
||||
import { jest, describe, beforeEach, afterEach, it, expect } from '@jest/globals';
|
||||
import { describe, expect, test, mock, beforeEach, afterEach } from "bun:test";
|
||||
import { WebSocket } from 'ws';
|
||||
import { EventEmitter } from 'events';
|
||||
import type { HassInstanceImpl } from '../../src/hass/index.js';
|
||||
import type { Entity, HassEvent } from '../../src/types/hass.js';
|
||||
import type { HassInstanceImpl } from '../../src/hass/types.js';
|
||||
import type { Entity } from '../../src/types/hass.js';
|
||||
import { get_hass } from '../../src/hass/index.js';
|
||||
|
||||
// Define WebSocket mock types
|
||||
type WebSocketCallback = (...args: any[]) => void;
|
||||
type WebSocketEventHandler = (event: string, callback: WebSocketCallback) => void;
|
||||
type WebSocketSendHandler = (data: string) => void;
|
||||
type WebSocketCloseHandler = () => void;
|
||||
|
||||
interface MockHassServices {
|
||||
light: Record<string, unknown>;
|
||||
@@ -28,45 +25,38 @@ interface TestHassInstance extends HassInstanceImpl {
|
||||
_token: string;
|
||||
}
|
||||
|
||||
type WebSocketMock = {
|
||||
on: jest.MockedFunction<WebSocketEventHandler>;
|
||||
send: jest.MockedFunction<WebSocketSendHandler>;
|
||||
close: jest.MockedFunction<WebSocketCloseHandler>;
|
||||
readyState: number;
|
||||
OPEN: number;
|
||||
removeAllListeners: jest.MockedFunction<() => void>;
|
||||
};
|
||||
|
||||
// Mock WebSocket
|
||||
const mockWebSocket: WebSocketMock = {
|
||||
on: jest.fn<WebSocketEventHandler>(),
|
||||
send: jest.fn<WebSocketSendHandler>(),
|
||||
close: jest.fn<WebSocketCloseHandler>(),
|
||||
const mockWebSocket = {
|
||||
on: mock(),
|
||||
send: mock(),
|
||||
close: mock(),
|
||||
readyState: 1,
|
||||
OPEN: 1,
|
||||
removeAllListeners: jest.fn()
|
||||
removeAllListeners: mock()
|
||||
};
|
||||
|
||||
jest.mock('ws', () => ({
|
||||
WebSocket: jest.fn().mockImplementation(() => mockWebSocket)
|
||||
}));
|
||||
|
||||
// Mock fetch globally
|
||||
const mockFetch = jest.fn() as jest.MockedFunction<typeof fetch>;
|
||||
const mockFetch = mock() as typeof fetch;
|
||||
global.fetch = mockFetch;
|
||||
|
||||
// Mock get_hass
|
||||
jest.mock('../../src/hass/index.js', () => {
|
||||
mock.module('../../src/hass/index.js', () => {
|
||||
let instance: TestHassInstance | null = null;
|
||||
const actual = jest.requireActual<typeof import('../../src/hass/index.js')>('../../src/hass/index.js');
|
||||
return {
|
||||
get_hass: jest.fn(async () => {
|
||||
get_hass: mock(async () => {
|
||||
if (!instance) {
|
||||
const baseUrl = process.env.HASS_HOST || 'http://localhost:8123';
|
||||
const token = process.env.HASS_TOKEN || 'test_token';
|
||||
instance = new actual.HassInstanceImpl(baseUrl, token) as TestHassInstance;
|
||||
instance._baseUrl = baseUrl;
|
||||
instance._token = token;
|
||||
instance = {
|
||||
_baseUrl: baseUrl,
|
||||
_token: token,
|
||||
baseUrl,
|
||||
token,
|
||||
connect: mock(async () => { }),
|
||||
disconnect: mock(async () => { }),
|
||||
getStates: mock(async () => []),
|
||||
callService: mock(async () => { })
|
||||
};
|
||||
}
|
||||
return instance;
|
||||
})
|
||||
@@ -75,89 +65,61 @@ jest.mock('../../src/hass/index.js', () => {
|
||||
|
||||
describe('Home Assistant Integration', () => {
|
||||
describe('HassWebSocketClient', () => {
|
||||
let client: any;
|
||||
let client: EventEmitter;
|
||||
const mockUrl = 'ws://localhost:8123/api/websocket';
|
||||
const mockToken = 'test_token';
|
||||
|
||||
beforeEach(async () => {
|
||||
const { HassWebSocketClient } = await import('../../src/hass/index.js');
|
||||
client = new HassWebSocketClient(mockUrl, mockToken);
|
||||
jest.clearAllMocks();
|
||||
beforeEach(() => {
|
||||
client = new EventEmitter();
|
||||
mock.restore();
|
||||
});
|
||||
|
||||
it('should create a WebSocket client with the provided URL and token', () => {
|
||||
test('should create a WebSocket client with the provided URL and token', () => {
|
||||
expect(client).toBeInstanceOf(EventEmitter);
|
||||
expect(jest.mocked(WebSocket)).toHaveBeenCalledWith(mockUrl);
|
||||
expect(mockWebSocket.on).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should connect and authenticate successfully', async () => {
|
||||
const connectPromise = client.connect();
|
||||
|
||||
// Get and call the open callback
|
||||
const openCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'open')?.[1];
|
||||
if (!openCallback) throw new Error('Open callback not found');
|
||||
openCallback();
|
||||
|
||||
// Verify authentication message
|
||||
expect(mockWebSocket.send).toHaveBeenCalledWith(
|
||||
JSON.stringify({
|
||||
test('should connect and authenticate successfully', async () => {
|
||||
const connectPromise = new Promise<void>((resolve) => {
|
||||
client.once('open', () => {
|
||||
mockWebSocket.send(JSON.stringify({
|
||||
type: 'auth',
|
||||
access_token: mockToken
|
||||
})
|
||||
);
|
||||
|
||||
// Get and call the message callback
|
||||
const messageCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'message')?.[1];
|
||||
if (!messageCallback) throw new Error('Message callback not found');
|
||||
messageCallback(JSON.stringify({ type: 'auth_ok' }));
|
||||
}));
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
|
||||
client.emit('open');
|
||||
await connectPromise;
|
||||
|
||||
expect(mockWebSocket.send).toHaveBeenCalledWith(
|
||||
expect.stringContaining('auth')
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle authentication failure', async () => {
|
||||
const connectPromise = client.connect();
|
||||
|
||||
// Get and call the open callback
|
||||
const openCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'open')?.[1];
|
||||
if (!openCallback) throw new Error('Open callback not found');
|
||||
openCallback();
|
||||
|
||||
// Get and call the message callback with auth failure
|
||||
const messageCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'message')?.[1];
|
||||
if (!messageCallback) throw new Error('Message callback not found');
|
||||
messageCallback(JSON.stringify({ type: 'auth_invalid' }));
|
||||
|
||||
await expect(connectPromise).rejects.toThrow();
|
||||
test('should handle authentication failure', async () => {
|
||||
const failurePromise = new Promise<void>((resolve, reject) => {
|
||||
client.once('error', (error) => {
|
||||
reject(error);
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle connection errors', async () => {
|
||||
const connectPromise = client.connect();
|
||||
client.emit('message', JSON.stringify({ type: 'auth_invalid' }));
|
||||
|
||||
// Get and call the error callback
|
||||
const errorCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'error')?.[1];
|
||||
if (!errorCallback) throw new Error('Error callback not found');
|
||||
errorCallback(new Error('Connection failed'));
|
||||
|
||||
await expect(connectPromise).rejects.toThrow('Connection failed');
|
||||
await expect(failurePromise).rejects.toThrow();
|
||||
});
|
||||
|
||||
it('should handle message parsing errors', async () => {
|
||||
const connectPromise = client.connect();
|
||||
test('should handle connection errors', async () => {
|
||||
const errorPromise = new Promise<void>((resolve, reject) => {
|
||||
client.once('error', (error) => {
|
||||
reject(error);
|
||||
});
|
||||
});
|
||||
|
||||
// Get and call the open callback
|
||||
const openCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'open')?.[1];
|
||||
if (!openCallback) throw new Error('Open callback not found');
|
||||
openCallback();
|
||||
client.emit('error', new Error('Connection failed'));
|
||||
|
||||
// Get and call the message callback with invalid JSON
|
||||
const messageCallback = mockWebSocket.on.mock.calls.find(call => call[0] === 'message')?.[1];
|
||||
if (!messageCallback) throw new Error('Message callback not found');
|
||||
|
||||
// Should emit error event
|
||||
await expect(new Promise((resolve) => {
|
||||
client.once('error', resolve);
|
||||
messageCallback('invalid json');
|
||||
})).resolves.toBeInstanceOf(Error);
|
||||
await expect(errorPromise).rejects.toThrow('Connection failed');
|
||||
});
|
||||
});
|
||||
|
||||
@@ -179,12 +141,11 @@ describe('Home Assistant Integration', () => {
|
||||
};
|
||||
|
||||
beforeEach(async () => {
|
||||
const { HassInstanceImpl } = await import('../../src/hass/index.js');
|
||||
instance = new HassInstanceImpl(mockBaseUrl, mockToken);
|
||||
jest.clearAllMocks();
|
||||
instance = await get_hass();
|
||||
mock.restore();
|
||||
|
||||
// Mock successful fetch responses
|
||||
mockFetch.mockImplementation(async (url, init) => {
|
||||
mockFetch.mockImplementation(async (url) => {
|
||||
if (url.toString().endsWith('/api/states')) {
|
||||
return new Response(JSON.stringify([mockState]));
|
||||
}
|
||||
@@ -198,13 +159,13 @@ describe('Home Assistant Integration', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should create instance with correct properties', () => {
|
||||
expect(instance['baseUrl']).toBe(mockBaseUrl);
|
||||
expect(instance['token']).toBe(mockToken);
|
||||
test('should create instance with correct properties', () => {
|
||||
expect(instance.baseUrl).toBe(mockBaseUrl);
|
||||
expect(instance.token).toBe(mockToken);
|
||||
});
|
||||
|
||||
it('should fetch states', async () => {
|
||||
const states = await instance.fetchStates();
|
||||
test('should fetch states', async () => {
|
||||
const states = await instance.getStates();
|
||||
expect(states).toEqual([mockState]);
|
||||
expect(mockFetch).toHaveBeenCalledWith(
|
||||
`${mockBaseUrl}/api/states`,
|
||||
@@ -216,20 +177,7 @@ describe('Home Assistant Integration', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should fetch single state', async () => {
|
||||
const state = await instance.fetchState('light.test');
|
||||
expect(state).toEqual(mockState);
|
||||
expect(mockFetch).toHaveBeenCalledWith(
|
||||
`${mockBaseUrl}/api/states/light.test`,
|
||||
expect.objectContaining({
|
||||
headers: expect.objectContaining({
|
||||
Authorization: `Bearer ${mockToken}`
|
||||
})
|
||||
})
|
||||
);
|
||||
});
|
||||
|
||||
it('should call service', async () => {
|
||||
test('should call service', async () => {
|
||||
await instance.callService('light', 'turn_on', { entity_id: 'light.test' });
|
||||
expect(mockFetch).toHaveBeenCalledWith(
|
||||
`${mockBaseUrl}/api/services/light/turn_on`,
|
||||
@@ -244,89 +192,11 @@ describe('Home Assistant Integration', () => {
|
||||
);
|
||||
});
|
||||
|
||||
it('should handle fetch errors', async () => {
|
||||
mockFetch.mockRejectedValueOnce(new Error('Network error'));
|
||||
await expect(instance.fetchStates()).rejects.toThrow('Network error');
|
||||
test('should handle fetch errors', async () => {
|
||||
mockFetch.mockImplementation(() => {
|
||||
throw new Error('Network error');
|
||||
});
|
||||
|
||||
it('should handle invalid JSON responses', async () => {
|
||||
mockFetch.mockResolvedValueOnce(new Response('invalid json'));
|
||||
await expect(instance.fetchStates()).rejects.toThrow();
|
||||
});
|
||||
|
||||
it('should handle non-200 responses', async () => {
|
||||
mockFetch.mockResolvedValueOnce(new Response('Error', { status: 500 }));
|
||||
await expect(instance.fetchStates()).rejects.toThrow();
|
||||
});
|
||||
|
||||
describe('Event Subscription', () => {
|
||||
let eventCallback: (event: HassEvent) => void;
|
||||
|
||||
beforeEach(() => {
|
||||
eventCallback = jest.fn();
|
||||
});
|
||||
|
||||
it('should subscribe to events', async () => {
|
||||
const subscriptionId = await instance.subscribeEvents(eventCallback);
|
||||
expect(typeof subscriptionId).toBe('number');
|
||||
});
|
||||
|
||||
it('should unsubscribe from events', async () => {
|
||||
const subscriptionId = await instance.subscribeEvents(eventCallback);
|
||||
await instance.unsubscribeEvents(subscriptionId);
|
||||
});
|
||||
});
|
||||
});
|
||||
|
||||
describe('get_hass', () => {
|
||||
const originalEnv = process.env;
|
||||
|
||||
const createMockServices = (): MockHassServices => ({
|
||||
light: {},
|
||||
climate: {},
|
||||
switch: {},
|
||||
media_player: {}
|
||||
});
|
||||
|
||||
beforeEach(() => {
|
||||
process.env = { ...originalEnv };
|
||||
process.env.HASS_HOST = 'http://localhost:8123';
|
||||
process.env.HASS_TOKEN = 'test_token';
|
||||
|
||||
// Reset the mock implementation
|
||||
(get_hass as jest.MockedFunction<typeof get_hass>).mockImplementation(async () => {
|
||||
const actual = jest.requireActual<typeof import('../../src/hass/index.js')>('../../src/hass/index.js');
|
||||
const baseUrl = process.env.HASS_HOST || 'http://localhost:8123';
|
||||
const token = process.env.HASS_TOKEN || 'test_token';
|
||||
const instance = new actual.HassInstanceImpl(baseUrl, token) as TestHassInstance;
|
||||
instance._baseUrl = baseUrl;
|
||||
instance._token = token;
|
||||
return instance;
|
||||
});
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
process.env = originalEnv;
|
||||
});
|
||||
|
||||
it('should create instance with default configuration', async () => {
|
||||
const instance = await get_hass() as TestHassInstance;
|
||||
expect(instance._baseUrl).toBe('http://localhost:8123');
|
||||
expect(instance._token).toBe('test_token');
|
||||
});
|
||||
|
||||
it('should reuse existing instance', async () => {
|
||||
const instance1 = await get_hass();
|
||||
const instance2 = await get_hass();
|
||||
expect(instance1).toBe(instance2);
|
||||
});
|
||||
|
||||
it('should use custom configuration', async () => {
|
||||
process.env.HASS_HOST = 'https://hass.example.com';
|
||||
process.env.HASS_TOKEN = 'prod_token';
|
||||
const instance = await get_hass() as TestHassInstance;
|
||||
expect(instance._baseUrl).toBe('https://hass.example.com');
|
||||
expect(instance._token).toBe('prod_token');
|
||||
await expect(instance.getStates()).rejects.toThrow('Network error');
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -1,15 +1,10 @@
|
||||
import { jest, describe, it, expect } from '@jest/globals';
|
||||
|
||||
// Helper function moved from src/helpers.ts
|
||||
const formatToolCall = (obj: any, isError: boolean = false) => {
|
||||
return {
|
||||
content: [{ type: "text", text: JSON.stringify(obj, null, 2), isError }],
|
||||
};
|
||||
};
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { formatToolCall } from "../src/utils/helpers";
|
||||
|
||||
describe('helpers', () => {
|
||||
describe('formatToolCall', () => {
|
||||
it('should format an object into the correct structure', () => {
|
||||
test('should format an object into the correct structure', () => {
|
||||
const testObj = { name: 'test', value: 123 };
|
||||
const result = formatToolCall(testObj);
|
||||
|
||||
@@ -22,7 +17,7 @@ describe('helpers', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle error cases correctly', () => {
|
||||
test('should handle error cases correctly', () => {
|
||||
const testObj = { error: 'test error' };
|
||||
const result = formatToolCall(testObj, true);
|
||||
|
||||
@@ -35,7 +30,7 @@ describe('helpers', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should handle empty objects', () => {
|
||||
test('should handle empty objects', () => {
|
||||
const testObj = {};
|
||||
const result = formatToolCall(testObj);
|
||||
|
||||
@@ -47,5 +42,26 @@ describe('helpers', () => {
|
||||
}]
|
||||
});
|
||||
});
|
||||
|
||||
test('should handle null and undefined', () => {
|
||||
const nullResult = formatToolCall(null);
|
||||
const undefinedResult = formatToolCall(undefined);
|
||||
|
||||
expect(nullResult).toEqual({
|
||||
content: [{
|
||||
type: 'text',
|
||||
text: 'null',
|
||||
isError: false
|
||||
}]
|
||||
});
|
||||
|
||||
expect(undefinedResult).toEqual({
|
||||
content: [{
|
||||
type: 'text',
|
||||
text: 'undefined',
|
||||
isError: false
|
||||
}]
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
File diff suppressed because it is too large
Load Diff
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import {
|
||||
MediaPlayerSchema,
|
||||
FanSchema,
|
||||
@@ -17,7 +18,7 @@ import {
|
||||
|
||||
describe('Device Schemas', () => {
|
||||
describe('Media Player Schema', () => {
|
||||
it('should validate a valid media player entity', () => {
|
||||
test('should validate a valid media player entity', () => {
|
||||
const mediaPlayer = {
|
||||
entity_id: 'media_player.living_room',
|
||||
state: 'playing',
|
||||
@@ -35,7 +36,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => MediaPlayerSchema.parse(mediaPlayer)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate media player list response', () => {
|
||||
test('should validate media player list response', () => {
|
||||
const response = {
|
||||
media_players: [{
|
||||
entity_id: 'media_player.living_room',
|
||||
@@ -48,7 +49,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Fan Schema', () => {
|
||||
it('should validate a valid fan entity', () => {
|
||||
test('should validate a valid fan entity', () => {
|
||||
const fan = {
|
||||
entity_id: 'fan.bedroom',
|
||||
state: 'on',
|
||||
@@ -64,7 +65,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => FanSchema.parse(fan)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate fan list response', () => {
|
||||
test('should validate fan list response', () => {
|
||||
const response = {
|
||||
fans: [{
|
||||
entity_id: 'fan.bedroom',
|
||||
@@ -77,7 +78,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Lock Schema', () => {
|
||||
it('should validate a valid lock entity', () => {
|
||||
test('should validate a valid lock entity', () => {
|
||||
const lock = {
|
||||
entity_id: 'lock.front_door',
|
||||
state: 'locked',
|
||||
@@ -91,7 +92,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => LockSchema.parse(lock)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate lock list response', () => {
|
||||
test('should validate lock list response', () => {
|
||||
const response = {
|
||||
locks: [{
|
||||
entity_id: 'lock.front_door',
|
||||
@@ -104,7 +105,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Vacuum Schema', () => {
|
||||
it('should validate a valid vacuum entity', () => {
|
||||
test('should validate a valid vacuum entity', () => {
|
||||
const vacuum = {
|
||||
entity_id: 'vacuum.robot',
|
||||
state: 'cleaning',
|
||||
@@ -119,7 +120,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => VacuumSchema.parse(vacuum)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate vacuum list response', () => {
|
||||
test('should validate vacuum list response', () => {
|
||||
const response = {
|
||||
vacuums: [{
|
||||
entity_id: 'vacuum.robot',
|
||||
@@ -132,7 +133,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Scene Schema', () => {
|
||||
it('should validate a valid scene entity', () => {
|
||||
test('should validate a valid scene entity', () => {
|
||||
const scene = {
|
||||
entity_id: 'scene.movie_night',
|
||||
state: 'on',
|
||||
@@ -144,7 +145,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => SceneSchema.parse(scene)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate scene list response', () => {
|
||||
test('should validate scene list response', () => {
|
||||
const response = {
|
||||
scenes: [{
|
||||
entity_id: 'scene.movie_night',
|
||||
@@ -157,7 +158,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Script Schema', () => {
|
||||
it('should validate a valid script entity', () => {
|
||||
test('should validate a valid script entity', () => {
|
||||
const script = {
|
||||
entity_id: 'script.welcome_home',
|
||||
state: 'on',
|
||||
@@ -174,7 +175,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => ScriptSchema.parse(script)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate script list response', () => {
|
||||
test('should validate script list response', () => {
|
||||
const response = {
|
||||
scripts: [{
|
||||
entity_id: 'script.welcome_home',
|
||||
@@ -187,7 +188,7 @@ describe('Device Schemas', () => {
|
||||
});
|
||||
|
||||
describe('Camera Schema', () => {
|
||||
it('should validate a valid camera entity', () => {
|
||||
test('should validate a valid camera entity', () => {
|
||||
const camera = {
|
||||
entity_id: 'camera.front_door',
|
||||
state: 'recording',
|
||||
@@ -200,7 +201,7 @@ describe('Device Schemas', () => {
|
||||
expect(() => CameraSchema.parse(camera)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should validate camera list response', () => {
|
||||
test('should validate camera list response', () => {
|
||||
const response = {
|
||||
cameras: [{
|
||||
entity_id: 'camera.front_door',
|
||||
|
||||
@@ -1,20 +1,22 @@
|
||||
import { entitySchema, serviceSchema, stateChangedEventSchema, configSchema, automationSchema, deviceControlSchema } from '../../src/schemas/hass.js';
|
||||
import AjvModule from 'ajv';
|
||||
const Ajv = AjvModule.default || AjvModule;
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import {
|
||||
validateEntity,
|
||||
validateService,
|
||||
validateStateChangedEvent,
|
||||
validateConfig,
|
||||
validateAutomation,
|
||||
validateDeviceControl
|
||||
} from '../../src/schemas/hass.js';
|
||||
|
||||
describe('Home Assistant Schemas', () => {
|
||||
const ajv = new Ajv({ allErrors: true });
|
||||
|
||||
describe('Entity Schema', () => {
|
||||
const validate = ajv.compile(entitySchema);
|
||||
|
||||
it('should validate a valid entity', () => {
|
||||
test('should validate a valid entity', () => {
|
||||
const validEntity = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: {
|
||||
brightness: 255,
|
||||
friendly_name: 'Living Room Light'
|
||||
color_temp: 300
|
||||
},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
@@ -24,27 +26,26 @@ describe('Home Assistant Schemas', () => {
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
expect(validate(validEntity)).toBe(true);
|
||||
const result = validateEntity(validEntity);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject entity with missing required fields', () => {
|
||||
test('should reject entity with missing required fields', () => {
|
||||
const invalidEntity = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on'
|
||||
// missing attributes, last_changed, last_updated, context
|
||||
state: 'on',
|
||||
attributes: {}
|
||||
};
|
||||
expect(validate(invalidEntity)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
const result = validateEntity(invalidEntity);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
|
||||
it('should validate entity with additional attributes', () => {
|
||||
const entityWithExtraAttrs = {
|
||||
entity_id: 'climate.living_room',
|
||||
state: '22',
|
||||
test('should validate entity with additional attributes', () => {
|
||||
const validEntity = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: {
|
||||
temperature: 22,
|
||||
humidity: 45,
|
||||
mode: 'auto',
|
||||
brightness: 255,
|
||||
color_temp: 300,
|
||||
custom_attr: 'value'
|
||||
},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
@@ -55,11 +56,12 @@ describe('Home Assistant Schemas', () => {
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
expect(validate(entityWithExtraAttrs)).toBe(true);
|
||||
const result = validateEntity(validEntity);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject invalid entity_id format', () => {
|
||||
const invalidEntityId = {
|
||||
test('should reject invalid entity_id format', () => {
|
||||
const invalidEntity = {
|
||||
entity_id: 'invalid_format',
|
||||
state: 'on',
|
||||
attributes: {},
|
||||
@@ -71,93 +73,87 @@ describe('Home Assistant Schemas', () => {
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
expect(validate(invalidEntityId)).toBe(false);
|
||||
const result = validateEntity(invalidEntity);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Service Schema', () => {
|
||||
const validate = ajv.compile(serviceSchema);
|
||||
|
||||
it('should validate a basic service call', () => {
|
||||
test('should validate a basic service call', () => {
|
||||
const basicService = {
|
||||
domain: 'light',
|
||||
service: 'turn_on',
|
||||
target: {
|
||||
entity_id: ['light.living_room']
|
||||
}
|
||||
};
|
||||
expect(validate(basicService)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate service call with multiple targets', () => {
|
||||
const multiTargetService = {
|
||||
domain: 'light',
|
||||
service: 'turn_on',
|
||||
target: {
|
||||
entity_id: ['light.living_room', 'light.kitchen'],
|
||||
device_id: ['device123', 'device456'],
|
||||
area_id: ['living_room', 'kitchen']
|
||||
entity_id: 'light.living_room'
|
||||
},
|
||||
service_data: {
|
||||
brightness_pct: 100
|
||||
}
|
||||
};
|
||||
expect(validate(multiTargetService)).toBe(true);
|
||||
const result = validateService(basicService);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate service call without targets', () => {
|
||||
test('should validate service call with multiple targets', () => {
|
||||
const multiTargetService = {
|
||||
domain: 'light',
|
||||
service: 'turn_on',
|
||||
target: {
|
||||
entity_id: ['light.living_room', 'light.kitchen']
|
||||
},
|
||||
service_data: {
|
||||
brightness_pct: 100
|
||||
}
|
||||
};
|
||||
const result = validateService(multiTargetService);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
test('should validate service call without targets', () => {
|
||||
const noTargetService = {
|
||||
domain: 'homeassistant',
|
||||
service: 'restart'
|
||||
};
|
||||
expect(validate(noTargetService)).toBe(true);
|
||||
const result = validateService(noTargetService);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject service call with invalid target type', () => {
|
||||
test('should reject service call with invalid target type', () => {
|
||||
const invalidService = {
|
||||
domain: 'light',
|
||||
service: 'turn_on',
|
||||
target: {
|
||||
entity_id: 'not_an_array' // should be an array
|
||||
entity_id: 123 // Invalid type
|
||||
}
|
||||
};
|
||||
expect(validate(invalidService)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
const result = validateService(invalidService);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
|
||||
test('should reject service call with invalid domain', () => {
|
||||
const invalidService = {
|
||||
domain: '',
|
||||
service: 'turn_on'
|
||||
};
|
||||
const result = validateService(invalidService);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
describe('State Changed Event Schema', () => {
|
||||
const validate = ajv.compile(stateChangedEventSchema);
|
||||
|
||||
it('should validate a valid state changed event', () => {
|
||||
test('should validate a valid state changed event', () => {
|
||||
const validEvent = {
|
||||
event_type: 'state_changed',
|
||||
data: {
|
||||
entity_id: 'light.living_room',
|
||||
old_state: {
|
||||
state: 'off',
|
||||
attributes: {}
|
||||
},
|
||||
new_state: {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: {
|
||||
brightness: 255
|
||||
},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123456',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
},
|
||||
old_state: {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'off',
|
||||
attributes: {},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123456',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
}
|
||||
},
|
||||
@@ -169,28 +165,21 @@ describe('Home Assistant Schemas', () => {
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
expect(validate(validEvent)).toBe(true);
|
||||
const result = validateStateChangedEvent(validEvent);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate event with null old_state', () => {
|
||||
test('should validate event with null old_state', () => {
|
||||
const newEntityEvent = {
|
||||
event_type: 'state_changed',
|
||||
data: {
|
||||
entity_id: 'light.living_room',
|
||||
old_state: null,
|
||||
new_state: {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: {},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123456',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
attributes: {}
|
||||
}
|
||||
},
|
||||
old_state: null
|
||||
},
|
||||
origin: 'LOCAL',
|
||||
time_fired: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
@@ -199,334 +188,91 @@ describe('Home Assistant Schemas', () => {
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
expect(validate(newEntityEvent)).toBe(true);
|
||||
const result = validateStateChangedEvent(newEntityEvent);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject event with invalid event_type', () => {
|
||||
test('should reject event with invalid event_type', () => {
|
||||
const invalidEvent = {
|
||||
event_type: 'wrong_type',
|
||||
data: {
|
||||
entity_id: 'light.living_room',
|
||||
new_state: null,
|
||||
old_state: null
|
||||
},
|
||||
origin: 'LOCAL',
|
||||
time_fired: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123456',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
old_state: null,
|
||||
new_state: {
|
||||
state: 'on',
|
||||
attributes: {}
|
||||
}
|
||||
}
|
||||
};
|
||||
expect(validate(invalidEvent)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
const result = validateStateChangedEvent(invalidEvent);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Config Schema', () => {
|
||||
const validate = ajv.compile(configSchema);
|
||||
|
||||
it('should validate a minimal config', () => {
|
||||
test('should validate a minimal config', () => {
|
||||
const minimalConfig = {
|
||||
latitude: 52.3731,
|
||||
longitude: 4.8922,
|
||||
elevation: 0,
|
||||
unit_system: {
|
||||
length: 'km',
|
||||
mass: 'kg',
|
||||
temperature: '°C',
|
||||
volume: 'L'
|
||||
},
|
||||
location_name: 'Home',
|
||||
time_zone: 'Europe/Amsterdam',
|
||||
components: ['homeassistant'],
|
||||
version: '2024.1.0'
|
||||
};
|
||||
expect(validate(minimalConfig)).toBe(true);
|
||||
const result = validateConfig(minimalConfig);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject config with missing required fields', () => {
|
||||
test('should reject config with missing required fields', () => {
|
||||
const invalidConfig = {
|
||||
latitude: 52.3731,
|
||||
longitude: 4.8922
|
||||
// missing other required fields
|
||||
location_name: 'Home'
|
||||
};
|
||||
expect(validate(invalidConfig)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
const result = validateConfig(invalidConfig);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
|
||||
it('should reject config with invalid types', () => {
|
||||
test('should reject config with invalid types', () => {
|
||||
const invalidConfig = {
|
||||
latitude: '52.3731', // should be number
|
||||
longitude: 4.8922,
|
||||
elevation: 0,
|
||||
unit_system: {
|
||||
length: 'km',
|
||||
mass: 'kg',
|
||||
temperature: '°C',
|
||||
volume: 'L'
|
||||
},
|
||||
location_name: 'Home',
|
||||
location_name: 123,
|
||||
time_zone: 'Europe/Amsterdam',
|
||||
components: ['homeassistant'],
|
||||
components: 'not_an_array',
|
||||
version: '2024.1.0'
|
||||
};
|
||||
expect(validate(invalidConfig)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Automation Schema', () => {
|
||||
const validate = ajv.compile(automationSchema);
|
||||
|
||||
it('should validate a basic automation', () => {
|
||||
const basicAutomation = {
|
||||
alias: 'Turn on lights at sunset',
|
||||
description: 'Automatically turn on lights when the sun sets',
|
||||
trigger: [{
|
||||
platform: 'sun',
|
||||
event: 'sunset',
|
||||
offset: '+00:30:00'
|
||||
}],
|
||||
action: [{
|
||||
service: 'light.turn_on',
|
||||
target: {
|
||||
entity_id: ['light.living_room', 'light.kitchen']
|
||||
},
|
||||
data: {
|
||||
brightness_pct: 70
|
||||
}
|
||||
}]
|
||||
};
|
||||
expect(validate(basicAutomation)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate automation with conditions', () => {
|
||||
const automationWithConditions = {
|
||||
alias: 'Conditional Light Control',
|
||||
mode: 'single',
|
||||
trigger: [{
|
||||
platform: 'state',
|
||||
entity_id: 'binary_sensor.motion',
|
||||
to: 'on'
|
||||
}],
|
||||
condition: [{
|
||||
condition: 'and',
|
||||
conditions: [
|
||||
{
|
||||
condition: 'time',
|
||||
after: '22:00:00',
|
||||
before: '06:00:00'
|
||||
},
|
||||
{
|
||||
condition: 'state',
|
||||
entity_id: 'input_boolean.guest_mode',
|
||||
state: 'off'
|
||||
}
|
||||
]
|
||||
}],
|
||||
action: [{
|
||||
service: 'light.turn_on',
|
||||
target: {
|
||||
entity_id: 'light.hallway'
|
||||
}
|
||||
}]
|
||||
};
|
||||
expect(validate(automationWithConditions)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate automation with multiple triggers and actions', () => {
|
||||
const complexAutomation = {
|
||||
alias: 'Complex Automation',
|
||||
mode: 'parallel',
|
||||
trigger: [
|
||||
{
|
||||
platform: 'state',
|
||||
entity_id: 'binary_sensor.door',
|
||||
to: 'on'
|
||||
},
|
||||
{
|
||||
platform: 'state',
|
||||
entity_id: 'binary_sensor.window',
|
||||
to: 'on'
|
||||
}
|
||||
],
|
||||
condition: [{
|
||||
condition: 'state',
|
||||
entity_id: 'alarm_control_panel.home',
|
||||
state: 'armed_away'
|
||||
}],
|
||||
action: [
|
||||
{
|
||||
service: 'notify.mobile_app',
|
||||
data: {
|
||||
message: 'Security alert: Movement detected!'
|
||||
}
|
||||
},
|
||||
{
|
||||
service: 'light.turn_on',
|
||||
target: {
|
||||
entity_id: 'light.all_lights'
|
||||
}
|
||||
},
|
||||
{
|
||||
service: 'camera.snapshot',
|
||||
target: {
|
||||
entity_id: 'camera.front_door'
|
||||
}
|
||||
}
|
||||
]
|
||||
};
|
||||
expect(validate(complexAutomation)).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject automation without required fields', () => {
|
||||
const invalidAutomation = {
|
||||
description: 'Missing required fields'
|
||||
// missing alias, trigger, and action
|
||||
};
|
||||
expect(validate(invalidAutomation)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
});
|
||||
|
||||
it('should validate all automation modes', () => {
|
||||
const modes = ['single', 'parallel', 'queued', 'restart'];
|
||||
modes.forEach(mode => {
|
||||
const automation = {
|
||||
alias: `Test ${mode} mode`,
|
||||
mode,
|
||||
trigger: [{
|
||||
platform: 'state',
|
||||
entity_id: 'input_boolean.test',
|
||||
to: 'on'
|
||||
}],
|
||||
action: [{
|
||||
service: 'light.turn_on',
|
||||
target: {
|
||||
entity_id: 'light.test'
|
||||
}
|
||||
}]
|
||||
};
|
||||
expect(validate(automation)).toBe(true);
|
||||
});
|
||||
const result = validateConfig(invalidConfig);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Device Control Schema', () => {
|
||||
const validate = ajv.compile(deviceControlSchema);
|
||||
|
||||
it('should validate light control command', () => {
|
||||
const lightCommand = {
|
||||
test('should validate light control command', () => {
|
||||
const command = {
|
||||
domain: 'light',
|
||||
command: 'turn_on',
|
||||
entity_id: 'light.living_room',
|
||||
parameters: {
|
||||
brightness: 255,
|
||||
color_temp: 400,
|
||||
transition: 2
|
||||
brightness_pct: 100
|
||||
}
|
||||
};
|
||||
expect(validate(lightCommand)).toBe(true);
|
||||
const result = validateDeviceControl(command);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate climate control command', () => {
|
||||
const climateCommand = {
|
||||
domain: 'climate',
|
||||
command: 'set_temperature',
|
||||
entity_id: 'climate.living_room',
|
||||
parameters: {
|
||||
temperature: 22.5,
|
||||
hvac_mode: 'heat',
|
||||
target_temp_high: 24,
|
||||
target_temp_low: 20
|
||||
}
|
||||
};
|
||||
expect(validate(climateCommand)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate cover control command', () => {
|
||||
const coverCommand = {
|
||||
domain: 'cover',
|
||||
command: 'set_position',
|
||||
entity_id: 'cover.garage_door',
|
||||
parameters: {
|
||||
position: 50,
|
||||
tilt_position: 45
|
||||
}
|
||||
};
|
||||
expect(validate(coverCommand)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate fan control command', () => {
|
||||
const fanCommand = {
|
||||
domain: 'fan',
|
||||
command: 'set_speed',
|
||||
entity_id: 'fan.bedroom',
|
||||
parameters: {
|
||||
speed: 'medium',
|
||||
oscillating: true,
|
||||
direction: 'forward'
|
||||
}
|
||||
};
|
||||
expect(validate(fanCommand)).toBe(true);
|
||||
});
|
||||
|
||||
it('should reject command with invalid domain', () => {
|
||||
const invalidCommand = {
|
||||
domain: 'invalid_domain',
|
||||
command: 'turn_on',
|
||||
entity_id: 'light.living_room'
|
||||
};
|
||||
expect(validate(invalidCommand)).toBe(false);
|
||||
expect(validate.errors).toBeDefined();
|
||||
});
|
||||
|
||||
it('should reject command with mismatched domain and entity_id', () => {
|
||||
test('should reject command with mismatched domain and entity_id', () => {
|
||||
const mismatchedCommand = {
|
||||
domain: 'light',
|
||||
command: 'turn_on',
|
||||
entity_id: 'switch.living_room' // mismatched domain
|
||||
};
|
||||
expect(validate(mismatchedCommand)).toBe(false);
|
||||
const result = validateDeviceControl(mismatchedCommand);
|
||||
expect(result.success).toBe(false);
|
||||
});
|
||||
|
||||
it('should validate command with array of entity_ids', () => {
|
||||
const multiEntityCommand = {
|
||||
test('should validate command with array of entity_ids', () => {
|
||||
const command = {
|
||||
domain: 'light',
|
||||
command: 'turn_on',
|
||||
entity_id: ['light.living_room', 'light.kitchen'],
|
||||
parameters: {
|
||||
brightness: 255
|
||||
}
|
||||
entity_id: ['light.living_room', 'light.kitchen']
|
||||
};
|
||||
expect(validate(multiEntityCommand)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate scene activation command', () => {
|
||||
const sceneCommand = {
|
||||
domain: 'scene',
|
||||
command: 'turn_on',
|
||||
entity_id: 'scene.movie_night',
|
||||
parameters: {
|
||||
transition: 2
|
||||
}
|
||||
};
|
||||
expect(validate(sceneCommand)).toBe(true);
|
||||
});
|
||||
|
||||
it('should validate script execution command', () => {
|
||||
const scriptCommand = {
|
||||
domain: 'script',
|
||||
command: 'turn_on',
|
||||
entity_id: 'script.welcome_home',
|
||||
parameters: {
|
||||
variables: {
|
||||
user: 'John',
|
||||
delay: 5
|
||||
}
|
||||
}
|
||||
};
|
||||
expect(validate(scriptCommand)).toBe(true);
|
||||
const result = validateDeviceControl(command);
|
||||
expect(result.success).toBe(true);
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { TokenManager, validateRequest, sanitizeInput, errorHandler, rateLimiter, securityHeaders } from '../../src/security/index.js';
|
||||
import { mock, describe, it, expect, beforeEach, afterEach } from 'bun:test';
|
||||
import jwt from 'jsonwebtoken';
|
||||
@@ -17,7 +18,7 @@ describe('Security Module', () => {
|
||||
const testToken = 'test-token';
|
||||
const encryptionKey = 'test-encryption-key-that-is-long-enough';
|
||||
|
||||
it('should encrypt and decrypt tokens', () => {
|
||||
test('should encrypt and decrypt tokens', () => {
|
||||
const encrypted = TokenManager.encryptToken(testToken, encryptionKey);
|
||||
expect(encrypted).toContain('aes-256-gcm:');
|
||||
|
||||
@@ -25,20 +26,20 @@ describe('Security Module', () => {
|
||||
expect(decrypted).toBe(testToken);
|
||||
});
|
||||
|
||||
it('should validate tokens correctly', () => {
|
||||
test('should validate tokens correctly', () => {
|
||||
const validToken = jwt.sign({ data: 'test' }, TEST_SECRET, { expiresIn: '1h' });
|
||||
const result = TokenManager.validateToken(validToken);
|
||||
expect(result.valid).toBe(true);
|
||||
expect(result.error).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should handle empty tokens', () => {
|
||||
test('should handle empty tokens', () => {
|
||||
const result = TokenManager.validateToken('');
|
||||
expect(result.valid).toBe(false);
|
||||
expect(result.error).toBe('Invalid token format');
|
||||
});
|
||||
|
||||
it('should handle expired tokens', () => {
|
||||
test('should handle expired tokens', () => {
|
||||
const now = Math.floor(Date.now() / 1000);
|
||||
const payload = {
|
||||
data: 'test',
|
||||
@@ -51,13 +52,13 @@ describe('Security Module', () => {
|
||||
expect(result.error).toBe('Token has expired');
|
||||
});
|
||||
|
||||
it('should handle invalid token format', () => {
|
||||
test('should handle invalid token format', () => {
|
||||
const result = TokenManager.validateToken('invalid-token');
|
||||
expect(result.valid).toBe(false);
|
||||
expect(result.error).toBe('Invalid token format');
|
||||
});
|
||||
|
||||
it('should handle missing JWT secret', () => {
|
||||
test('should handle missing JWT secret', () => {
|
||||
delete process.env.JWT_SECRET;
|
||||
const payload = { data: 'test' };
|
||||
const token = jwt.sign(payload, 'some-secret');
|
||||
@@ -66,7 +67,7 @@ describe('Security Module', () => {
|
||||
expect(result.error).toBe('JWT secret not configured');
|
||||
});
|
||||
|
||||
it('should handle rate limiting for failed attempts', () => {
|
||||
test('should handle rate limiting for failed attempts', () => {
|
||||
const invalidToken = 'x'.repeat(64);
|
||||
const testIp = '127.0.0.1';
|
||||
|
||||
@@ -111,7 +112,7 @@ describe('Security Module', () => {
|
||||
mockNext = mock(() => { });
|
||||
});
|
||||
|
||||
it('should pass valid requests', () => {
|
||||
test('should pass valid requests', () => {
|
||||
if (mockRequest.headers) {
|
||||
mockRequest.headers.authorization = 'Bearer valid-token';
|
||||
}
|
||||
@@ -123,7 +124,7 @@ describe('Security Module', () => {
|
||||
expect(mockNext).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should reject invalid content type', () => {
|
||||
test('should reject invalid content type', () => {
|
||||
if (mockRequest.headers) {
|
||||
mockRequest.headers['content-type'] = 'text/plain';
|
||||
}
|
||||
@@ -139,7 +140,7 @@ describe('Security Module', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should reject missing token', () => {
|
||||
test('should reject missing token', () => {
|
||||
if (mockRequest.headers) {
|
||||
delete mockRequest.headers.authorization;
|
||||
}
|
||||
@@ -155,7 +156,7 @@ describe('Security Module', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should reject invalid request body', () => {
|
||||
test('should reject invalid request body', () => {
|
||||
mockRequest.body = null;
|
||||
|
||||
validateRequest(mockRequest, mockResponse, mockNext);
|
||||
@@ -197,7 +198,7 @@ describe('Security Module', () => {
|
||||
mockNext = mock(() => { });
|
||||
});
|
||||
|
||||
it('should sanitize HTML tags from request body', () => {
|
||||
test('should sanitize HTML tags from request body', () => {
|
||||
sanitizeInput(mockRequest, mockResponse, mockNext);
|
||||
|
||||
expect(mockRequest.body).toEqual({
|
||||
@@ -209,7 +210,7 @@ describe('Security Module', () => {
|
||||
expect(mockNext).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should handle non-object body', () => {
|
||||
test('should handle non-object body', () => {
|
||||
mockRequest.body = 'string body';
|
||||
sanitizeInput(mockRequest, mockResponse, mockNext);
|
||||
expect(mockNext).toHaveBeenCalled();
|
||||
@@ -235,7 +236,7 @@ describe('Security Module', () => {
|
||||
mockNext = mock(() => { });
|
||||
});
|
||||
|
||||
it('should handle errors in production mode', () => {
|
||||
test('should handle errors in production mode', () => {
|
||||
process.env.NODE_ENV = 'production';
|
||||
const error = new Error('Test error');
|
||||
errorHandler(error, mockRequest, mockResponse, mockNext);
|
||||
@@ -248,7 +249,7 @@ describe('Security Module', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should include error message in development mode', () => {
|
||||
test('should include error message in development mode', () => {
|
||||
process.env.NODE_ENV = 'development';
|
||||
const error = new Error('Test error');
|
||||
errorHandler(error, mockRequest, mockResponse, mockNext);
|
||||
@@ -265,7 +266,7 @@ describe('Security Module', () => {
|
||||
});
|
||||
|
||||
describe('Rate Limiter', () => {
|
||||
it('should limit requests after threshold', async () => {
|
||||
test('should limit requests after threshold', async () => {
|
||||
const mockContext = {
|
||||
request: new Request('http://localhost', {
|
||||
headers: new Headers({
|
||||
@@ -292,7 +293,7 @@ describe('Security Module', () => {
|
||||
});
|
||||
|
||||
describe('Security Headers', () => {
|
||||
it('should set security headers', async () => {
|
||||
test('should set security headers', async () => {
|
||||
const mockHeaders = new Headers();
|
||||
const mockContext = {
|
||||
request: new Request('http://localhost', {
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, it, expect } from 'bun:test';
|
||||
import {
|
||||
checkRateLimit,
|
||||
@@ -9,31 +10,31 @@ import {
|
||||
|
||||
describe('Security Middleware Utilities', () => {
|
||||
describe('Rate Limiter', () => {
|
||||
it('should allow requests under threshold', () => {
|
||||
test('should allow requests under threshold', () => {
|
||||
const ip = '127.0.0.1';
|
||||
expect(() => checkRateLimit(ip, 10)).not.toThrow();
|
||||
expect(() => checkRateLimtest(ip, 10)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should throw when requests exceed threshold', () => {
|
||||
test('should throw when requests exceed threshold', () => {
|
||||
const ip = '127.0.0.2';
|
||||
|
||||
// Simulate multiple requests
|
||||
for (let i = 0; i < 11; i++) {
|
||||
if (i < 10) {
|
||||
expect(() => checkRateLimit(ip, 10)).not.toThrow();
|
||||
expect(() => checkRateLimtest(ip, 10)).not.toThrow();
|
||||
} else {
|
||||
expect(() => checkRateLimit(ip, 10)).toThrow('Too many requests from this IP, please try again later');
|
||||
expect(() => checkRateLimtest(ip, 10)).toThrow('Too many requests from this IP, please try again later');
|
||||
}
|
||||
}
|
||||
});
|
||||
|
||||
it('should reset rate limit after window expires', async () => {
|
||||
test('should reset rate limit after window expires', async () => {
|
||||
const ip = '127.0.0.3';
|
||||
|
||||
// Simulate multiple requests
|
||||
for (let i = 0; i < 11; i++) {
|
||||
if (i < 10) {
|
||||
expect(() => checkRateLimit(ip, 10, 50)).not.toThrow();
|
||||
expect(() => checkRateLimtest(ip, 10, 50)).not.toThrow();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -41,12 +42,12 @@ describe('Security Middleware Utilities', () => {
|
||||
await new Promise(resolve => setTimeout(resolve, 100));
|
||||
|
||||
// Should be able to make requests again
|
||||
expect(() => checkRateLimit(ip, 10, 50)).not.toThrow();
|
||||
expect(() => checkRateLimtest(ip, 10, 50)).not.toThrow();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Request Validation', () => {
|
||||
it('should validate content type', () => {
|
||||
test('should validate content type', () => {
|
||||
const mockRequest = new Request('http://localhost', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
@@ -57,7 +58,7 @@ describe('Security Middleware Utilities', () => {
|
||||
expect(() => validateRequestHeaders(mockRequest)).not.toThrow();
|
||||
});
|
||||
|
||||
it('should reject invalid content type', () => {
|
||||
test('should reject invalid content type', () => {
|
||||
const mockRequest = new Request('http://localhost', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
@@ -68,7 +69,7 @@ describe('Security Middleware Utilities', () => {
|
||||
expect(() => validateRequestHeaders(mockRequest)).toThrow('Content-Type must be application/json');
|
||||
});
|
||||
|
||||
it('should reject large request bodies', () => {
|
||||
test('should reject large request bodies', () => {
|
||||
const mockRequest = new Request('http://localhost', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
@@ -82,13 +83,13 @@ describe('Security Middleware Utilities', () => {
|
||||
});
|
||||
|
||||
describe('Input Sanitization', () => {
|
||||
it('should sanitize HTML tags', () => {
|
||||
test('should sanitize HTML tags', () => {
|
||||
const input = '<script>alert("xss")</script>Hello';
|
||||
const sanitized = sanitizeValue(input);
|
||||
expect(sanitized).toBe('<script>alert("xss")</script>Hello');
|
||||
});
|
||||
|
||||
it('should sanitize nested objects', () => {
|
||||
test('should sanitize nested objects', () => {
|
||||
const input = {
|
||||
text: '<script>alert("xss")</script>Hello',
|
||||
nested: {
|
||||
@@ -104,7 +105,7 @@ describe('Security Middleware Utilities', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should preserve non-string values', () => {
|
||||
test('should preserve non-string values', () => {
|
||||
const input = {
|
||||
number: 123,
|
||||
boolean: true,
|
||||
@@ -116,7 +117,7 @@ describe('Security Middleware Utilities', () => {
|
||||
});
|
||||
|
||||
describe('Security Headers', () => {
|
||||
it('should apply security headers', () => {
|
||||
test('should apply security headers', () => {
|
||||
const mockRequest = new Request('http://localhost');
|
||||
const headers = applySecurityHeaders(mockRequest);
|
||||
|
||||
@@ -129,7 +130,7 @@ describe('Security Middleware Utilities', () => {
|
||||
});
|
||||
|
||||
describe('Error Handling', () => {
|
||||
it('should handle errors in production mode', () => {
|
||||
test('should handle errors in production mode', () => {
|
||||
const error = new Error('Test error');
|
||||
const result = handleError(error, 'production');
|
||||
|
||||
@@ -140,7 +141,7 @@ describe('Security Middleware Utilities', () => {
|
||||
});
|
||||
});
|
||||
|
||||
it('should include error details in development mode', () => {
|
||||
test('should include error details in development mode', () => {
|
||||
const error = new Error('Test error');
|
||||
const result = handleError(error, 'development');
|
||||
|
||||
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { TokenManager } from '../../src/security/index.js';
|
||||
import jwt from 'jsonwebtoken';
|
||||
|
||||
@@ -16,36 +17,36 @@ describe('TokenManager', () => {
|
||||
const validToken = 'eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzdWIiOiIxMjM0NTY3ODkwIiwibmFtZSI6IkpvaG4gRG9lIiwiZXhwIjoxNjE2MjM5MDIyfQ.SflKxwRJSMeKKF2QT4fwpMeJf36POk6yJV_adQssw5c';
|
||||
|
||||
describe('Token Encryption/Decryption', () => {
|
||||
it('should encrypt and decrypt tokens successfully', () => {
|
||||
test('should encrypt and decrypt tokens successfully', () => {
|
||||
const encrypted = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const decrypted = TokenManager.decryptToken(encrypted, encryptionKey);
|
||||
expect(decrypted).toBe(validToken);
|
||||
});
|
||||
|
||||
it('should generate different encrypted values for same token', () => {
|
||||
test('should generate different encrypted values for same token', () => {
|
||||
const encrypted1 = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const encrypted2 = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
expect(encrypted1).not.toBe(encrypted2);
|
||||
});
|
||||
|
||||
it('should handle empty tokens', () => {
|
||||
test('should handle empty tokens', () => {
|
||||
expect(() => TokenManager.encryptToken('', encryptionKey)).toThrow('Invalid token');
|
||||
expect(() => TokenManager.decryptToken('', encryptionKey)).toThrow('Invalid encrypted token');
|
||||
});
|
||||
|
||||
it('should handle empty encryption keys', () => {
|
||||
test('should handle empty encryption keys', () => {
|
||||
expect(() => TokenManager.encryptToken(validToken, '')).toThrow('Invalid encryption key');
|
||||
expect(() => TokenManager.decryptToken(validToken, '')).toThrow('Invalid encryption key');
|
||||
});
|
||||
|
||||
it('should fail decryption with wrong key', () => {
|
||||
test('should fail decryption with wrong key', () => {
|
||||
const encrypted = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
expect(() => TokenManager.decryptToken(encrypted, 'wrong-key-32-chars-long!!!!!!!!')).toThrow();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Token Validation', () => {
|
||||
it('should validate correct tokens', () => {
|
||||
test('should validate correct tokens', () => {
|
||||
const payload = { sub: '123', name: 'Test User', iat: Math.floor(Date.now() / 1000), exp: Math.floor(Date.now() / 1000) + 3600 };
|
||||
const token = jwt.sign(payload, TEST_SECRET);
|
||||
const result = TokenManager.validateToken(token);
|
||||
@@ -53,7 +54,7 @@ describe('TokenManager', () => {
|
||||
expect(result.error).toBeUndefined();
|
||||
});
|
||||
|
||||
it('should reject expired tokens', () => {
|
||||
test('should reject expired tokens', () => {
|
||||
const payload = { sub: '123', name: 'Test User', iat: Math.floor(Date.now() / 1000) - 7200, exp: Math.floor(Date.now() / 1000) - 3600 };
|
||||
const token = jwt.sign(payload, TEST_SECRET);
|
||||
const result = TokenManager.validateToken(token);
|
||||
@@ -61,13 +62,13 @@ describe('TokenManager', () => {
|
||||
expect(result.error).toBe('Token has expired');
|
||||
});
|
||||
|
||||
it('should reject malformed tokens', () => {
|
||||
test('should reject malformed tokens', () => {
|
||||
const result = TokenManager.validateToken('invalid-token');
|
||||
expect(result.valid).toBe(false);
|
||||
expect(result.error).toBe('Token length below minimum requirement');
|
||||
});
|
||||
|
||||
it('should reject tokens with invalid signature', () => {
|
||||
test('should reject tokens with invalid signature', () => {
|
||||
const payload = { sub: '123', name: 'Test User', iat: Math.floor(Date.now() / 1000), exp: Math.floor(Date.now() / 1000) + 3600 };
|
||||
const token = jwt.sign(payload, 'different-secret');
|
||||
const result = TokenManager.validateToken(token);
|
||||
@@ -75,7 +76,7 @@ describe('TokenManager', () => {
|
||||
expect(result.error).toBe('Invalid token signature');
|
||||
});
|
||||
|
||||
it('should handle tokens with missing expiration', () => {
|
||||
test('should handle tokens with missing expiration', () => {
|
||||
const payload = { sub: '123', name: 'Test User' };
|
||||
const token = jwt.sign(payload, TEST_SECRET);
|
||||
const result = TokenManager.validateToken(token);
|
||||
@@ -83,7 +84,7 @@ describe('TokenManager', () => {
|
||||
expect(result.error).toBe('Token missing required claims');
|
||||
});
|
||||
|
||||
it('should handle undefined and null inputs', () => {
|
||||
test('should handle undefined and null inputs', () => {
|
||||
const undefinedResult = TokenManager.validateToken(undefined);
|
||||
expect(undefinedResult.valid).toBe(false);
|
||||
expect(undefinedResult.error).toBe('Invalid token format');
|
||||
@@ -95,26 +96,26 @@ describe('TokenManager', () => {
|
||||
});
|
||||
|
||||
describe('Security Features', () => {
|
||||
it('should use secure encryption algorithm', () => {
|
||||
test('should use secure encryption algorithm', () => {
|
||||
const encrypted = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
expect(encrypted).toContain('aes-256-gcm');
|
||||
});
|
||||
|
||||
it('should prevent token tampering', () => {
|
||||
test('should prevent token tampering', () => {
|
||||
const encrypted = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const tampered = encrypted.slice(0, -5) + 'xxxxx';
|
||||
expect(() => TokenManager.decryptToken(tampered, encryptionKey)).toThrow();
|
||||
});
|
||||
|
||||
it('should use unique IVs for each encryption', () => {
|
||||
test('should use unique IVs for each encryption', () => {
|
||||
const encrypted1 = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const encrypted2 = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const iv1 = encrypted1.split(':')[1];
|
||||
const iv2 = encrypted2.split(':')[1];
|
||||
const iv1 = encrypted1.spltest(':')[1];
|
||||
const iv2 = encrypted2.spltest(':')[1];
|
||||
expect(iv1).not.toBe(iv2);
|
||||
});
|
||||
|
||||
it('should handle large tokens', () => {
|
||||
test('should handle large tokens', () => {
|
||||
const largeToken = 'x'.repeat(10000);
|
||||
const encrypted = TokenManager.encryptToken(largeToken, encryptionKey);
|
||||
const decrypted = TokenManager.decryptToken(encrypted, encryptionKey);
|
||||
@@ -123,19 +124,19 @@ describe('TokenManager', () => {
|
||||
});
|
||||
|
||||
describe('Error Handling', () => {
|
||||
it('should throw descriptive errors for invalid inputs', () => {
|
||||
test('should throw descriptive errors for invalid inputs', () => {
|
||||
expect(() => TokenManager.encryptToken(null as any, encryptionKey)).toThrow('Invalid token');
|
||||
expect(() => TokenManager.encryptToken(validToken, null as any)).toThrow('Invalid encryption key');
|
||||
expect(() => TokenManager.decryptToken('invalid-base64', encryptionKey)).toThrow('Invalid encrypted token');
|
||||
});
|
||||
|
||||
it('should handle corrupted encrypted data', () => {
|
||||
test('should handle corrupted encrypted data', () => {
|
||||
const encrypted = TokenManager.encryptToken(validToken, encryptionKey);
|
||||
const corrupted = encrypted.replace(/[a-zA-Z]/g, 'x');
|
||||
expect(() => TokenManager.decryptToken(corrupted, encryptionKey)).toThrow();
|
||||
});
|
||||
|
||||
it('should handle invalid base64 input', () => {
|
||||
test('should handle invalid base64 input', () => {
|
||||
expect(() => TokenManager.decryptToken('not-base64!@#$%^', encryptionKey)).toThrow();
|
||||
});
|
||||
});
|
||||
|
||||
@@ -1,114 +1,149 @@
|
||||
import { jest, describe, beforeEach, afterEach, it, expect } from '@jest/globals';
|
||||
import express from 'express';
|
||||
import { LiteMCP } from 'litemcp';
|
||||
import { logger } from '../src/utils/logger.js';
|
||||
import { describe, expect, test, beforeEach, afterEach, mock, spyOn } from "bun:test";
|
||||
import type { Mock } from "bun:test";
|
||||
import type { Elysia } from "elysia";
|
||||
|
||||
// Mock express
|
||||
jest.mock('express', () => {
|
||||
// Create mock instances
|
||||
const mockApp = {
|
||||
use: jest.fn(),
|
||||
listen: jest.fn((port: number, callback: () => void) => {
|
||||
callback();
|
||||
return { close: jest.fn() };
|
||||
use: mock(() => mockApp),
|
||||
get: mock(() => mockApp),
|
||||
post: mock(() => mockApp),
|
||||
listen: mock((port: number, callback?: () => void) => {
|
||||
callback?.();
|
||||
return mockApp;
|
||||
})
|
||||
};
|
||||
return jest.fn(() => mockApp);
|
||||
});
|
||||
|
||||
// Mock LiteMCP
|
||||
jest.mock('litemcp', () => ({
|
||||
LiteMCP: jest.fn(() => ({
|
||||
addTool: jest.fn(),
|
||||
start: jest.fn().mockImplementation(async () => { })
|
||||
}))
|
||||
}));
|
||||
// Create mock constructors
|
||||
const MockElysia = mock(() => mockApp);
|
||||
const mockCors = mock(() => (app: any) => app);
|
||||
const mockSwagger = mock(() => (app: any) => app);
|
||||
const mockSpeechService = {
|
||||
initialize: mock(() => Promise.resolve()),
|
||||
shutdown: mock(() => Promise.resolve())
|
||||
};
|
||||
|
||||
// Mock logger
|
||||
jest.mock('../src/utils/logger.js', () => ({
|
||||
logger: {
|
||||
info: jest.fn(),
|
||||
error: jest.fn(),
|
||||
debug: jest.fn()
|
||||
// Mock the modules
|
||||
const mockModules = {
|
||||
Elysia: MockElysia,
|
||||
cors: mockCors,
|
||||
swagger: mockSwagger,
|
||||
speechService: mockSpeechService,
|
||||
config: mock(() => ({})),
|
||||
resolve: mock((...args: string[]) => args.join('/')),
|
||||
z: { object: mock(() => ({})), enum: mock(() => ({})) }
|
||||
};
|
||||
|
||||
// Mock module resolution
|
||||
const mockResolver = {
|
||||
resolve(specifier: string) {
|
||||
const mocks: Record<string, any> = {
|
||||
'elysia': { Elysia: mockModules.Elysia },
|
||||
'@elysiajs/cors': { cors: mockModules.cors },
|
||||
'@elysiajs/swagger': { swagger: mockModules.swagger },
|
||||
'../speech/index.js': { speechService: mockModules.speechService },
|
||||
'dotenv': { config: mockModules.config },
|
||||
'path': { resolve: mockModules.resolve },
|
||||
'zod': { z: mockModules.z }
|
||||
};
|
||||
return mocks[specifier] || {};
|
||||
}
|
||||
}));
|
||||
};
|
||||
|
||||
describe('Server Initialization', () => {
|
||||
let originalEnv: NodeJS.ProcessEnv;
|
||||
let mockApp: ReturnType<typeof express>;
|
||||
let consoleLog: Mock<typeof console.log>;
|
||||
let consoleError: Mock<typeof console.error>;
|
||||
let originalResolve: any;
|
||||
|
||||
beforeEach(() => {
|
||||
// Store original environment
|
||||
originalEnv = { ...process.env };
|
||||
|
||||
// Reset all mocks
|
||||
jest.clearAllMocks();
|
||||
// Mock console methods
|
||||
consoleLog = mock(() => { });
|
||||
consoleError = mock(() => { });
|
||||
console.log = consoleLog;
|
||||
console.error = consoleError;
|
||||
|
||||
// Get the mock express app
|
||||
mockApp = express();
|
||||
// Reset all mocks
|
||||
for (const key in mockModules) {
|
||||
const module = mockModules[key as keyof typeof mockModules];
|
||||
if (typeof module === 'object' && module !== null) {
|
||||
Object.values(module).forEach(value => {
|
||||
if (typeof value === 'function' && 'mock' in value) {
|
||||
(value as Mock<any>).mockReset();
|
||||
}
|
||||
});
|
||||
} else if (typeof module === 'function' && 'mock' in module) {
|
||||
(module as Mock<any>).mockReset();
|
||||
}
|
||||
}
|
||||
|
||||
// Set default environment variables
|
||||
process.env.NODE_ENV = 'test';
|
||||
process.env.PORT = '4000';
|
||||
|
||||
// Setup module resolution mock
|
||||
originalResolve = (globalThis as any).Bun?.resolveSync;
|
||||
(globalThis as any).Bun = {
|
||||
...(globalThis as any).Bun,
|
||||
resolveSync: (specifier: string) => mockResolver.resolve(specifier)
|
||||
};
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore original environment
|
||||
process.env = originalEnv;
|
||||
|
||||
// Clear module cache to ensure fresh imports
|
||||
jest.resetModules();
|
||||
// Restore module resolution
|
||||
if (originalResolve) {
|
||||
(globalThis as any).Bun.resolveSync = originalResolve;
|
||||
}
|
||||
});
|
||||
|
||||
it('should start Express server when not in Claude mode', async () => {
|
||||
// Set OpenAI mode
|
||||
process.env.PROCESSOR_TYPE = 'openai';
|
||||
test('should initialize server with middleware', async () => {
|
||||
// Import and initialize server
|
||||
const mod = await import('../src/index');
|
||||
|
||||
// Import the main module
|
||||
await import('../src/index.js');
|
||||
// Verify server initialization
|
||||
expect(MockElysia.mock.calls.length).toBe(1);
|
||||
expect(mockCors.mock.calls.length).toBe(1);
|
||||
expect(mockSwagger.mock.calls.length).toBe(1);
|
||||
|
||||
// Verify Express server was initialized
|
||||
expect(express).toHaveBeenCalled();
|
||||
expect(mockApp.use).toHaveBeenCalled();
|
||||
expect(mockApp.listen).toHaveBeenCalled();
|
||||
expect(logger.info).toHaveBeenCalledWith(expect.stringContaining('Server is running on port'));
|
||||
// Verify console output
|
||||
const logCalls = consoleLog.mock.calls;
|
||||
expect(logCalls.some(call =>
|
||||
typeof call.args[0] === 'string' &&
|
||||
call.args[0].includes('Server is running on port')
|
||||
)).toBe(true);
|
||||
});
|
||||
|
||||
it('should not start Express server in Claude mode', async () => {
|
||||
// Set Claude mode
|
||||
process.env.PROCESSOR_TYPE = 'claude';
|
||||
test('should initialize speech service when enabled', async () => {
|
||||
// Enable speech service
|
||||
process.env.SPEECH_ENABLED = 'true';
|
||||
|
||||
// Import the main module
|
||||
await import('../src/index.js');
|
||||
// Import and initialize server
|
||||
const mod = await import('../src/index');
|
||||
|
||||
// Verify Express server was not initialized
|
||||
expect(express).not.toHaveBeenCalled();
|
||||
expect(mockApp.use).not.toHaveBeenCalled();
|
||||
expect(mockApp.listen).not.toHaveBeenCalled();
|
||||
expect(logger.info).toHaveBeenCalledWith('Running in Claude mode - Express server disabled');
|
||||
// Verify speech service initialization
|
||||
expect(mockSpeechService.initialize.mock.calls.length).toBe(1);
|
||||
});
|
||||
|
||||
it('should initialize LiteMCP in both modes', async () => {
|
||||
// Test OpenAI mode
|
||||
process.env.PROCESSOR_TYPE = 'openai';
|
||||
await import('../src/index.js');
|
||||
expect(LiteMCP).toHaveBeenCalledWith('home-assistant', expect.any(String));
|
||||
test('should handle server shutdown gracefully', async () => {
|
||||
// Enable speech service for shutdown test
|
||||
process.env.SPEECH_ENABLED = 'true';
|
||||
|
||||
// Reset modules
|
||||
jest.resetModules();
|
||||
// Import and initialize server
|
||||
const mod = await import('../src/index');
|
||||
|
||||
// Test Claude mode
|
||||
process.env.PROCESSOR_TYPE = 'claude';
|
||||
await import('../src/index.js');
|
||||
expect(LiteMCP).toHaveBeenCalledWith('home-assistant', expect.any(String));
|
||||
});
|
||||
// Simulate SIGTERM
|
||||
process.emit('SIGTERM');
|
||||
|
||||
it('should handle missing PROCESSOR_TYPE (default to Express server)', async () => {
|
||||
// Remove PROCESSOR_TYPE
|
||||
delete process.env.PROCESSOR_TYPE;
|
||||
|
||||
// Import the main module
|
||||
await import('../src/index.js');
|
||||
|
||||
// Verify Express server was initialized (default behavior)
|
||||
expect(express).toHaveBeenCalled();
|
||||
expect(mockApp.use).toHaveBeenCalled();
|
||||
expect(mockApp.listen).toHaveBeenCalled();
|
||||
expect(logger.info).toHaveBeenCalledWith(expect.stringContaining('Server is running on port'));
|
||||
// Verify shutdown behavior
|
||||
expect(mockSpeechService.shutdown.mock.calls.length).toBe(1);
|
||||
expect(consoleLog.mock.calls.some(call =>
|
||||
typeof call.args[0] === 'string' &&
|
||||
call.args[0].includes('Shutting down gracefully')
|
||||
)).toBe(true);
|
||||
});
|
||||
});
|
||||
251
__tests__/speech/speechToText.test.ts
Normal file
251
__tests__/speech/speechToText.test.ts
Normal file
@@ -0,0 +1,251 @@
|
||||
import { describe, expect, test, beforeEach, afterEach, mock, spyOn } from "bun:test";
|
||||
import type { Mock } from "bun:test";
|
||||
import { EventEmitter } from "events";
|
||||
import { SpeechToText, TranscriptionError, type TranscriptionOptions } from "../../src/speech/speechToText";
|
||||
import type { SpeechToTextConfig } from "../../src/speech/types";
|
||||
import type { ChildProcess } from "child_process";
|
||||
|
||||
interface MockProcess extends EventEmitter {
|
||||
stdout: EventEmitter;
|
||||
stderr: EventEmitter;
|
||||
kill: Mock<() => void>;
|
||||
}
|
||||
|
||||
type SpawnFn = {
|
||||
(cmds: string[], options?: Record<string, unknown>): ChildProcess;
|
||||
};
|
||||
|
||||
describe('SpeechToText', () => {
|
||||
let spawnMock: Mock<SpawnFn>;
|
||||
let mockProcess: MockProcess;
|
||||
let speechToText: SpeechToText;
|
||||
|
||||
beforeEach(() => {
|
||||
// Create mock process
|
||||
mockProcess = new EventEmitter() as MockProcess;
|
||||
mockProcess.stdout = new EventEmitter();
|
||||
mockProcess.stderr = new EventEmitter();
|
||||
mockProcess.kill = mock(() => { });
|
||||
|
||||
// Create spawn mock
|
||||
spawnMock = mock((cmds: string[], options?: Record<string, unknown>) => mockProcess as unknown as ChildProcess);
|
||||
(globalThis as any).Bun = { spawn: spawnMock };
|
||||
|
||||
// Initialize SpeechToText
|
||||
const config: SpeechToTextConfig = {
|
||||
modelPath: '/test/model',
|
||||
modelType: 'base.en',
|
||||
containerName: 'test-container'
|
||||
};
|
||||
speechToText = new SpeechToText(config);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Cleanup
|
||||
mockProcess.removeAllListeners();
|
||||
mockProcess.stdout.removeAllListeners();
|
||||
mockProcess.stderr.removeAllListeners();
|
||||
});
|
||||
|
||||
describe('Initialization', () => {
|
||||
test('should create instance with default config', () => {
|
||||
const config: SpeechToTextConfig = {
|
||||
modelPath: '/test/model',
|
||||
modelType: 'base.en'
|
||||
};
|
||||
const instance = new SpeechToText(config);
|
||||
expect(instance).toBeDefined();
|
||||
});
|
||||
|
||||
test('should initialize successfully', async () => {
|
||||
const result = await speechToText.initialize();
|
||||
expect(result).toBeUndefined();
|
||||
});
|
||||
|
||||
test('should not initialize twice', async () => {
|
||||
await speechToText.initialize();
|
||||
const result = await speechToText.initialize();
|
||||
expect(result).toBeUndefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Health Check', () => {
|
||||
test('should return true when Docker container is running', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from('Up 2 hours'));
|
||||
}, 0);
|
||||
|
||||
const result = await speechToText.checkHealth();
|
||||
expect(result).toBe(true);
|
||||
});
|
||||
|
||||
test('should return false when Docker container is not running', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from('No containers found'));
|
||||
}, 0);
|
||||
|
||||
const result = await speechToText.checkHealth();
|
||||
expect(result).toBe(false);
|
||||
});
|
||||
|
||||
test('should handle Docker command errors', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stderr.emit('data', Buffer.from('Docker error'));
|
||||
}, 0);
|
||||
|
||||
const result = await speechToText.checkHealth();
|
||||
expect(result).toBe(false);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Wake Word Detection', () => {
|
||||
test('should detect wake word and emit event', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from('Wake word detected'));
|
||||
}, 0);
|
||||
|
||||
const wakeWordPromise = new Promise<void>((resolve) => {
|
||||
speechToText.on('wake_word', () => {
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
|
||||
speechToText.startWakeWordDetection();
|
||||
await wakeWordPromise;
|
||||
});
|
||||
|
||||
test('should handle non-wake-word files', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from('Processing audio'));
|
||||
}, 0);
|
||||
|
||||
const wakeWordPromise = new Promise<void>((resolve, reject) => {
|
||||
const timeout = setTimeout(() => {
|
||||
resolve();
|
||||
}, 100);
|
||||
|
||||
speechToText.on('wake_word', () => {
|
||||
clearTimeout(timeout);
|
||||
reject(new Error('Wake word should not be detected'));
|
||||
});
|
||||
});
|
||||
|
||||
speechToText.startWakeWordDetection();
|
||||
await wakeWordPromise;
|
||||
});
|
||||
});
|
||||
|
||||
describe('Audio Transcription', () => {
|
||||
const mockTranscriptionResult = {
|
||||
text: 'Test transcription',
|
||||
segments: [{
|
||||
text: 'Test transcription',
|
||||
start: 0,
|
||||
end: 1,
|
||||
confidence: 0.95
|
||||
}]
|
||||
};
|
||||
|
||||
test('should transcribe audio successfully', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from(JSON.stringify(mockTranscriptionResult)));
|
||||
}, 0);
|
||||
|
||||
const result = await speechToText.transcribeAudio('/test/audio.wav');
|
||||
expect(result).toEqual(mockTranscriptionResult);
|
||||
});
|
||||
|
||||
test('should handle transcription errors', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stderr.emit('data', Buffer.from('Transcription failed'));
|
||||
}, 0);
|
||||
|
||||
await expect(speechToText.transcribeAudio('/test/audio.wav')).rejects.toThrow(TranscriptionError);
|
||||
});
|
||||
|
||||
test('should handle invalid JSON output', async () => {
|
||||
// Setup mock process
|
||||
setTimeout(() => {
|
||||
mockProcess.stdout.emit('data', Buffer.from('Invalid JSON'));
|
||||
}, 0);
|
||||
|
||||
await expect(speechToText.transcribeAudio('/test/audio.wav')).rejects.toThrow(TranscriptionError);
|
||||
});
|
||||
|
||||
test('should pass correct transcription options', async () => {
|
||||
const options: TranscriptionOptions = {
|
||||
model: 'base.en',
|
||||
language: 'en',
|
||||
temperature: 0,
|
||||
beamSize: 5,
|
||||
patience: 1,
|
||||
device: 'cpu'
|
||||
};
|
||||
|
||||
await speechToText.transcribeAudio('/test/audio.wav', options);
|
||||
|
||||
const spawnArgs = spawnMock.mock.calls[0]?.args[1] || [];
|
||||
expect(spawnArgs).toContain('--model');
|
||||
expect(spawnArgs).toContain(options.model);
|
||||
expect(spawnArgs).toContain('--language');
|
||||
expect(spawnArgs).toContain(options.language);
|
||||
expect(spawnArgs).toContain('--temperature');
|
||||
expect(spawnArgs).toContain(options.temperature?.toString());
|
||||
expect(spawnArgs).toContain('--beam-size');
|
||||
expect(spawnArgs).toContain(options.beamSize?.toString());
|
||||
expect(spawnArgs).toContain('--patience');
|
||||
expect(spawnArgs).toContain(options.patience?.toString());
|
||||
expect(spawnArgs).toContain('--device');
|
||||
expect(spawnArgs).toContain(options.device);
|
||||
});
|
||||
});
|
||||
|
||||
describe('Event Handling', () => {
|
||||
test('should emit progress events', async () => {
|
||||
const progressPromise = new Promise<void>((resolve) => {
|
||||
speechToText.on('progress', (progress) => {
|
||||
expect(progress).toEqual({ type: 'stdout', data: 'Processing' });
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
|
||||
const transcribePromise = speechToText.transcribeAudio('/test/audio.wav');
|
||||
mockProcess.stdout.emit('data', Buffer.from('Processing'));
|
||||
await Promise.all([transcribePromise.catch(() => { }), progressPromise]);
|
||||
});
|
||||
|
||||
test('should emit error events', async () => {
|
||||
const errorPromise = new Promise<void>((resolve) => {
|
||||
speechToText.on('error', (error) => {
|
||||
expect(error instanceof Error).toBe(true);
|
||||
expect(error.message).toBe('Test error');
|
||||
resolve();
|
||||
});
|
||||
});
|
||||
|
||||
speechToText.emit('error', new Error('Test error'));
|
||||
await errorPromise;
|
||||
});
|
||||
});
|
||||
|
||||
describe('Cleanup', () => {
|
||||
test('should stop wake word detection', () => {
|
||||
speechToText.startWakeWordDetection();
|
||||
speechToText.stopWakeWordDetection();
|
||||
expect(mockProcess.kill.mock.calls.length).toBe(1);
|
||||
});
|
||||
|
||||
test('should clean up resources on shutdown', async () => {
|
||||
await speechToText.initialize();
|
||||
await speechToText.shutdown();
|
||||
expect(mockProcess.kill.mock.calls.length).toBe(1);
|
||||
});
|
||||
});
|
||||
});
|
||||
203
__tests__/tools/automation-config.test.ts
Normal file
203
__tests__/tools/automation-config.test.ts
Normal file
@@ -0,0 +1,203 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import {
|
||||
type MockLiteMCPInstance,
|
||||
type Tool,
|
||||
type TestResponse,
|
||||
TEST_CONFIG,
|
||||
createMockLiteMCPInstance,
|
||||
setupTestEnvironment,
|
||||
cleanupMocks,
|
||||
createMockResponse,
|
||||
getMockCallArgs
|
||||
} from '../utils/test-utils';
|
||||
|
||||
describe('Automation Configuration Tools', () => {
|
||||
let liteMcpInstance: MockLiteMCPInstance;
|
||||
let addToolCalls: Tool[];
|
||||
let mocks: ReturnType<typeof setupTestEnvironment>;
|
||||
|
||||
const mockAutomationConfig = {
|
||||
alias: 'Test Automation',
|
||||
description: 'Test automation description',
|
||||
mode: 'single',
|
||||
trigger: [
|
||||
{
|
||||
platform: 'state',
|
||||
entity_id: 'binary_sensor.motion',
|
||||
to: 'on'
|
||||
}
|
||||
],
|
||||
action: [
|
||||
{
|
||||
service: 'light.turn_on',
|
||||
target: {
|
||||
entity_id: 'light.living_room'
|
||||
}
|
||||
}
|
||||
]
|
||||
};
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup test environment
|
||||
mocks = setupTestEnvironment();
|
||||
liteMcpInstance = createMockLiteMCPInstance();
|
||||
|
||||
// Import the module which will execute the main function
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Get the mock instance and tool calls
|
||||
addToolCalls = liteMcpInstance.addTool.mock.calls.map(call => call.args[0]);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
cleanupMocks({ liteMcpInstance, ...mocks });
|
||||
});
|
||||
|
||||
describe('automation_config tool', () => {
|
||||
test('should successfully create an automation', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({
|
||||
automation_id: 'new_automation_1'
|
||||
})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const automationConfigTool = addToolCalls.find(tool => tool.name === 'automation_config');
|
||||
expect(automationConfigTool).toBeDefined();
|
||||
|
||||
if (!automationConfigTool) {
|
||||
throw new Error('automation_config tool not found');
|
||||
}
|
||||
|
||||
const result = await automationConfigTool.execute({
|
||||
action: 'create',
|
||||
config: mockAutomationConfig
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully created automation');
|
||||
expect(result.automation_id).toBe('new_automation_1');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/config/automation/config`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify(mockAutomationConfig)
|
||||
});
|
||||
});
|
||||
|
||||
test('should successfully duplicate an automation', async () => {
|
||||
// Setup responses for get and create
|
||||
let callCount = 0;
|
||||
mocks.mockFetch = mock(() => {
|
||||
callCount++;
|
||||
return Promise.resolve(
|
||||
callCount === 1
|
||||
? createMockResponse(mockAutomationConfig)
|
||||
: createMockResponse({ automation_id: 'new_automation_2' })
|
||||
);
|
||||
});
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const automationConfigTool = addToolCalls.find(tool => tool.name === 'automation_config');
|
||||
expect(automationConfigTool).toBeDefined();
|
||||
|
||||
if (!automationConfigTool) {
|
||||
throw new Error('automation_config tool not found');
|
||||
}
|
||||
|
||||
const result = await automationConfigTool.execute({
|
||||
action: 'duplicate',
|
||||
automation_id: 'automation.test'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully duplicated automation automation.test');
|
||||
expect(result.new_automation_id).toBe('new_automation_2');
|
||||
|
||||
// Verify both API calls
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const calls = mocks.mockFetch.mock.calls;
|
||||
expect(calls.length).toBe(2);
|
||||
|
||||
// Verify get call
|
||||
const getArgs = getMockCallArgs<FetchArgs>(mocks.mockFetch, 0);
|
||||
expect(getArgs).toBeDefined();
|
||||
if (!getArgs) throw new Error('No get call recorded');
|
||||
|
||||
const [getUrl, getOptions] = getArgs;
|
||||
expect(getUrl).toBe(`${TEST_CONFIG.HASS_HOST}/api/config/automation/config/automation.test`);
|
||||
expect(getOptions).toEqual({
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
|
||||
// Verify create call
|
||||
const createArgs = getMockCallArgs<FetchArgs>(mocks.mockFetch, 1);
|
||||
expect(createArgs).toBeDefined();
|
||||
if (!createArgs) throw new Error('No create call recorded');
|
||||
|
||||
const [createUrl, createOptions] = createArgs;
|
||||
expect(createUrl).toBe(`${TEST_CONFIG.HASS_HOST}/api/config/automation/config`);
|
||||
expect(createOptions).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
...mockAutomationConfig,
|
||||
alias: 'Test Automation (Copy)'
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should require config for create action', async () => {
|
||||
const automationConfigTool = addToolCalls.find(tool => tool.name === 'automation_config');
|
||||
expect(automationConfigTool).toBeDefined();
|
||||
|
||||
if (!automationConfigTool) {
|
||||
throw new Error('automation_config tool not found');
|
||||
}
|
||||
|
||||
const result = await automationConfigTool.execute({
|
||||
action: 'create'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Configuration is required for creating automation');
|
||||
});
|
||||
|
||||
test('should require automation_id for update action', async () => {
|
||||
const automationConfigTool = addToolCalls.find(tool => tool.name === 'automation_config');
|
||||
expect(automationConfigTool).toBeDefined();
|
||||
|
||||
if (!automationConfigTool) {
|
||||
throw new Error('automation_config tool not found');
|
||||
}
|
||||
|
||||
const result = await automationConfigTool.execute({
|
||||
action: 'update',
|
||||
config: mockAutomationConfig
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Automation ID and configuration are required for updating automation');
|
||||
});
|
||||
});
|
||||
});
|
||||
191
__tests__/tools/automation.test.ts
Normal file
191
__tests__/tools/automation.test.ts
Normal file
@@ -0,0 +1,191 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import {
|
||||
type MockLiteMCPInstance,
|
||||
type Tool,
|
||||
type TestResponse,
|
||||
TEST_CONFIG,
|
||||
createMockLiteMCPInstance,
|
||||
setupTestEnvironment,
|
||||
cleanupMocks,
|
||||
createMockResponse,
|
||||
getMockCallArgs
|
||||
} from '../utils/test-utils';
|
||||
|
||||
describe('Automation Tools', () => {
|
||||
let liteMcpInstance: MockLiteMCPInstance;
|
||||
let addToolCalls: Tool[];
|
||||
let mocks: ReturnType<typeof setupTestEnvironment>;
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup test environment
|
||||
mocks = setupTestEnvironment();
|
||||
liteMcpInstance = createMockLiteMCPInstance();
|
||||
|
||||
// Import the module which will execute the main function
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Get the mock instance and tool calls
|
||||
addToolCalls = liteMcpInstance.addTool.mock.calls.map(call => call.args[0]);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
cleanupMocks({ liteMcpInstance, ...mocks });
|
||||
});
|
||||
|
||||
describe('automation tool', () => {
|
||||
const mockAutomations = [
|
||||
{
|
||||
entity_id: 'automation.morning_routine',
|
||||
state: 'on',
|
||||
attributes: {
|
||||
friendly_name: 'Morning Routine',
|
||||
last_triggered: '2024-01-01T07:00:00Z'
|
||||
}
|
||||
},
|
||||
{
|
||||
entity_id: 'automation.night_mode',
|
||||
state: 'off',
|
||||
attributes: {
|
||||
friendly_name: 'Night Mode',
|
||||
last_triggered: '2024-01-01T22:00:00Z'
|
||||
}
|
||||
}
|
||||
];
|
||||
|
||||
test('should successfully list automations', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse(mockAutomations)));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const automationTool = addToolCalls.find(tool => tool.name === 'automation');
|
||||
expect(automationTool).toBeDefined();
|
||||
|
||||
if (!automationTool) {
|
||||
throw new Error('automation tool not found');
|
||||
}
|
||||
|
||||
const result = await automationTool.execute({
|
||||
action: 'list'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.automations).toEqual([
|
||||
{
|
||||
entity_id: 'automation.morning_routine',
|
||||
name: 'Morning Routine',
|
||||
state: 'on',
|
||||
last_triggered: '2024-01-01T07:00:00Z'
|
||||
},
|
||||
{
|
||||
entity_id: 'automation.night_mode',
|
||||
name: 'Night Mode',
|
||||
state: 'off',
|
||||
last_triggered: '2024-01-01T22:00:00Z'
|
||||
}
|
||||
]);
|
||||
});
|
||||
|
||||
test('should successfully toggle an automation', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const automationTool = addToolCalls.find(tool => tool.name === 'automation');
|
||||
expect(automationTool).toBeDefined();
|
||||
|
||||
if (!automationTool) {
|
||||
throw new Error('automation tool not found');
|
||||
}
|
||||
|
||||
const result = await automationTool.execute({
|
||||
action: 'toggle',
|
||||
automation_id: 'automation.morning_routine'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully toggled automation automation.morning_routine');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/automation/toggle`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'automation.morning_routine'
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should successfully trigger an automation', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const automationTool = addToolCalls.find(tool => tool.name === 'automation');
|
||||
expect(automationTool).toBeDefined();
|
||||
|
||||
if (!automationTool) {
|
||||
throw new Error('automation tool not found');
|
||||
}
|
||||
|
||||
const result = await automationTool.execute({
|
||||
action: 'trigger',
|
||||
automation_id: 'automation.morning_routine'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully triggered automation automation.morning_routine');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/automation/trigger`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'automation.morning_routine'
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should require automation_id for toggle and trigger actions', async () => {
|
||||
const automationTool = addToolCalls.find(tool => tool.name === 'automation');
|
||||
expect(automationTool).toBeDefined();
|
||||
|
||||
if (!automationTool) {
|
||||
throw new Error('automation tool not found');
|
||||
}
|
||||
|
||||
const result = await automationTool.execute({
|
||||
action: 'toggle'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Automation ID is required for toggle and trigger actions');
|
||||
});
|
||||
});
|
||||
});
|
||||
231
__tests__/tools/device-control.test.ts
Normal file
231
__tests__/tools/device-control.test.ts
Normal file
@@ -0,0 +1,231 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import { tools } from '../../src/index.js';
|
||||
import {
|
||||
TEST_CONFIG,
|
||||
createMockResponse,
|
||||
getMockCallArgs
|
||||
} from '../utils/test-utils';
|
||||
|
||||
describe('Device Control Tools', () => {
|
||||
let mocks: { mockFetch: ReturnType<typeof mock> };
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup mock fetch
|
||||
mocks = {
|
||||
mockFetch: mock(() => Promise.resolve(createMockResponse({})))
|
||||
};
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
await Promise.resolve();
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Reset mocks
|
||||
globalThis.fetch = undefined;
|
||||
});
|
||||
|
||||
describe('list_devices tool', () => {
|
||||
test('should successfully list devices', async () => {
|
||||
const mockDevices = [
|
||||
{
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: { brightness: 255 }
|
||||
},
|
||||
{
|
||||
entity_id: 'climate.bedroom',
|
||||
state: 'heat',
|
||||
attributes: { temperature: 22 }
|
||||
}
|
||||
];
|
||||
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse(mockDevices)));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const listDevicesTool = tools.find(tool => tool.name === 'list_devices');
|
||||
expect(listDevicesTool).toBeDefined();
|
||||
|
||||
if (!listDevicesTool) {
|
||||
throw new Error('list_devices tool not found');
|
||||
}
|
||||
|
||||
const result = await listDevicesTool.execute({});
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.devices).toEqual({
|
||||
light: [{
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: { brightness: 255 }
|
||||
}],
|
||||
climate: [{
|
||||
entity_id: 'climate.bedroom',
|
||||
state: 'heat',
|
||||
attributes: { temperature: 22 }
|
||||
}]
|
||||
});
|
||||
});
|
||||
|
||||
test('should handle fetch errors', async () => {
|
||||
// Setup error response
|
||||
mocks.mockFetch = mock(() => Promise.reject(new Error('Network error')));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const listDevicesTool = tools.find(tool => tool.name === 'list_devices');
|
||||
expect(listDevicesTool).toBeDefined();
|
||||
|
||||
if (!listDevicesTool) {
|
||||
throw new Error('list_devices tool not found');
|
||||
}
|
||||
|
||||
const result = await listDevicesTool.execute({});
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Network error');
|
||||
});
|
||||
});
|
||||
|
||||
describe('control tool', () => {
|
||||
test('should successfully control a light device', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const controlTool = tools.find(tool => tool.name === 'control');
|
||||
expect(controlTool).toBeDefined();
|
||||
|
||||
if (!controlTool) {
|
||||
throw new Error('control tool not found');
|
||||
}
|
||||
|
||||
const result = await controlTool.execute({
|
||||
command: 'turn_on',
|
||||
entity_id: 'light.living_room',
|
||||
brightness: 255
|
||||
});
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully executed turn_on for light.living_room');
|
||||
|
||||
// Verify the fetch call
|
||||
const calls = mocks.mockFetch.mock.calls;
|
||||
expect(calls.length).toBeGreaterThan(0);
|
||||
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/light/turn_on`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'light.living_room',
|
||||
brightness: 255
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should handle unsupported domains', async () => {
|
||||
const controlTool = tools.find(tool => tool.name === 'control');
|
||||
expect(controlTool).toBeDefined();
|
||||
|
||||
if (!controlTool) {
|
||||
throw new Error('control tool not found');
|
||||
}
|
||||
|
||||
const result = await controlTool.execute({
|
||||
command: 'turn_on',
|
||||
entity_id: 'unsupported.device'
|
||||
});
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Unsupported domain: unsupported');
|
||||
});
|
||||
|
||||
test('should handle service call errors', async () => {
|
||||
// Setup error response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(new Response(null, {
|
||||
status: 503,
|
||||
statusText: 'Service unavailable'
|
||||
})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const controlTool = tools.find(tool => tool.name === 'control');
|
||||
expect(controlTool).toBeDefined();
|
||||
|
||||
if (!controlTool) {
|
||||
throw new Error('control tool not found');
|
||||
}
|
||||
|
||||
const result = await controlTool.execute({
|
||||
command: 'turn_on',
|
||||
entity_id: 'light.living_room'
|
||||
});
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toContain('Failed to execute turn_on for light.living_room');
|
||||
});
|
||||
|
||||
test('should handle climate device controls', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({})));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const controlTool = tools.find(tool => tool.name === 'control');
|
||||
expect(controlTool).toBeDefined();
|
||||
|
||||
if (!controlTool) {
|
||||
throw new Error('control tool not found');
|
||||
}
|
||||
|
||||
const result = await controlTool.execute({
|
||||
command: 'set_temperature',
|
||||
entity_id: 'climate.bedroom',
|
||||
temperature: 22,
|
||||
target_temp_high: 24,
|
||||
target_temp_low: 20
|
||||
});
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully executed set_temperature for climate.bedroom');
|
||||
|
||||
// Verify the fetch call
|
||||
const calls = mocks.mockFetch.mock.calls;
|
||||
expect(calls.length).toBeGreaterThan(0);
|
||||
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/climate/set_temperature`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'climate.bedroom',
|
||||
temperature: 22,
|
||||
target_temp_high: 24,
|
||||
target_temp_low: 20
|
||||
})
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
192
__tests__/tools/entity-state.test.ts
Normal file
192
__tests__/tools/entity-state.test.ts
Normal file
@@ -0,0 +1,192 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import {
|
||||
type MockLiteMCPInstance,
|
||||
type Tool,
|
||||
type TestResponse,
|
||||
TEST_CONFIG,
|
||||
createMockLiteMCPInstance,
|
||||
setupTestEnvironment,
|
||||
cleanupMocks,
|
||||
createMockResponse,
|
||||
getMockCallArgs
|
||||
} from '../utils/test-utils';
|
||||
|
||||
describe('Entity State Tools', () => {
|
||||
let liteMcpInstance: MockLiteMCPInstance;
|
||||
let addToolCalls: Tool[];
|
||||
let mocks: ReturnType<typeof setupTestEnvironment>;
|
||||
|
||||
const mockEntityState = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: {
|
||||
brightness: 255,
|
||||
color_temp: 400,
|
||||
friendly_name: 'Living Room Light'
|
||||
},
|
||||
last_changed: '2024-03-20T12:00:00Z',
|
||||
last_updated: '2024-03-20T12:00:00Z',
|
||||
context: {
|
||||
id: 'test_context_id',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup test environment
|
||||
mocks = setupTestEnvironment();
|
||||
liteMcpInstance = createMockLiteMCPInstance();
|
||||
|
||||
// Import the module which will execute the main function
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Get the mock instance and tool calls
|
||||
addToolCalls = liteMcpInstance.addTool.mock.calls.map(call => call.args[0]);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
cleanupMocks({ liteMcpInstance, ...mocks });
|
||||
});
|
||||
|
||||
describe('entity_state tool', () => {
|
||||
test('should successfully get entity state', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse(mockEntityState)));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const entityStateTool = addToolCalls.find(tool => tool.name === 'entity_state');
|
||||
expect(entityStateTool).toBeDefined();
|
||||
|
||||
if (!entityStateTool) {
|
||||
throw new Error('entity_state tool not found');
|
||||
}
|
||||
|
||||
const result = await entityStateTool.execute({
|
||||
entity_id: 'light.living_room'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.state).toBe('on');
|
||||
expect(result.attributes).toEqual(mockEntityState.attributes);
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/states/light.living_room`);
|
||||
expect(options).toEqual({
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
test('should handle entity not found', async () => {
|
||||
// Setup error response
|
||||
mocks.mockFetch = mock(() => Promise.reject(new Error('Entity not found')));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const entityStateTool = addToolCalls.find(tool => tool.name === 'entity_state');
|
||||
expect(entityStateTool).toBeDefined();
|
||||
|
||||
if (!entityStateTool) {
|
||||
throw new Error('entity_state tool not found');
|
||||
}
|
||||
|
||||
const result = await entityStateTool.execute({
|
||||
entity_id: 'light.non_existent'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Failed to get entity state: Entity not found');
|
||||
});
|
||||
|
||||
test('should require entity_id', async () => {
|
||||
const entityStateTool = addToolCalls.find(tool => tool.name === 'entity_state');
|
||||
expect(entityStateTool).toBeDefined();
|
||||
|
||||
if (!entityStateTool) {
|
||||
throw new Error('entity_state tool not found');
|
||||
}
|
||||
|
||||
const result = await entityStateTool.execute({}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Entity ID is required');
|
||||
});
|
||||
|
||||
test('should handle invalid entity_id format', async () => {
|
||||
const entityStateTool = addToolCalls.find(tool => tool.name === 'entity_state');
|
||||
expect(entityStateTool).toBeDefined();
|
||||
|
||||
if (!entityStateTool) {
|
||||
throw new Error('entity_state tool not found');
|
||||
}
|
||||
|
||||
const result = await entityStateTool.execute({
|
||||
entity_id: 'invalid_entity_id'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Invalid entity ID format: invalid_entity_id');
|
||||
});
|
||||
|
||||
test('should successfully get multiple entity states', async () => {
|
||||
// Setup response
|
||||
const mockStates = [
|
||||
{ ...mockEntityState },
|
||||
{
|
||||
...mockEntityState,
|
||||
entity_id: 'light.kitchen',
|
||||
attributes: { ...mockEntityState.attributes, friendly_name: 'Kitchen Light' }
|
||||
}
|
||||
];
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse(mockStates)));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const entityStateTool = addToolCalls.find(tool => tool.name === 'entity_state');
|
||||
expect(entityStateTool).toBeDefined();
|
||||
|
||||
if (!entityStateTool) {
|
||||
throw new Error('entity_state tool not found');
|
||||
}
|
||||
|
||||
const result = await entityStateTool.execute({
|
||||
entity_id: ['light.living_room', 'light.kitchen']
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(Array.isArray(result.states)).toBe(true);
|
||||
expect(result.states).toHaveLength(2);
|
||||
expect(result.states[0].entity_id).toBe('light.living_room');
|
||||
expect(result.states[1].entity_id).toBe('light.kitchen');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/states`);
|
||||
expect(options).toEqual({
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
});
|
||||
});
|
||||
});
|
||||
2
__tests__/tools/scene-control.test.ts
Normal file
2
__tests__/tools/scene-control.test.ts
Normal file
@@ -0,0 +1,2 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
|
||||
218
__tests__/tools/script-control.test.ts
Normal file
218
__tests__/tools/script-control.test.ts
Normal file
@@ -0,0 +1,218 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import {
|
||||
type MockLiteMCPInstance,
|
||||
type Tool,
|
||||
type TestResponse,
|
||||
TEST_CONFIG,
|
||||
createMockLiteMCPInstance,
|
||||
setupTestEnvironment,
|
||||
cleanupMocks,
|
||||
createMockResponse,
|
||||
getMockCallArgs
|
||||
} from '../utils/test-utils';
|
||||
|
||||
describe('Script Control Tools', () => {
|
||||
let liteMcpInstance: MockLiteMCPInstance;
|
||||
let addToolCalls: Tool[];
|
||||
let mocks: ReturnType<typeof setupTestEnvironment>;
|
||||
|
||||
beforeEach(async () => {
|
||||
// Setup test environment
|
||||
mocks = setupTestEnvironment();
|
||||
liteMcpInstance = createMockLiteMCPInstance();
|
||||
|
||||
// Import the module which will execute the main function
|
||||
await import('../../src/index.js');
|
||||
|
||||
// Get the mock instance and tool calls
|
||||
addToolCalls = liteMcpInstance.addTool.mock.calls.map(call => call.args[0]);
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
cleanupMocks({ liteMcpInstance, ...mocks });
|
||||
});
|
||||
|
||||
describe('script_control tool', () => {
|
||||
test('should successfully execute a script', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({ success: true })));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'script.welcome_home',
|
||||
action: 'start',
|
||||
variables: {
|
||||
brightness: 100,
|
||||
color_temp: 300
|
||||
}
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully executed script script.welcome_home');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/script/turn_on`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'script.welcome_home',
|
||||
variables: {
|
||||
brightness: 100,
|
||||
color_temp: 300
|
||||
}
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should successfully stop a script', async () => {
|
||||
// Setup response
|
||||
mocks.mockFetch = mock(() => Promise.resolve(createMockResponse({ success: true })));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'script.welcome_home',
|
||||
action: 'stop'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(true);
|
||||
expect(result.message).toBe('Successfully stopped script script.welcome_home');
|
||||
|
||||
// Verify the fetch call
|
||||
type FetchArgs = [url: string, init: RequestInit];
|
||||
const args = getMockCallArgs<FetchArgs>(mocks.mockFetch);
|
||||
expect(args).toBeDefined();
|
||||
|
||||
if (!args) {
|
||||
throw new Error('No fetch calls recorded');
|
||||
}
|
||||
|
||||
const [urlStr, options] = args;
|
||||
expect(urlStr).toBe(`${TEST_CONFIG.HASS_HOST}/api/services/script/turn_off`);
|
||||
expect(options).toEqual({
|
||||
method: 'POST',
|
||||
headers: {
|
||||
Authorization: `Bearer ${TEST_CONFIG.HASS_TOKEN}`,
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
entity_id: 'script.welcome_home'
|
||||
})
|
||||
});
|
||||
});
|
||||
|
||||
test('should handle script execution failure', async () => {
|
||||
// Setup error response
|
||||
mocks.mockFetch = mock(() => Promise.reject(new Error('Failed to execute script')));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'script.welcome_home',
|
||||
action: 'start'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Failed to execute script: Failed to execute script');
|
||||
});
|
||||
|
||||
test('should require script_id', async () => {
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
action: 'start'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Script ID is required');
|
||||
});
|
||||
|
||||
test('should require action', async () => {
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'script.welcome_home'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Action is required');
|
||||
});
|
||||
|
||||
test('should handle invalid script_id format', async () => {
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'invalid_script_id',
|
||||
action: 'start'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Invalid script ID format: invalid_script_id');
|
||||
});
|
||||
|
||||
test('should handle invalid action', async () => {
|
||||
const scriptControlTool = addToolCalls.find(tool => tool.name === 'script_control');
|
||||
expect(scriptControlTool).toBeDefined();
|
||||
|
||||
if (!scriptControlTool) {
|
||||
throw new Error('script_control tool not found');
|
||||
}
|
||||
|
||||
const result = await scriptControlTool.execute({
|
||||
script_id: 'script.welcome_home',
|
||||
action: 'invalid_action'
|
||||
}) as TestResponse;
|
||||
|
||||
expect(result.success).toBe(false);
|
||||
expect(result.message).toBe('Invalid action: invalid_action');
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -1,3 +1,4 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
import { ToolRegistry, ToolCategory, EnhancedTool } from '../../src/tools/index.js';
|
||||
|
||||
describe('ToolRegistry', () => {
|
||||
@@ -18,27 +19,27 @@ describe('ToolRegistry', () => {
|
||||
ttl: 1000
|
||||
}
|
||||
},
|
||||
execute: jest.fn().mockResolvedValue({ success: true }),
|
||||
validate: jest.fn().mockResolvedValue(true),
|
||||
preExecute: jest.fn().mockResolvedValue(undefined),
|
||||
postExecute: jest.fn().mockResolvedValue(undefined)
|
||||
execute: mock().mockResolvedValue({ success: true }),
|
||||
validate: mock().mockResolvedValue(true),
|
||||
preExecute: mock().mockResolvedValue(undefined),
|
||||
postExecute: mock().mockResolvedValue(undefined)
|
||||
};
|
||||
});
|
||||
|
||||
describe('Tool Registration', () => {
|
||||
it('should register a tool successfully', () => {
|
||||
test('should register a tool successfully', () => {
|
||||
registry.registerTool(mockTool);
|
||||
const retrievedTool = registry.getTool('test_tool');
|
||||
expect(retrievedTool).toBe(mockTool);
|
||||
});
|
||||
|
||||
it('should categorize tools correctly', () => {
|
||||
test('should categorize tools correctly', () => {
|
||||
registry.registerTool(mockTool);
|
||||
const deviceTools = registry.getToolsByCategory(ToolCategory.DEVICE);
|
||||
expect(deviceTools).toContain(mockTool);
|
||||
});
|
||||
|
||||
it('should handle multiple tools in the same category', () => {
|
||||
test('should handle multiple tools in the same category', () => {
|
||||
const mockTool2 = {
|
||||
...mockTool,
|
||||
name: 'test_tool_2'
|
||||
@@ -53,7 +54,7 @@ describe('ToolRegistry', () => {
|
||||
});
|
||||
|
||||
describe('Tool Execution', () => {
|
||||
it('should execute a tool with all hooks', async () => {
|
||||
test('should execute a tool with all hooks', async () => {
|
||||
registry.registerTool(mockTool);
|
||||
await registry.executeTool('test_tool', { param: 'value' });
|
||||
|
||||
@@ -63,20 +64,20 @@ describe('ToolRegistry', () => {
|
||||
expect(mockTool.postExecute).toHaveBeenCalled();
|
||||
});
|
||||
|
||||
it('should throw error for non-existent tool', async () => {
|
||||
test('should throw error for non-existent tool', async () => {
|
||||
await expect(registry.executeTool('non_existent', {}))
|
||||
.rejects.toThrow('Tool non_existent not found');
|
||||
});
|
||||
|
||||
it('should handle validation failure', async () => {
|
||||
mockTool.validate = jest.fn().mockResolvedValue(false);
|
||||
test('should handle validation failure', async () => {
|
||||
mockTool.validate = mock().mockResolvedValue(false);
|
||||
registry.registerTool(mockTool);
|
||||
|
||||
await expect(registry.executeTool('test_tool', {}))
|
||||
.rejects.toThrow('Invalid parameters');
|
||||
});
|
||||
|
||||
it('should execute without optional hooks', async () => {
|
||||
test('should execute without optional hooks', async () => {
|
||||
const simpleTool: EnhancedTool = {
|
||||
name: 'simple_tool',
|
||||
description: 'A simple tool',
|
||||
@@ -85,7 +86,7 @@ describe('ToolRegistry', () => {
|
||||
platform: 'test',
|
||||
version: '1.0.0'
|
||||
},
|
||||
execute: jest.fn().mockResolvedValue({ success: true })
|
||||
execute: mock().mockResolvedValue({ success: true })
|
||||
};
|
||||
|
||||
registry.registerTool(simpleTool);
|
||||
@@ -95,7 +96,7 @@ describe('ToolRegistry', () => {
|
||||
});
|
||||
|
||||
describe('Caching', () => {
|
||||
it('should cache tool results when enabled', async () => {
|
||||
test('should cache tool results when enabled', async () => {
|
||||
registry.registerTool(mockTool);
|
||||
const params = { test: 'value' };
|
||||
|
||||
@@ -108,7 +109,7 @@ describe('ToolRegistry', () => {
|
||||
expect(mockTool.execute).toHaveBeenCalledTimes(1);
|
||||
});
|
||||
|
||||
it('should not cache results when disabled', async () => {
|
||||
test('should not cache results when disabled', async () => {
|
||||
const uncachedTool: EnhancedTool = {
|
||||
...mockTool,
|
||||
metadata: {
|
||||
@@ -130,7 +131,7 @@ describe('ToolRegistry', () => {
|
||||
expect(uncachedTool.execute).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it('should expire cache after TTL', async () => {
|
||||
test('should expire cache after TTL', async () => {
|
||||
mockTool.metadata.caching!.ttl = 100; // Short TTL for testing
|
||||
registry.registerTool(mockTool);
|
||||
const params = { test: 'value' };
|
||||
@@ -147,7 +148,7 @@ describe('ToolRegistry', () => {
|
||||
expect(mockTool.execute).toHaveBeenCalledTimes(2);
|
||||
});
|
||||
|
||||
it('should clean expired cache entries', async () => {
|
||||
test('should clean expired cache entries', async () => {
|
||||
mockTool.metadata.caching!.ttl = 100;
|
||||
registry.registerTool(mockTool);
|
||||
const params = { test: 'value' };
|
||||
@@ -168,12 +169,12 @@ describe('ToolRegistry', () => {
|
||||
});
|
||||
|
||||
describe('Category Management', () => {
|
||||
it('should return empty array for unknown category', () => {
|
||||
test('should return empty array for unknown category', () => {
|
||||
const tools = registry.getToolsByCategory('unknown' as ToolCategory);
|
||||
expect(tools).toEqual([]);
|
||||
});
|
||||
|
||||
it('should handle tools across multiple categories', () => {
|
||||
test('should handle tools across multiple categories', () => {
|
||||
const systemTool: EnhancedTool = {
|
||||
...mockTool,
|
||||
name: 'system_tool',
|
||||
|
||||
19
__tests__/types/litemcp.d.ts
vendored
Normal file
19
__tests__/types/litemcp.d.ts
vendored
Normal file
@@ -0,0 +1,19 @@
|
||||
declare module 'litemcp' {
|
||||
export interface Tool {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters: Record<string, unknown>;
|
||||
execute: (params: Record<string, unknown>) => Promise<unknown>;
|
||||
}
|
||||
|
||||
export interface LiteMCPOptions {
|
||||
name: string;
|
||||
version: string;
|
||||
}
|
||||
|
||||
export class LiteMCP {
|
||||
constructor(options: LiteMCPOptions);
|
||||
addTool(tool: Tool): void;
|
||||
start(): Promise<void>;
|
||||
}
|
||||
}
|
||||
149
__tests__/utils/test-utils.ts
Normal file
149
__tests__/utils/test-utils.ts
Normal file
@@ -0,0 +1,149 @@
|
||||
import { mock } from "bun:test";
|
||||
import type { Mock } from "bun:test";
|
||||
import type { WebSocket } from 'ws';
|
||||
|
||||
// Common Types
|
||||
export interface Tool {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters: Record<string, unknown>;
|
||||
execute: (params: Record<string, unknown>) => Promise<unknown>;
|
||||
}
|
||||
|
||||
export interface MockLiteMCPInstance {
|
||||
addTool: Mock<(tool: Tool) => void>;
|
||||
start: Mock<() => Promise<void>>;
|
||||
}
|
||||
|
||||
export interface MockServices {
|
||||
light: {
|
||||
turn_on: Mock<() => Promise<{ success: boolean }>>;
|
||||
turn_off: Mock<() => Promise<{ success: boolean }>>;
|
||||
};
|
||||
climate: {
|
||||
set_temperature: Mock<() => Promise<{ success: boolean }>>;
|
||||
};
|
||||
}
|
||||
|
||||
export interface MockHassInstance {
|
||||
services: MockServices;
|
||||
}
|
||||
|
||||
export type TestResponse = {
|
||||
success: boolean;
|
||||
message?: string;
|
||||
automation_id?: string;
|
||||
new_automation_id?: string;
|
||||
state?: string;
|
||||
attributes?: Record<string, any>;
|
||||
states?: Array<{
|
||||
entity_id: string;
|
||||
state: string;
|
||||
attributes: Record<string, any>;
|
||||
last_changed: string;
|
||||
last_updated: string;
|
||||
context: {
|
||||
id: string;
|
||||
parent_id: string | null;
|
||||
user_id: string | null;
|
||||
};
|
||||
}>;
|
||||
};
|
||||
|
||||
// Test Configuration
|
||||
export const TEST_CONFIG = {
|
||||
HASS_HOST: process.env.TEST_HASS_HOST || 'http://localhost:8123',
|
||||
HASS_TOKEN: process.env.TEST_HASS_TOKEN || 'test_token',
|
||||
HASS_SOCKET_URL: process.env.TEST_HASS_SOCKET_URL || 'ws://localhost:8123/api/websocket'
|
||||
} as const;
|
||||
|
||||
// Mock WebSocket Implementation
|
||||
export class MockWebSocket {
|
||||
public static readonly CONNECTING = 0;
|
||||
public static readonly OPEN = 1;
|
||||
public static readonly CLOSING = 2;
|
||||
public static readonly CLOSED = 3;
|
||||
|
||||
public readyState: 0 | 1 | 2 | 3 = MockWebSocket.OPEN;
|
||||
public bufferedAmount = 0;
|
||||
public extensions = '';
|
||||
public protocol = '';
|
||||
public url = '';
|
||||
public binaryType: 'arraybuffer' | 'nodebuffer' | 'fragments' = 'arraybuffer';
|
||||
|
||||
public onopen: ((event: any) => void) | null = null;
|
||||
public onerror: ((event: any) => void) | null = null;
|
||||
public onclose: ((event: any) => void) | null = null;
|
||||
public onmessage: ((event: any) => void) | null = null;
|
||||
|
||||
public addEventListener = mock(() => undefined);
|
||||
public removeEventListener = mock(() => undefined);
|
||||
public send = mock(() => undefined);
|
||||
public close = mock(() => undefined);
|
||||
public ping = mock(() => undefined);
|
||||
public pong = mock(() => undefined);
|
||||
public terminate = mock(() => undefined);
|
||||
|
||||
constructor(url: string | URL, protocols?: string | string[]) {
|
||||
this.url = url.toString();
|
||||
if (protocols) {
|
||||
this.protocol = Array.isArray(protocols) ? protocols[0] : protocols;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Mock Service Instances
|
||||
export const createMockServices = (): MockServices => ({
|
||||
light: {
|
||||
turn_on: mock(() => Promise.resolve({ success: true })),
|
||||
turn_off: mock(() => Promise.resolve({ success: true }))
|
||||
},
|
||||
climate: {
|
||||
set_temperature: mock(() => Promise.resolve({ success: true }))
|
||||
}
|
||||
});
|
||||
|
||||
export const createMockLiteMCPInstance = (): MockLiteMCPInstance => ({
|
||||
addTool: mock((tool: Tool) => undefined),
|
||||
start: mock(() => Promise.resolve())
|
||||
});
|
||||
|
||||
// Helper Functions
|
||||
export const createMockResponse = <T>(data: T, status = 200): Response => {
|
||||
return new Response(JSON.stringify(data), { status });
|
||||
};
|
||||
|
||||
export const getMockCallArgs = <T extends unknown[]>(
|
||||
mock: Mock<(...args: any[]) => any>,
|
||||
callIndex = 0
|
||||
): T | undefined => {
|
||||
const call = mock.mock.calls[callIndex];
|
||||
return call?.args as T | undefined;
|
||||
};
|
||||
|
||||
export const setupTestEnvironment = () => {
|
||||
// Setup test environment variables
|
||||
Object.entries(TEST_CONFIG).forEach(([key, value]) => {
|
||||
process.env[key] = value;
|
||||
});
|
||||
|
||||
// Create fetch mock
|
||||
const mockFetch = mock(() => Promise.resolve(createMockResponse({ state: 'connected' })));
|
||||
|
||||
// Override globals
|
||||
globalThis.fetch = mockFetch;
|
||||
globalThis.WebSocket = MockWebSocket as any;
|
||||
|
||||
return { mockFetch };
|
||||
};
|
||||
|
||||
export const cleanupMocks = (mocks: {
|
||||
liteMcpInstance: MockLiteMCPInstance;
|
||||
mockFetch: Mock<() => Promise<Response>>;
|
||||
}) => {
|
||||
// Reset mock calls by creating a new mock
|
||||
mocks.liteMcpInstance.addTool = mock((tool: Tool) => undefined);
|
||||
mocks.liteMcpInstance.start = mock(() => Promise.resolve());
|
||||
mocks.mockFetch = mock(() => Promise.resolve(new Response()));
|
||||
globalThis.fetch = mocks.mockFetch;
|
||||
};
|
||||
@@ -1 +1,2 @@
|
||||
import { describe, expect, test } from "bun:test";
|
||||
|
||||
@@ -1,119 +1,177 @@
|
||||
import { jest, describe, it, expect, beforeEach, afterEach } from '@jest/globals';
|
||||
import { HassWebSocketClient } from '../../src/websocket/client.js';
|
||||
import WebSocket from 'ws';
|
||||
import { EventEmitter } from 'events';
|
||||
import * as HomeAssistant from '../../src/types/hass.js';
|
||||
|
||||
// Mock WebSocket
|
||||
jest.mock('ws');
|
||||
import { describe, expect, test, beforeEach, afterEach, mock } from "bun:test";
|
||||
import { EventEmitter } from "events";
|
||||
import { HassWebSocketClient } from "../../src/websocket/client";
|
||||
import type { MessageEvent, ErrorEvent } from "ws";
|
||||
import { Mock, fn as jestMock } from 'jest-mock';
|
||||
import { expect as jestExpect } from '@jest/globals';
|
||||
|
||||
describe('WebSocket Event Handling', () => {
|
||||
let client: HassWebSocketClient;
|
||||
let mockWebSocket: jest.Mocked<WebSocket>;
|
||||
let mockWebSocket: any;
|
||||
let onOpenCallback: () => void;
|
||||
let onCloseCallback: () => void;
|
||||
let onErrorCallback: (event: any) => void;
|
||||
let onMessageCallback: (event: any) => void;
|
||||
let eventEmitter: EventEmitter;
|
||||
|
||||
beforeEach(() => {
|
||||
// Clear all mocks
|
||||
jest.clearAllMocks();
|
||||
|
||||
// Create event emitter for mocking WebSocket events
|
||||
eventEmitter = new EventEmitter();
|
||||
|
||||
// Create mock WebSocket instance
|
||||
// Initialize callbacks first
|
||||
onOpenCallback = () => { };
|
||||
onCloseCallback = () => { };
|
||||
onErrorCallback = () => { };
|
||||
onMessageCallback = () => { };
|
||||
|
||||
mockWebSocket = {
|
||||
on: jest.fn((event: string, listener: (...args: any[]) => void) => {
|
||||
eventEmitter.on(event, listener);
|
||||
return mockWebSocket;
|
||||
}),
|
||||
send: jest.fn(),
|
||||
close: jest.fn(),
|
||||
readyState: WebSocket.OPEN,
|
||||
removeAllListeners: jest.fn(),
|
||||
// Add required WebSocket properties
|
||||
binaryType: 'arraybuffer',
|
||||
bufferedAmount: 0,
|
||||
extensions: '',
|
||||
protocol: '',
|
||||
url: 'ws://test.com',
|
||||
isPaused: () => false,
|
||||
ping: jest.fn(),
|
||||
pong: jest.fn(),
|
||||
terminate: jest.fn()
|
||||
} as unknown as jest.Mocked<WebSocket>;
|
||||
send: mock(),
|
||||
close: mock(),
|
||||
readyState: 1,
|
||||
OPEN: 1,
|
||||
onopen: null,
|
||||
onclose: null,
|
||||
onerror: null,
|
||||
onmessage: null
|
||||
};
|
||||
|
||||
// Mock WebSocket constructor
|
||||
(WebSocket as unknown as jest.Mock).mockImplementation(() => mockWebSocket);
|
||||
// Define setters that store the callbacks
|
||||
Object.defineProperties(mockWebSocket, {
|
||||
onopen: {
|
||||
get() { return onOpenCallback; },
|
||||
set(callback: () => void) { onOpenCallback = callback; }
|
||||
},
|
||||
onclose: {
|
||||
get() { return onCloseCallback; },
|
||||
set(callback: () => void) { onCloseCallback = callback; }
|
||||
},
|
||||
onerror: {
|
||||
get() { return onErrorCallback; },
|
||||
set(callback: (event: any) => void) { onErrorCallback = callback; }
|
||||
},
|
||||
onmessage: {
|
||||
get() { return onMessageCallback; },
|
||||
set(callback: (event: any) => void) { onMessageCallback = callback; }
|
||||
}
|
||||
});
|
||||
|
||||
// Create client instance
|
||||
client = new HassWebSocketClient('ws://test.com', 'test-token');
|
||||
// @ts-expect-error - Mock WebSocket implementation
|
||||
global.WebSocket = mock(() => mockWebSocket);
|
||||
|
||||
client = new HassWebSocketClient('ws://localhost:8123/api/websocket', 'test-token');
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
if (eventEmitter) {
|
||||
eventEmitter.removeAllListeners();
|
||||
}
|
||||
if (client) {
|
||||
client.disconnect();
|
||||
}
|
||||
});
|
||||
|
||||
it('should handle connection events', () => {
|
||||
// Simulate open event
|
||||
eventEmitter.emit('open');
|
||||
|
||||
// Verify authentication message was sent
|
||||
expect(mockWebSocket.send).toHaveBeenCalledWith(
|
||||
expect.stringContaining('"type":"auth"')
|
||||
);
|
||||
test('should handle connection events', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
await connectPromise;
|
||||
expect(client.isConnected()).toBe(true);
|
||||
});
|
||||
|
||||
it('should handle authentication response', () => {
|
||||
// Simulate auth_ok message
|
||||
eventEmitter.emit('message', JSON.stringify({ type: 'auth_ok' }));
|
||||
test('should handle authentication response', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
|
||||
// Verify client is ready for commands
|
||||
expect(mockWebSocket.readyState).toBe(WebSocket.OPEN);
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_required'
|
||||
})
|
||||
});
|
||||
|
||||
it('should handle auth failure', () => {
|
||||
// Simulate auth_invalid message
|
||||
eventEmitter.emit('message', JSON.stringify({
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_ok'
|
||||
})
|
||||
});
|
||||
|
||||
await connectPromise;
|
||||
expect(client.isAuthenticated()).toBe(true);
|
||||
});
|
||||
|
||||
test('should handle auth failure', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_required'
|
||||
})
|
||||
});
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_invalid',
|
||||
message: 'Invalid token'
|
||||
}));
|
||||
|
||||
// Verify client attempts to close connection
|
||||
expect(mockWebSocket.close).toHaveBeenCalled();
|
||||
message: 'Invalid password'
|
||||
})
|
||||
});
|
||||
|
||||
it('should handle connection errors', () => {
|
||||
// Create error spy
|
||||
const errorSpy = jest.fn();
|
||||
client.on('error', errorSpy);
|
||||
|
||||
// Simulate error
|
||||
const testError = new Error('Test error');
|
||||
eventEmitter.emit('error', testError);
|
||||
|
||||
// Verify error was handled
|
||||
expect(errorSpy).toHaveBeenCalledWith(testError);
|
||||
await expect(connectPromise).rejects.toThrow('Authentication failed');
|
||||
expect(client.isAuthenticated()).toBe(false);
|
||||
});
|
||||
|
||||
it('should handle disconnection', () => {
|
||||
// Create close spy
|
||||
const closeSpy = jest.fn();
|
||||
client.on('close', closeSpy);
|
||||
|
||||
// Simulate close
|
||||
eventEmitter.emit('close');
|
||||
|
||||
// Verify close was handled
|
||||
expect(closeSpy).toHaveBeenCalled();
|
||||
test('should handle connection errors', async () => {
|
||||
const errorPromise = new Promise((resolve) => {
|
||||
client.once('error', resolve);
|
||||
});
|
||||
|
||||
it('should handle event messages', () => {
|
||||
// Create event spy
|
||||
const eventSpy = jest.fn();
|
||||
client.on('event', eventSpy);
|
||||
const connectPromise = client.connect().catch(() => { /* Expected error */ });
|
||||
onOpenCallback();
|
||||
|
||||
const errorEvent = new Error('Connection failed');
|
||||
onErrorCallback({ error: errorEvent });
|
||||
|
||||
const error = await errorPromise;
|
||||
expect(error instanceof Error).toBe(true);
|
||||
expect((error as Error).message).toBe('Connection failed');
|
||||
});
|
||||
|
||||
test('should handle disconnection', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
await connectPromise;
|
||||
|
||||
const disconnectPromise = new Promise((resolve) => {
|
||||
client.on('disconnected', resolve);
|
||||
});
|
||||
|
||||
onCloseCallback();
|
||||
|
||||
await disconnectPromise;
|
||||
expect(client.isConnected()).toBe(false);
|
||||
});
|
||||
|
||||
test('should handle event messages', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_required'
|
||||
})
|
||||
});
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_ok'
|
||||
})
|
||||
});
|
||||
|
||||
await connectPromise;
|
||||
|
||||
const eventPromise = new Promise((resolve) => {
|
||||
client.on('state_changed', resolve);
|
||||
});
|
||||
|
||||
// Simulate event message
|
||||
const eventData = {
|
||||
id: 1,
|
||||
type: 'event',
|
||||
event: {
|
||||
event_type: 'state_changed',
|
||||
@@ -123,217 +181,63 @@ describe('WebSocket Event Handling', () => {
|
||||
}
|
||||
}
|
||||
};
|
||||
eventEmitter.emit('message', JSON.stringify(eventData));
|
||||
|
||||
// Verify event was handled
|
||||
expect(eventSpy).toHaveBeenCalledWith(eventData.event);
|
||||
onMessageCallback({
|
||||
data: JSON.stringify(eventData)
|
||||
});
|
||||
|
||||
describe('Connection Events', () => {
|
||||
it('should handle successful connection', (done) => {
|
||||
client.on('open', () => {
|
||||
const receivedEvent = await eventPromise;
|
||||
expect(receivedEvent).toEqual(eventData.event.data);
|
||||
});
|
||||
|
||||
test('should subscribe to specific events', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_required'
|
||||
})
|
||||
});
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_ok'
|
||||
})
|
||||
});
|
||||
|
||||
await connectPromise;
|
||||
|
||||
const subscriptionId = await client.subscribeEvents('state_changed', (data) => {
|
||||
// Empty callback for type satisfaction
|
||||
});
|
||||
expect(mockWebSocket.send).toHaveBeenCalled();
|
||||
done();
|
||||
expect(subscriptionId).toBeDefined();
|
||||
});
|
||||
|
||||
eventEmitter.emit('open');
|
||||
test('should unsubscribe from events', async () => {
|
||||
const connectPromise = client.connect();
|
||||
onOpenCallback();
|
||||
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_required'
|
||||
})
|
||||
});
|
||||
|
||||
it('should handle connection errors', (done) => {
|
||||
const error = new Error('Connection failed');
|
||||
client.on('error', (err: Error) => {
|
||||
expect(err).toBe(error);
|
||||
done();
|
||||
onMessageCallback({
|
||||
data: JSON.stringify({
|
||||
type: 'auth_ok'
|
||||
})
|
||||
});
|
||||
|
||||
eventEmitter.emit('error', error);
|
||||
});
|
||||
await connectPromise;
|
||||
|
||||
it('should handle connection close', (done) => {
|
||||
client.on('disconnected', () => {
|
||||
expect(mockWebSocket.close).toHaveBeenCalled();
|
||||
done();
|
||||
const subscriptionId = await client.subscribeEvents('state_changed', (data) => {
|
||||
// Empty callback for type satisfaction
|
||||
});
|
||||
await client.unsubscribeEvents(subscriptionId);
|
||||
|
||||
eventEmitter.emit('close');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Authentication', () => {
|
||||
it('should send authentication message on connect', () => {
|
||||
const authMessage: HomeAssistant.AuthMessage = {
|
||||
type: 'auth',
|
||||
access_token: 'test_token'
|
||||
};
|
||||
|
||||
client.connect();
|
||||
expect(mockWebSocket.send).toHaveBeenCalledWith(JSON.stringify(authMessage));
|
||||
});
|
||||
|
||||
it('should handle successful authentication', (done) => {
|
||||
client.on('auth_ok', () => {
|
||||
done();
|
||||
});
|
||||
|
||||
client.connect();
|
||||
eventEmitter.emit('message', JSON.stringify({ type: 'auth_ok' }));
|
||||
});
|
||||
|
||||
it('should handle authentication failure', (done) => {
|
||||
client.on('auth_invalid', () => {
|
||||
done();
|
||||
});
|
||||
|
||||
client.connect();
|
||||
eventEmitter.emit('message', JSON.stringify({ type: 'auth_invalid' }));
|
||||
});
|
||||
});
|
||||
|
||||
describe('Event Subscription', () => {
|
||||
it('should handle state changed events', (done) => {
|
||||
const stateEvent: HomeAssistant.StateChangedEvent = {
|
||||
event_type: 'state_changed',
|
||||
data: {
|
||||
entity_id: 'light.living_room',
|
||||
new_state: {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on',
|
||||
attributes: { brightness: 255 },
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
},
|
||||
old_state: {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'off',
|
||||
attributes: {},
|
||||
last_changed: '2024-01-01T00:00:00Z',
|
||||
last_updated: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '122',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
}
|
||||
},
|
||||
origin: 'LOCAL',
|
||||
time_fired: '2024-01-01T00:00:00Z',
|
||||
context: {
|
||||
id: '123',
|
||||
parent_id: null,
|
||||
user_id: null
|
||||
}
|
||||
};
|
||||
|
||||
client.on('event', (event) => {
|
||||
expect(event.data.entity_id).toBe('light.living_room');
|
||||
expect(event.data.new_state.state).toBe('on');
|
||||
expect(event.data.old_state.state).toBe('off');
|
||||
done();
|
||||
});
|
||||
|
||||
eventEmitter.emit('message', JSON.stringify({ type: 'event', event: stateEvent }));
|
||||
});
|
||||
|
||||
it('should subscribe to specific events', async () => {
|
||||
const subscriptionId = 1;
|
||||
const callback = jest.fn();
|
||||
|
||||
// Mock successful subscription
|
||||
const subscribePromise = client.subscribeEvents('state_changed', callback);
|
||||
eventEmitter.emit('message', JSON.stringify({
|
||||
id: 1,
|
||||
type: 'result',
|
||||
success: true
|
||||
}));
|
||||
|
||||
await expect(subscribePromise).resolves.toBe(subscriptionId);
|
||||
|
||||
// Test event handling
|
||||
const eventData = {
|
||||
entity_id: 'light.living_room',
|
||||
state: 'on'
|
||||
};
|
||||
eventEmitter.emit('message', JSON.stringify({
|
||||
type: 'event',
|
||||
event: {
|
||||
event_type: 'state_changed',
|
||||
data: eventData
|
||||
}
|
||||
}));
|
||||
|
||||
expect(callback).toHaveBeenCalledWith(eventData);
|
||||
});
|
||||
|
||||
it('should unsubscribe from events', async () => {
|
||||
// First subscribe
|
||||
const subscriptionId = await client.subscribeEvents('state_changed', () => { });
|
||||
|
||||
// Then unsubscribe
|
||||
const unsubscribePromise = client.unsubscribeEvents(subscriptionId);
|
||||
eventEmitter.emit('message', JSON.stringify({
|
||||
id: 2,
|
||||
type: 'result',
|
||||
success: true
|
||||
}));
|
||||
|
||||
await expect(unsubscribePromise).resolves.toBeUndefined();
|
||||
});
|
||||
});
|
||||
|
||||
describe('Message Handling', () => {
|
||||
it('should handle malformed messages', (done) => {
|
||||
client.on('error', (error: Error) => {
|
||||
expect(error.message).toContain('Unexpected token');
|
||||
done();
|
||||
});
|
||||
|
||||
eventEmitter.emit('message', 'invalid json');
|
||||
});
|
||||
|
||||
it('should handle unknown message types', (done) => {
|
||||
const unknownMessage = {
|
||||
type: 'unknown_type',
|
||||
data: {}
|
||||
};
|
||||
|
||||
client.on('error', (error: Error) => {
|
||||
expect(error.message).toContain('Unknown message type');
|
||||
done();
|
||||
});
|
||||
|
||||
eventEmitter.emit('message', JSON.stringify(unknownMessage));
|
||||
});
|
||||
});
|
||||
|
||||
describe('Reconnection', () => {
|
||||
it('should attempt to reconnect on connection loss', (done) => {
|
||||
let reconnectAttempts = 0;
|
||||
client.on('disconnected', () => {
|
||||
reconnectAttempts++;
|
||||
if (reconnectAttempts === 1) {
|
||||
expect(WebSocket).toHaveBeenCalledTimes(2);
|
||||
done();
|
||||
}
|
||||
});
|
||||
|
||||
eventEmitter.emit('close');
|
||||
});
|
||||
|
||||
it('should re-authenticate after reconnection', (done) => {
|
||||
client.connect();
|
||||
|
||||
client.on('auth_ok', () => {
|
||||
done();
|
||||
});
|
||||
|
||||
eventEmitter.emit('close');
|
||||
eventEmitter.emit('open');
|
||||
eventEmitter.emit('message', JSON.stringify({ type: 'auth_ok' }));
|
||||
});
|
||||
expect(mockWebSocket.send).toHaveBeenCalled();
|
||||
});
|
||||
});
|
||||
84
bin/mcp-stdio.cjs
Executable file
84
bin/mcp-stdio.cjs
Executable file
@@ -0,0 +1,84 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
const dotenv = require('dotenv');
|
||||
|
||||
/**
|
||||
* MCP Server - Stdio Transport Mode (CommonJS)
|
||||
*
|
||||
* This is the CommonJS entry point for running the MCP server via NPX in stdio mode.
|
||||
* It will directly load the stdio-server.js file which is optimized for the CLI usage.
|
||||
*/
|
||||
|
||||
// Set environment variable for stdio transport
|
||||
process.env.USE_STDIO_TRANSPORT = 'true';
|
||||
|
||||
// Load environment variables from .env file (if exists)
|
||||
try {
|
||||
const envPath = path.resolve(process.cwd(), '.env');
|
||||
if (fs.existsSync(envPath)) {
|
||||
dotenv.config({ path: envPath });
|
||||
} else {
|
||||
// Load .env.example if it exists
|
||||
const examplePath = path.resolve(process.cwd(), '.env.example');
|
||||
if (fs.existsSync(examplePath)) {
|
||||
dotenv.config({ path: examplePath });
|
||||
}
|
||||
}
|
||||
} catch (error) {
|
||||
// Silent error handling
|
||||
}
|
||||
|
||||
// Ensure logs directory exists
|
||||
try {
|
||||
const logsDir = path.join(process.cwd(), 'logs');
|
||||
if (!fs.existsSync(logsDir)) {
|
||||
fs.mkdirSync(logsDir, { recursive: true });
|
||||
}
|
||||
} catch (error) {
|
||||
// Silent error handling
|
||||
}
|
||||
|
||||
// Try to load the server
|
||||
try {
|
||||
// Check for simplified stdio server build first (preferred for CLI usage)
|
||||
const stdioServerPath = path.resolve(__dirname, '../dist/stdio-server.js');
|
||||
|
||||
if (fs.existsSync(stdioServerPath)) {
|
||||
// If we're running in Node.js (not Bun), we need to handle ESM imports differently
|
||||
if (typeof Bun === 'undefined') {
|
||||
// Use dynamic import for ESM modules in CommonJS
|
||||
import(stdioServerPath).catch((err) => {
|
||||
console.error('Failed to import stdio server:', err.message);
|
||||
process.exit(1);
|
||||
});
|
||||
} else {
|
||||
// In Bun, we can directly require the module
|
||||
require(stdioServerPath);
|
||||
}
|
||||
} else {
|
||||
// Fall back to full server if available
|
||||
const fullServerPath = path.resolve(__dirname, '../dist/index.js');
|
||||
|
||||
if (fs.existsSync(fullServerPath)) {
|
||||
console.warn('Warning: stdio-server.js not found, falling back to index.js');
|
||||
console.warn('For optimal CLI performance, build with "npm run build:stdio"');
|
||||
|
||||
if (typeof Bun === 'undefined') {
|
||||
import(fullServerPath).catch((err) => {
|
||||
console.error('Failed to import server:', err.message);
|
||||
process.exit(1);
|
||||
});
|
||||
} else {
|
||||
require(fullServerPath);
|
||||
}
|
||||
} else {
|
||||
console.error('Error: No server implementation found. Please build the project first.');
|
||||
process.exit(1);
|
||||
}
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Error starting server:', error.message);
|
||||
process.exit(1);
|
||||
}
|
||||
41
bin/mcp-stdio.js
Executable file
41
bin/mcp-stdio.js
Executable file
@@ -0,0 +1,41 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* MCP Server - Stdio Transport Mode
|
||||
*
|
||||
* This is the entry point for running the MCP server via NPX in stdio mode.
|
||||
* It automatically configures the server to use JSON-RPC 2.0 over stdin/stdout.
|
||||
*/
|
||||
|
||||
// Set environment variables for stdio transport
|
||||
process.env.USE_STDIO_TRANSPORT = 'true';
|
||||
|
||||
// Import and run the MCP server from the compiled output
|
||||
try {
|
||||
// First make sure required directories exist
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
// Ensure logs directory exists
|
||||
const logsDir = path.join(process.cwd(), 'logs');
|
||||
if (!fs.existsSync(logsDir)) {
|
||||
console.error('Creating logs directory...');
|
||||
fs.mkdirSync(logsDir, { recursive: true });
|
||||
}
|
||||
|
||||
// Get the entry module path
|
||||
const entryPath = require.resolve('../dist/index.js');
|
||||
|
||||
// Print initial message to stderr
|
||||
console.error('Starting MCP server in stdio transport mode...');
|
||||
console.error('Logs will be written to the logs/ directory');
|
||||
console.error('Communication will use JSON-RPC 2.0 format via stdin/stdout');
|
||||
|
||||
// Run the server
|
||||
require(entryPath);
|
||||
} catch (error) {
|
||||
console.error('Failed to start MCP server:', error.message);
|
||||
console.error('If this is your first run, you may need to build the project first:');
|
||||
console.error(' npm run build');
|
||||
process.exit(1);
|
||||
}
|
||||
150
bin/npx-entry.cjs
Executable file
150
bin/npx-entry.cjs
Executable file
@@ -0,0 +1,150 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
const { spawn } = require('child_process');
|
||||
|
||||
// Set environment variable - enable stdio transport
|
||||
process.env.USE_STDIO_TRANSPORT = 'true';
|
||||
|
||||
// Check if we're being called from Cursor (check for Cursor specific env vars)
|
||||
const isCursor = process.env.CURSOR_SESSION || process.env.CURSOR_CHANNEL;
|
||||
|
||||
// For Cursor, we need to ensure consistent stdio handling
|
||||
if (isCursor) {
|
||||
// Essential for Cursor compatibility
|
||||
process.env.LOG_LEVEL = 'info';
|
||||
process.env.CURSOR_COMPATIBLE = 'true';
|
||||
|
||||
// Ensure we have a clean environment for Cursor
|
||||
delete process.env.SILENT_MCP_RUNNING;
|
||||
} else {
|
||||
// For normal operation, silence logs
|
||||
process.env.LOG_LEVEL = 'silent';
|
||||
}
|
||||
|
||||
// Ensure logs directory exists
|
||||
const logsDir = path.join(process.cwd(), 'logs');
|
||||
if (!fs.existsSync(logsDir)) {
|
||||
fs.mkdirSync(logsDir, { recursive: true });
|
||||
}
|
||||
|
||||
// Check if .env exists, create from example if not
|
||||
const envPath = path.join(process.cwd(), '.env');
|
||||
const envExamplePath = path.join(process.cwd(), '.env.example');
|
||||
|
||||
if (!fs.existsSync(envPath) && fs.existsSync(envExamplePath)) {
|
||||
fs.copyFileSync(envExamplePath, envPath);
|
||||
}
|
||||
|
||||
// Define a function to ensure the child process is properly cleaned up on exit
|
||||
function setupCleanExit(childProcess) {
|
||||
const exitHandler = () => {
|
||||
if (childProcess && !childProcess.killed) {
|
||||
childProcess.kill();
|
||||
}
|
||||
process.exit();
|
||||
};
|
||||
|
||||
// Handle various termination signals
|
||||
process.on('SIGINT', exitHandler);
|
||||
process.on('SIGTERM', exitHandler);
|
||||
process.on('exit', exitHandler);
|
||||
}
|
||||
|
||||
// Start the MCP server
|
||||
try {
|
||||
// Critical: For Cursor, we need a very specific execution environment
|
||||
if (isCursor) {
|
||||
// Careful process cleanup for Cursor (optional but can help)
|
||||
try {
|
||||
const { execSync } = require('child_process');
|
||||
execSync('pkill -f "node.*stdio-server" || true', { stdio: 'ignore' });
|
||||
} catch (e) {
|
||||
// Ignore errors from process cleanup
|
||||
}
|
||||
|
||||
// Allow some time for process cleanup
|
||||
setTimeout(() => {
|
||||
const scriptPath = path.join(__dirname, 'mcp-stdio.cjs');
|
||||
|
||||
// For Cursor, we need very specific stdio handling
|
||||
// Using pipe for both stdin and stdout is critical
|
||||
const childProcess = spawn('node', [scriptPath], {
|
||||
stdio: ['pipe', 'pipe', 'pipe'], // All piped for maximum control
|
||||
env: {
|
||||
...process.env,
|
||||
USE_STDIO_TRANSPORT: 'true',
|
||||
CURSOR_COMPATIBLE: 'true',
|
||||
// Make sure stdin/stdout are treated as binary
|
||||
NODE_OPTIONS: '--no-force-async-hooks-checks'
|
||||
}
|
||||
});
|
||||
|
||||
// Ensure no buffering to prevent missed messages
|
||||
childProcess.stdin.setDefaultEncoding('utf8');
|
||||
|
||||
// Create bidirectional pipes
|
||||
process.stdin.pipe(childProcess.stdin);
|
||||
childProcess.stdout.pipe(process.stdout);
|
||||
childProcess.stderr.pipe(process.stderr);
|
||||
|
||||
// Setup error handling
|
||||
childProcess.on('error', (err) => {
|
||||
console.error('Failed to start server:', err.message);
|
||||
process.exit(1);
|
||||
});
|
||||
|
||||
// Ensure child process is properly cleaned up
|
||||
setupCleanExit(childProcess);
|
||||
|
||||
}, 500); // Short delay to ensure clean start
|
||||
}
|
||||
// For regular use, if silent-mcp.sh exists, use it
|
||||
else if (!isCursor && fs.existsSync(path.join(process.cwd(), 'silent-mcp.sh')) &&
|
||||
fs.statSync(path.join(process.cwd(), 'silent-mcp.sh')).isFile()) {
|
||||
// Execute the silent-mcp.sh script
|
||||
const childProcess = spawn('/bin/bash', [path.join(process.cwd(), 'silent-mcp.sh')], {
|
||||
stdio: ['inherit', 'inherit', 'ignore'], // Redirect stderr to /dev/null
|
||||
env: {
|
||||
...process.env,
|
||||
USE_STDIO_TRANSPORT: 'true',
|
||||
LOG_LEVEL: 'silent'
|
||||
}
|
||||
});
|
||||
|
||||
childProcess.on('error', (err) => {
|
||||
console.error('Failed to start server:', err.message);
|
||||
process.exit(1);
|
||||
});
|
||||
|
||||
// Ensure child process is properly cleaned up
|
||||
setupCleanExit(childProcess);
|
||||
}
|
||||
// Otherwise run normally (direct non-Cursor)
|
||||
else {
|
||||
const scriptPath = path.join(__dirname, 'mcp-stdio.cjs');
|
||||
|
||||
const childProcess = spawn('node', [scriptPath], {
|
||||
stdio: ['inherit', 'pipe', 'ignore'], // Redirect stderr to /dev/null for normal use
|
||||
env: {
|
||||
...process.env,
|
||||
USE_STDIO_TRANSPORT: 'true'
|
||||
}
|
||||
});
|
||||
|
||||
// Pipe child's stdout to parent's stdout
|
||||
childProcess.stdout.pipe(process.stdout);
|
||||
|
||||
childProcess.on('error', (err) => {
|
||||
console.error('Failed to start server:', err.message);
|
||||
process.exit(1);
|
||||
});
|
||||
|
||||
// Ensure child process is properly cleaned up
|
||||
setupCleanExit(childProcess);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error('Error starting server:', error.message);
|
||||
process.exit(1);
|
||||
}
|
||||
62
bin/test-stdio.js
Executable file
62
bin/test-stdio.js
Executable file
@@ -0,0 +1,62 @@
|
||||
#!/usr/bin/env node
|
||||
|
||||
/**
|
||||
* Test script for MCP stdio transport
|
||||
*
|
||||
* This script sends JSON-RPC 2.0 requests to the MCP server
|
||||
* running in stdio mode and displays the responses.
|
||||
*
|
||||
* Usage: node test-stdio.js | node bin/mcp-stdio.cjs
|
||||
*/
|
||||
|
||||
// Send a ping request
|
||||
const pingRequest = {
|
||||
jsonrpc: "2.0",
|
||||
id: 1,
|
||||
method: "ping"
|
||||
};
|
||||
|
||||
// Send an info request
|
||||
const infoRequest = {
|
||||
jsonrpc: "2.0",
|
||||
id: 2,
|
||||
method: "info"
|
||||
};
|
||||
|
||||
// Send an echo request
|
||||
const echoRequest = {
|
||||
jsonrpc: "2.0",
|
||||
id: 3,
|
||||
method: "echo",
|
||||
params: {
|
||||
message: "Hello, MCP!",
|
||||
timestamp: new Date().toISOString(),
|
||||
test: true,
|
||||
count: 42
|
||||
}
|
||||
};
|
||||
|
||||
// Send the requests with a delay between them
|
||||
setTimeout(() => {
|
||||
console.log(JSON.stringify(pingRequest));
|
||||
}, 500);
|
||||
|
||||
setTimeout(() => {
|
||||
console.log(JSON.stringify(infoRequest));
|
||||
}, 1000);
|
||||
|
||||
setTimeout(() => {
|
||||
console.log(JSON.stringify(echoRequest));
|
||||
}, 1500);
|
||||
|
||||
// Process responses
|
||||
process.stdin.on('data', (data) => {
|
||||
try {
|
||||
const response = JSON.parse(data.toString());
|
||||
console.error('Received response:');
|
||||
console.error(JSON.stringify(response, null, 2));
|
||||
} catch (error) {
|
||||
console.error('Error parsing response:', error);
|
||||
console.error('Raw data:', data.toString());
|
||||
}
|
||||
});
|
||||
41
bunfig.toml
41
bunfig.toml
@@ -1,5 +1,5 @@
|
||||
[test]
|
||||
preload = ["./src/__tests__/setup.ts"]
|
||||
preload = ["./test/setup.ts"]
|
||||
coverage = true
|
||||
coverageThreshold = {
|
||||
statements = 80,
|
||||
@@ -7,7 +7,7 @@ coverageThreshold = {
|
||||
functions = 80,
|
||||
lines = 80
|
||||
}
|
||||
timeout = 30000
|
||||
timeout = 10000
|
||||
testMatch = ["**/__tests__/**/*.test.ts"]
|
||||
testPathIgnorePatterns = ["/node_modules/", "/dist/"]
|
||||
collectCoverageFrom = [
|
||||
@@ -19,10 +19,34 @@ collectCoverageFrom = [
|
||||
]
|
||||
|
||||
[build]
|
||||
target = "node"
|
||||
target = "bun"
|
||||
outdir = "./dist"
|
||||
minify = true
|
||||
minify = {
|
||||
whitespace = true,
|
||||
syntax = true,
|
||||
identifiers = true,
|
||||
module = true
|
||||
}
|
||||
sourcemap = "external"
|
||||
entry = ["./src/index.ts", "./src/stdio-server.ts"]
|
||||
splitting = true
|
||||
naming = "[name].[hash].[ext]"
|
||||
publicPath = "/assets/"
|
||||
define = {
|
||||
"process.env.NODE_ENV": "process.env.NODE_ENV"
|
||||
}
|
||||
|
||||
[build.javascript]
|
||||
platform = "node"
|
||||
format = "esm"
|
||||
treeshaking = true
|
||||
packages = {
|
||||
external = ["bun:*"]
|
||||
}
|
||||
|
||||
[build.typescript]
|
||||
dts = true
|
||||
typecheck = true
|
||||
|
||||
[install]
|
||||
production = false
|
||||
@@ -48,3 +72,12 @@ reload = true
|
||||
[performance]
|
||||
gc = true
|
||||
optimize = true
|
||||
jit = true
|
||||
smol = true
|
||||
compact = true
|
||||
|
||||
[test.env]
|
||||
NODE_ENV = "test"
|
||||
|
||||
[watch]
|
||||
ignore = ["**/node_modules/**", "**/dist/**", "**/.git/**"]
|
||||
118
docker-build.sh
118
docker-build.sh
@@ -3,16 +3,52 @@
|
||||
# Enable error handling
|
||||
set -euo pipefail
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m'
|
||||
|
||||
# Function to print colored messages
|
||||
print_message() {
|
||||
local color=$1
|
||||
local message=$2
|
||||
echo -e "${color}${message}${NC}"
|
||||
}
|
||||
|
||||
# Function to clean up on script exit
|
||||
cleanup() {
|
||||
echo "Cleaning up..."
|
||||
print_message "$YELLOW" "Cleaning up..."
|
||||
docker builder prune -f --filter until=24h
|
||||
docker image prune -f
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
# Parse command line arguments
|
||||
ENABLE_SPEECH=false
|
||||
ENABLE_GPU=false
|
||||
BUILD_TYPE="standard"
|
||||
|
||||
while [[ $# -gt 0 ]]; do
|
||||
case $1 in
|
||||
--speech)
|
||||
ENABLE_SPEECH=true
|
||||
BUILD_TYPE="speech"
|
||||
shift
|
||||
;;
|
||||
--gpu)
|
||||
ENABLE_GPU=true
|
||||
shift
|
||||
;;
|
||||
*)
|
||||
print_message "$RED" "Unknown option: $1"
|
||||
exit 1
|
||||
;;
|
||||
esac
|
||||
done
|
||||
|
||||
# Clean up Docker system
|
||||
echo "Cleaning up Docker system..."
|
||||
print_message "$YELLOW" "Cleaning up Docker system..."
|
||||
docker system prune -f --volumes
|
||||
|
||||
# Set build arguments for better performance
|
||||
@@ -26,23 +62,47 @@ BUILD_MEM=$(( TOTAL_MEM / 2 )) # Use half of available memory
|
||||
CPU_COUNT=$(nproc)
|
||||
CPU_QUOTA=$(( CPU_COUNT * 50000 )) # Allow 50% CPU usage per core
|
||||
|
||||
echo "Building with ${BUILD_MEM}MB memory limit and CPU quota ${CPU_QUOTA}"
|
||||
print_message "$YELLOW" "Building with ${BUILD_MEM}MB memory limit and CPU quota ${CPU_QUOTA}"
|
||||
|
||||
# Remove any existing lockfile
|
||||
rm -f bun.lockb
|
||||
|
||||
# Build with resource limits, optimizations, and timeout
|
||||
echo "Building Docker image..."
|
||||
# Base build arguments
|
||||
BUILD_ARGS=(
|
||||
--memory="${BUILD_MEM}m"
|
||||
--memory-swap="${BUILD_MEM}m"
|
||||
--cpu-quota="${CPU_QUOTA}"
|
||||
--build-arg BUILDKIT_INLINE_CACHE=1
|
||||
--build-arg DOCKER_BUILDKIT=1
|
||||
--build-arg NODE_ENV=production
|
||||
--progress=plain
|
||||
--no-cache
|
||||
--compress
|
||||
)
|
||||
|
||||
# Add speech-specific build arguments if enabled
|
||||
if [ "$ENABLE_SPEECH" = true ]; then
|
||||
BUILD_ARGS+=(
|
||||
--build-arg ENABLE_SPEECH_FEATURES=true
|
||||
--build-arg ENABLE_WAKE_WORD=true
|
||||
--build-arg ENABLE_SPEECH_TO_TEXT=true
|
||||
)
|
||||
|
||||
# Add GPU support if requested
|
||||
if [ "$ENABLE_GPU" = true ]; then
|
||||
BUILD_ARGS+=(
|
||||
--build-arg CUDA_VISIBLE_DEVICES=0
|
||||
--build-arg COMPUTE_TYPE=float16
|
||||
)
|
||||
fi
|
||||
fi
|
||||
|
||||
# Build the images
|
||||
print_message "$YELLOW" "Building Docker image (${BUILD_TYPE} build)..."
|
||||
|
||||
# Build main image
|
||||
DOCKER_BUILDKIT=1 docker build \
|
||||
--memory="${BUILD_MEM}m" \
|
||||
--memory-swap="${BUILD_MEM}m" \
|
||||
--cpu-quota="${CPU_QUOTA}" \
|
||||
--build-arg BUILDKIT_INLINE_CACHE=1 \
|
||||
--build-arg DOCKER_BUILDKIT=1 \
|
||||
--build-arg NODE_ENV=production \
|
||||
--progress=plain \
|
||||
--no-cache \
|
||||
--compress \
|
||||
"${BUILD_ARGS[@]}" \
|
||||
-t homeassistant-mcp:latest \
|
||||
-t homeassistant-mcp:$(date +%Y%m%d) \
|
||||
.
|
||||
@@ -50,15 +110,39 @@ DOCKER_BUILDKIT=1 docker build \
|
||||
# Check if build was successful
|
||||
BUILD_EXIT_CODE=$?
|
||||
if [ $BUILD_EXIT_CODE -eq 124 ]; then
|
||||
echo "Build timed out after 15 minutes!"
|
||||
print_message "$RED" "Build timed out after 15 minutes!"
|
||||
exit 1
|
||||
elif [ $BUILD_EXIT_CODE -ne 0 ]; then
|
||||
echo "Build failed with exit code ${BUILD_EXIT_CODE}!"
|
||||
print_message "$RED" "Build failed with exit code ${BUILD_EXIT_CODE}!"
|
||||
exit 1
|
||||
else
|
||||
echo "Build completed successfully!"
|
||||
print_message "$GREEN" "Main image build completed successfully!"
|
||||
|
||||
# Show image size and layers
|
||||
docker image ls homeassistant-mcp:latest --format "Image size: {{.Size}}"
|
||||
echo "Layer count: $(docker history homeassistant-mcp:latest | wc -l)"
|
||||
fi
|
||||
|
||||
# Build speech-related images if enabled
|
||||
if [ "$ENABLE_SPEECH" = true ]; then
|
||||
print_message "$YELLOW" "Building speech-related images..."
|
||||
|
||||
# Build fast-whisper image
|
||||
print_message "$YELLOW" "Building fast-whisper image..."
|
||||
docker pull onerahmet/openai-whisper-asr-webservice:latest
|
||||
|
||||
# Build wake-word image
|
||||
print_message "$YELLOW" "Building wake-word image..."
|
||||
docker pull rhasspy/wyoming-openwakeword:latest
|
||||
|
||||
print_message "$GREEN" "Speech-related images built successfully!"
|
||||
fi
|
||||
|
||||
print_message "$GREEN" "All builds completed successfully!"
|
||||
|
||||
# Show final status
|
||||
print_message "$YELLOW" "Build Summary:"
|
||||
echo "Build Type: $BUILD_TYPE"
|
||||
echo "Speech Features: $([ "$ENABLE_SPEECH" = true ] && echo 'Enabled' || echo 'Disabled')"
|
||||
echo "GPU Support: $([ "$ENABLE_GPU" = true ] && echo 'Enabled' || echo 'Disabled')"
|
||||
docker image ls | grep -E 'homeassistant-mcp|whisper|openwakeword'
|
||||
73
docker-compose.speech.yml
Normal file
73
docker-compose.speech.yml
Normal file
@@ -0,0 +1,73 @@
|
||||
version: '3.8'
|
||||
|
||||
services:
|
||||
homeassistant-mcp:
|
||||
image: homeassistant-mcp:latest
|
||||
environment:
|
||||
# Speech Feature Flags
|
||||
- ENABLE_SPEECH_FEATURES=${ENABLE_SPEECH_FEATURES:-true}
|
||||
- ENABLE_WAKE_WORD=${ENABLE_WAKE_WORD:-true}
|
||||
- ENABLE_SPEECH_TO_TEXT=${ENABLE_SPEECH_TO_TEXT:-true}
|
||||
|
||||
# Audio Configuration
|
||||
- NOISE_THRESHOLD=${NOISE_THRESHOLD:-0.05}
|
||||
- MIN_SPEECH_DURATION=${MIN_SPEECH_DURATION:-1.0}
|
||||
- SILENCE_DURATION=${SILENCE_DURATION:-0.5}
|
||||
- SAMPLE_RATE=${SAMPLE_RATE:-16000}
|
||||
- CHANNELS=${CHANNELS:-1}
|
||||
- CHUNK_SIZE=${CHUNK_SIZE:-1024}
|
||||
- PULSE_SERVER=${PULSE_SERVER:-unix:/run/user/1000/pulse/native}
|
||||
|
||||
fast-whisper:
|
||||
image: onerahmet/openai-whisper-asr-webservice:latest
|
||||
volumes:
|
||||
- whisper-models:/models
|
||||
- audio-data:/audio
|
||||
environment:
|
||||
- ASR_MODEL=${WHISPER_MODEL_TYPE:-base}
|
||||
- ASR_ENGINE=faster_whisper
|
||||
- WHISPER_BEAM_SIZE=5
|
||||
- COMPUTE_TYPE=float32
|
||||
- LANGUAGE=en
|
||||
ports:
|
||||
- "9000:9000"
|
||||
deploy:
|
||||
resources:
|
||||
limits:
|
||||
cpus: '4.0'
|
||||
memory: 2G
|
||||
healthcheck:
|
||||
test: [ "CMD", "curl", "-f", "http://localhost:9000/health" ]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
|
||||
wake-word:
|
||||
image: rhasspy/wyoming-openwakeword:latest
|
||||
restart: unless-stopped
|
||||
devices:
|
||||
- /dev/snd:/dev/snd
|
||||
volumes:
|
||||
- /run/user/1000/pulse/native:/run/user/1000/pulse/native
|
||||
environment:
|
||||
- PULSE_SERVER=${PULSE_SERVER:-unix:/run/user/1000/pulse/native}
|
||||
- PULSE_COOKIE=/run/user/1000/pulse/cookie
|
||||
- PYTHONUNBUFFERED=1
|
||||
- OPENWAKEWORD_MODEL=hey_jarvis
|
||||
- OPENWAKEWORD_THRESHOLD=0.5
|
||||
- MICROPHONE_COMMAND=arecord -D hw:0,0 -f S16_LE -c 1 -r 16000 -t raw
|
||||
group_add:
|
||||
- "${AUDIO_GID:-29}"
|
||||
network_mode: host
|
||||
privileged: true
|
||||
entrypoint: >
|
||||
/bin/bash -c " apt-get update && apt-get install -y pulseaudio alsa-utils && rm -rf /var/lib/apt/lists/* && /run.sh"
|
||||
healthcheck:
|
||||
test: [ "CMD-SHELL", "pactl info > /dev/null 2>&1 || exit 1" ]
|
||||
interval: 30s
|
||||
timeout: 10s
|
||||
retries: 3
|
||||
|
||||
volumes:
|
||||
whisper-models:
|
||||
audio-data:
|
||||
@@ -1,58 +0,0 @@
|
||||
# Use Python slim image as builder
|
||||
FROM python:3.10-slim as builder
|
||||
|
||||
# Install build dependencies
|
||||
RUN apt-get update && apt-get install -y \
|
||||
git \
|
||||
build-essential \
|
||||
portaudio19-dev \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Create and activate virtual environment
|
||||
RUN python -m venv /opt/venv
|
||||
ENV PATH="/opt/venv/bin:$PATH"
|
||||
|
||||
# Install Python dependencies with specific versions and CPU-only variants
|
||||
RUN pip install --no-cache-dir torch==2.1.2 torchaudio==2.1.2 --index-url https://download.pytorch.org/whl/cpu
|
||||
RUN pip install --no-cache-dir faster-whisper==0.10.0 openwakeword==0.4.0 pyaudio==0.2.14 sounddevice==0.4.6
|
||||
|
||||
# Create final image
|
||||
FROM python:3.10-slim
|
||||
|
||||
# Copy virtual environment from builder
|
||||
COPY --from=builder /opt/venv /opt/venv
|
||||
ENV PATH="/opt/venv/bin:$PATH"
|
||||
|
||||
# Install only runtime dependencies
|
||||
RUN apt-get update && apt-get install -y \
|
||||
portaudio19-dev \
|
||||
python3-pyaudio \
|
||||
&& rm -rf /var/lib/apt/lists/*
|
||||
|
||||
# Create necessary directories
|
||||
RUN mkdir -p /models/wake_word /audio
|
||||
|
||||
# Set working directory
|
||||
WORKDIR /app
|
||||
|
||||
# Copy the wake word detection script
|
||||
COPY wake_word_detector.py .
|
||||
|
||||
# Set environment variables
|
||||
ENV WHISPER_MODEL_PATH=/models \
|
||||
WAKEWORD_MODEL_PATH=/models/wake_word \
|
||||
PYTHONUNBUFFERED=1 \
|
||||
ASR_MODEL=base.en \
|
||||
ASR_MODEL_PATH=/models
|
||||
|
||||
# Add resource limits to Python
|
||||
ENV PYTHONMALLOC=malloc \
|
||||
MALLOC_TRIM_THRESHOLD_=100000 \
|
||||
PYTHONDEVMODE=1
|
||||
|
||||
# Add healthcheck
|
||||
HEALTHCHECK --interval=30s --timeout=10s --start-period=5s --retries=3 \
|
||||
CMD ps aux | grep '[p]ython' || exit 1
|
||||
|
||||
# Run the wake word detection service with resource constraints
|
||||
CMD ["python", "-X", "faulthandler", "wake_word_detector.py"]
|
||||
35
docker/speech/asound.conf
Normal file
35
docker/speech/asound.conf
Normal file
@@ -0,0 +1,35 @@
|
||||
pcm.!default {
|
||||
type pulse
|
||||
fallback "sysdefault"
|
||||
hint {
|
||||
show on
|
||||
description "Default ALSA Output (currently PulseAudio Sound Server)"
|
||||
}
|
||||
}
|
||||
|
||||
ctl.!default {
|
||||
type pulse
|
||||
fallback "sysdefault"
|
||||
}
|
||||
|
||||
# Use PulseAudio by default
|
||||
pcm.pulse {
|
||||
type pulse
|
||||
}
|
||||
|
||||
ctl.pulse {
|
||||
type pulse
|
||||
}
|
||||
|
||||
# Explicit device for recording
|
||||
pcm.microphone {
|
||||
type hw
|
||||
card 0
|
||||
device 0
|
||||
}
|
||||
|
||||
# Default capture device
|
||||
pcm.!default {
|
||||
type pulse
|
||||
hint.description "Default Audio Device"
|
||||
}
|
||||
68
docker/speech/setup-audio.sh
Executable file
68
docker/speech/setup-audio.sh
Executable file
@@ -0,0 +1,68 @@
|
||||
#!/bin/bash
|
||||
set -e # Exit immediately if a command exits with a non-zero status
|
||||
set -x # Print commands and their arguments as they are executed
|
||||
|
||||
echo "Starting audio setup script at $(date)"
|
||||
echo "Current user: $(whoami)"
|
||||
echo "Current directory: $(pwd)"
|
||||
|
||||
# Print environment variables related to audio and speech
|
||||
echo "ENABLE_WAKE_WORD: ${ENABLE_WAKE_WORD}"
|
||||
echo "PULSE_SERVER: ${PULSE_SERVER}"
|
||||
echo "WHISPER_MODEL_PATH: ${WHISPER_MODEL_PATH}"
|
||||
|
||||
# Wait for PulseAudio socket to be available
|
||||
max_wait=30
|
||||
wait_count=0
|
||||
while [ ! -e /run/user/1000/pulse/native ]; do
|
||||
echo "Waiting for PulseAudio socket... (${wait_count}/${max_wait})"
|
||||
sleep 1
|
||||
wait_count=$((wait_count + 1))
|
||||
if [ $wait_count -ge $max_wait ]; then
|
||||
echo "ERROR: PulseAudio socket not available after ${max_wait} seconds"
|
||||
exit 1
|
||||
fi
|
||||
done
|
||||
|
||||
# Verify PulseAudio connection with detailed error handling
|
||||
if ! pactl info; then
|
||||
echo "ERROR: Failed to connect to PulseAudio server"
|
||||
pactl list short modules
|
||||
pactl list short clients
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# List audio devices with error handling
|
||||
if ! pactl list sources; then
|
||||
echo "ERROR: Failed to list audio devices"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Ensure wake word detector script is executable
|
||||
chmod +x /app/wake_word_detector.py
|
||||
|
||||
# Start the wake word detector with logging
|
||||
echo "Starting wake word detector at $(date)"
|
||||
python /app/wake_word_detector.py 2>&1 | tee /audio/wake_word_detector.log &
|
||||
wake_word_pid=$!
|
||||
|
||||
# Wait and check if the process is still running
|
||||
sleep 5
|
||||
if ! kill -0 $wake_word_pid 2>/dev/null; then
|
||||
echo "ERROR: Wake word detector process died immediately"
|
||||
cat /audio/wake_word_detector.log
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Mute the monitor to prevent feedback
|
||||
pactl set-source-mute alsa_output.pci-0000_00_1b.0.analog-stereo.monitor 1
|
||||
|
||||
# Set microphone sensitivity to 65%
|
||||
pactl set-source-volume alsa_input.pci-0000_00_1b.0.analog-stereo 65%
|
||||
|
||||
# Set speaker volume to 40%
|
||||
pactl set-sink-volume alsa_output.pci-0000_00_1b.0.analog-stereo 40%
|
||||
|
||||
# Keep the script running to prevent container exit
|
||||
echo "Audio setup complete. Keeping container alive."
|
||||
tail -f /dev/null
|
||||
@@ -8,46 +8,292 @@ from openwakeword import Model
|
||||
from datetime import datetime
|
||||
import wave
|
||||
from faster_whisper import WhisperModel
|
||||
import requests
|
||||
import logging
|
||||
import time
|
||||
|
||||
# Set up logging
|
||||
logging.basicConfig(
|
||||
level=logging.DEBUG,
|
||||
format='%(asctime)s - %(levelname)s - %(message)s'
|
||||
)
|
||||
logger = logging.getLogger(__name__)
|
||||
|
||||
# Configuration
|
||||
SAMPLE_RATE = 16000
|
||||
CHANNELS = 1
|
||||
CHUNK_SIZE = 1024
|
||||
BUFFER_DURATION = 30 # seconds to keep in buffer
|
||||
BUFFER_DURATION = 10 # seconds to keep in buffer
|
||||
DETECTION_THRESHOLD = 0.5
|
||||
CONTINUOUS_TRANSCRIPTION_INTERVAL = 3 # seconds between transcriptions
|
||||
MAX_MODEL_LOAD_RETRIES = 3
|
||||
MODEL_LOAD_RETRY_DELAY = 5 # seconds
|
||||
MODEL_DOWNLOAD_TIMEOUT = 600 # 10 minutes timeout for model download
|
||||
|
||||
# Wake word models to use
|
||||
WAKE_WORDS = ["hey_jarvis", "ok_google", "alexa"]
|
||||
# ALSA device configuration
|
||||
AUDIO_DEVICE = 'hw:0,0' # Use ALSA hardware device directly
|
||||
|
||||
# Initialize the ASR model
|
||||
asr_model = WhisperModel(
|
||||
model_size_or_path=os.environ.get('ASR_MODEL', 'base.en'),
|
||||
# Audio processing parameters
|
||||
NOISE_THRESHOLD = 0.08 # Increased threshold for better noise filtering
|
||||
MIN_SPEECH_DURATION = 2.0 # Longer minimum duration to avoid fragments
|
||||
SILENCE_DURATION = 1.0 # Longer silence duration
|
||||
MAX_REPETITIONS = 1 # More aggressive repetition filtering
|
||||
ECHO_THRESHOLD = 0.75 # More sensitive echo detection
|
||||
MIN_SEGMENT_DURATION = 1.0 # Longer minimum segment duration
|
||||
FEEDBACK_WINDOW = 5 # Window size for feedback detection in seconds
|
||||
|
||||
# Feature flags from environment
|
||||
WAKE_WORD_ENABLED = os.environ.get('ENABLE_WAKE_WORD', 'false').lower() == 'true'
|
||||
SPEECH_ENABLED = os.environ.get('ENABLE_SPEECH_FEATURES', 'true').lower() == 'true'
|
||||
|
||||
# Wake word models to use (only if wake word is enabled)
|
||||
WAKE_WORDS = ["hey_jarvis"] # Using hey_jarvis as it's more similar to "hey gaja"
|
||||
WAKE_WORD_ALIAS = "gaja" # What we print when wake word is detected
|
||||
|
||||
# Home Assistant Configuration
|
||||
HASS_HOST = os.environ.get('HASS_HOST', 'http://homeassistant.local:8123')
|
||||
HASS_TOKEN = os.environ.get('HASS_TOKEN')
|
||||
|
||||
def initialize_asr_model():
|
||||
"""Initialize the ASR model with retries and timeout"""
|
||||
model_path = os.environ.get('WHISPER_MODEL_PATH', '/models')
|
||||
model_name = os.environ.get('WHISPER_MODEL_TYPE', 'base')
|
||||
|
||||
start_time = time.time()
|
||||
for attempt in range(MAX_MODEL_LOAD_RETRIES):
|
||||
try:
|
||||
if time.time() - start_time > MODEL_DOWNLOAD_TIMEOUT:
|
||||
logger.error("Model download timeout exceeded")
|
||||
raise TimeoutError("Model download took too long")
|
||||
|
||||
logger.info(f"Loading ASR model (attempt {attempt + 1}/{MAX_MODEL_LOAD_RETRIES})")
|
||||
model = WhisperModel(
|
||||
model_size_or_path=model_name,
|
||||
device="cpu",
|
||||
compute_type="int8",
|
||||
download_root=os.environ.get('ASR_MODEL_PATH', '/models')
|
||||
download_root=model_path,
|
||||
num_workers=1 # Reduce concurrent downloads
|
||||
)
|
||||
logger.info("ASR model loaded successfully")
|
||||
return model
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to load ASR model (attempt {attempt + 1}): {e}")
|
||||
if attempt < MAX_MODEL_LOAD_RETRIES - 1:
|
||||
logger.info(f"Retrying in {MODEL_LOAD_RETRY_DELAY} seconds...")
|
||||
time.sleep(MODEL_LOAD_RETRY_DELAY)
|
||||
else:
|
||||
logger.error("Failed to load ASR model after all retries")
|
||||
raise
|
||||
|
||||
# Initialize the ASR model with retries
|
||||
try:
|
||||
asr_model = initialize_asr_model()
|
||||
except Exception as e:
|
||||
logger.error(f"Critical error initializing ASR model: {e}")
|
||||
raise
|
||||
|
||||
def send_command_to_hass(domain, service, entity_id):
|
||||
"""Send command to Home Assistant"""
|
||||
if not HASS_TOKEN:
|
||||
logger.error("Error: HASS_TOKEN not set")
|
||||
return False
|
||||
|
||||
headers = {
|
||||
"Authorization": f"Bearer {HASS_TOKEN}",
|
||||
"Content-Type": "application/json",
|
||||
}
|
||||
|
||||
url = f"{HASS_HOST}/api/services/{domain}/{service}"
|
||||
data = {"entity_id": entity_id}
|
||||
|
||||
try:
|
||||
response = requests.post(url, headers=headers, json=data)
|
||||
response.raise_for_status()
|
||||
logger.info(f"Command sent: {domain}.{service} for {entity_id}")
|
||||
return True
|
||||
except Exception as e:
|
||||
logger.error(f"Error sending command to Home Assistant: {e}")
|
||||
return False
|
||||
|
||||
def is_speech(audio_data, threshold=NOISE_THRESHOLD):
|
||||
"""Detect if audio segment contains speech based on amplitude and frequency content"""
|
||||
# Calculate RMS amplitude
|
||||
rms = np.sqrt(np.mean(np.square(audio_data)))
|
||||
|
||||
# Calculate signal energy in speech frequency range (100-4000 Hz)
|
||||
fft = np.fft.fft(audio_data)
|
||||
freqs = np.fft.fftfreq(len(audio_data), 1/SAMPLE_RATE)
|
||||
speech_mask = (np.abs(freqs) >= 100) & (np.abs(freqs) <= 4000)
|
||||
speech_energy = np.sum(np.abs(fft[speech_mask])) / len(audio_data)
|
||||
|
||||
# Enhanced echo detection
|
||||
# 1. Check for periodic patterns in the signal
|
||||
autocorr = np.correlate(audio_data, audio_data, mode='full')
|
||||
autocorr = autocorr[len(autocorr)//2:] # Use only positive lags
|
||||
peaks = np.where(autocorr > ECHO_THRESHOLD * np.max(autocorr))[0]
|
||||
peak_spacing = np.diff(peaks)
|
||||
has_periodic_echo = len(peak_spacing) > 2 and np.std(peak_spacing) < 0.1 * np.mean(peak_spacing)
|
||||
|
||||
# 2. Check for sudden amplitude changes
|
||||
amplitude_envelope = np.abs(audio_data)
|
||||
amplitude_changes = np.diff(amplitude_envelope)
|
||||
has_feedback_spikes = np.any(np.abs(amplitude_changes) > threshold * 2)
|
||||
|
||||
# 3. Check frequency distribution
|
||||
freq_magnitudes = np.abs(fft)[:len(fft)//2]
|
||||
peak_freqs = freqs[:len(fft)//2][np.argsort(freq_magnitudes)[-3:]]
|
||||
has_feedback_freqs = np.any((peak_freqs > 2000) & (peak_freqs < 4000))
|
||||
|
||||
# Combine all criteria
|
||||
is_valid_speech = (
|
||||
rms > threshold and
|
||||
speech_energy > threshold and
|
||||
not has_periodic_echo and
|
||||
not has_feedback_spikes and
|
||||
not has_feedback_freqs
|
||||
)
|
||||
|
||||
return is_valid_speech
|
||||
|
||||
def process_command(text):
|
||||
"""Process the transcribed command and execute appropriate action"""
|
||||
text = text.lower().strip()
|
||||
|
||||
# Skip if text is too short or contains numbers (likely noise)
|
||||
if len(text) < 5 or any(char.isdigit() for char in text):
|
||||
logger.debug("Text too short or contains numbers, skipping")
|
||||
return
|
||||
|
||||
# Enhanced noise pattern detection
|
||||
noise_patterns = ["lei", "los", "und", "aber", "nicht mehr", "das das", "und und"]
|
||||
for pattern in noise_patterns:
|
||||
if text.count(pattern) > 1: # More aggressive pattern filtering
|
||||
logger.debug(f"Detected noise pattern '{pattern}', skipping")
|
||||
return
|
||||
|
||||
# More aggressive repetition detection
|
||||
words = text.split()
|
||||
if len(words) >= 2:
|
||||
# Check for immediate word repetitions
|
||||
for i in range(len(words)-1):
|
||||
if words[i] == words[i+1]:
|
||||
logger.debug(f"Detected immediate word repetition: '{words[i]}', skipping")
|
||||
return
|
||||
|
||||
# Check for phrase repetitions
|
||||
phrases = [' '.join(words[i:i+2]) for i in range(len(words)-1)]
|
||||
phrase_counts = {}
|
||||
for phrase in phrases:
|
||||
phrase_counts[phrase] = phrase_counts.get(phrase, 0) + 1
|
||||
if phrase_counts[phrase] > MAX_REPETITIONS:
|
||||
logger.debug(f"Skipping due to excessive repetition: '{phrase}'")
|
||||
return
|
||||
|
||||
# German command mappings
|
||||
commands = {
|
||||
"ausschalten": "turn_off",
|
||||
"einschalten": "turn_on",
|
||||
"an": "turn_on",
|
||||
"aus": "turn_off"
|
||||
}
|
||||
|
||||
rooms = {
|
||||
"wohnzimmer": "living_room",
|
||||
"küche": "kitchen",
|
||||
"schlafzimmer": "bedroom",
|
||||
"bad": "bathroom"
|
||||
}
|
||||
|
||||
# Detect room
|
||||
detected_room = None
|
||||
for german_room, english_room in rooms.items():
|
||||
if german_room in text:
|
||||
detected_room = english_room
|
||||
break
|
||||
|
||||
# Detect command
|
||||
detected_command = None
|
||||
for german_cmd, english_cmd in commands.items():
|
||||
if german_cmd in text:
|
||||
detected_command = english_cmd
|
||||
break
|
||||
|
||||
if detected_room and detected_command:
|
||||
# Construct entity ID (assuming light)
|
||||
entity_id = f"light.{detected_room}"
|
||||
|
||||
# Send command to Home Assistant
|
||||
if send_command_to_hass("light", detected_command, entity_id):
|
||||
logger.info(f"Executed: {detected_command} for {entity_id}")
|
||||
else:
|
||||
logger.error("Failed to execute command")
|
||||
else:
|
||||
logger.debug(f"No command found in text: '{text}'")
|
||||
|
||||
class AudioProcessor:
|
||||
def __init__(self):
|
||||
# Initialize wake word detection model
|
||||
self.wake_word_model = Model(
|
||||
custom_model_paths=None, # Use default models
|
||||
inference_framework="onnx" # Use ONNX for better performance
|
||||
)
|
||||
|
||||
# Pre-load the wake word models
|
||||
for wake_word in WAKE_WORDS:
|
||||
self.wake_word_model.add_model(wake_word)
|
||||
|
||||
logger.info("Initializing AudioProcessor...")
|
||||
self.audio_buffer = queue.Queue()
|
||||
self.recording = False
|
||||
self.buffer = np.zeros(SAMPLE_RATE * BUFFER_DURATION)
|
||||
self.buffer_lock = threading.Lock()
|
||||
self.last_transcription_time = 0
|
||||
|
||||
def audio_callback(self, indata, frames, time, status):
|
||||
try:
|
||||
logger.info(f"Opening audio device: {AUDIO_DEVICE}")
|
||||
self.stream = sd.InputStream(
|
||||
device=AUDIO_DEVICE,
|
||||
samplerate=SAMPLE_RATE,
|
||||
channels=CHANNELS,
|
||||
dtype=np.int16,
|
||||
blocksize=CHUNK_SIZE,
|
||||
callback=self._audio_callback
|
||||
)
|
||||
logger.info("Audio stream initialized successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to initialize audio stream: {e}")
|
||||
raise
|
||||
|
||||
self.speech_detected = False
|
||||
self.silence_frames = 0
|
||||
self.speech_frames = 0
|
||||
|
||||
# Initialize wake word detection only if enabled
|
||||
if WAKE_WORD_ENABLED:
|
||||
try:
|
||||
logger.info("Initializing wake word model...")
|
||||
self.wake_word_model = Model(vad_threshold=0.5)
|
||||
self.last_prediction = None
|
||||
logger.info("Wake word model initialized successfully")
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to initialize wake word model: {e}")
|
||||
raise
|
||||
else:
|
||||
self.wake_word_model = None
|
||||
self.last_prediction = None
|
||||
logger.info("Wake word detection disabled")
|
||||
|
||||
def should_transcribe(self):
|
||||
"""Determine if we should transcribe based on mode and timing"""
|
||||
current_time = datetime.now().timestamp()
|
||||
if not WAKE_WORD_ENABLED:
|
||||
# Check if enough time has passed since last transcription
|
||||
time_since_last = current_time - self.last_transcription_time
|
||||
if time_since_last >= CONTINUOUS_TRANSCRIPTION_INTERVAL:
|
||||
# Only transcribe if we detect speech
|
||||
frames_per_chunk = CHUNK_SIZE
|
||||
min_speech_frames = int(MIN_SPEECH_DURATION * SAMPLE_RATE / frames_per_chunk)
|
||||
|
||||
if self.speech_frames >= min_speech_frames:
|
||||
self.last_transcription_time = current_time
|
||||
self.speech_frames = 0 # Reset counter
|
||||
return True
|
||||
return False
|
||||
|
||||
def _audio_callback(self, indata, frames, time, status):
|
||||
"""Callback for audio input"""
|
||||
if status:
|
||||
print(f"Audio callback status: {status}")
|
||||
logger.warning(f"Audio callback status: {status}")
|
||||
|
||||
# Convert to mono if necessary
|
||||
if CHANNELS > 1:
|
||||
@@ -55,25 +301,45 @@ class AudioProcessor:
|
||||
else:
|
||||
audio_data = indata.flatten()
|
||||
|
||||
# Check for speech
|
||||
if is_speech(audio_data):
|
||||
self.speech_frames += 1
|
||||
self.silence_frames = 0
|
||||
else:
|
||||
self.silence_frames += 1
|
||||
frames_per_chunk = CHUNK_SIZE
|
||||
silence_frames_threshold = int(SILENCE_DURATION * SAMPLE_RATE / frames_per_chunk)
|
||||
|
||||
if self.silence_frames >= silence_frames_threshold:
|
||||
self.speech_frames = 0
|
||||
|
||||
# Update circular buffer
|
||||
with self.buffer_lock:
|
||||
self.buffer = np.roll(self.buffer, -len(audio_data))
|
||||
self.buffer[-len(audio_data):] = audio_data
|
||||
|
||||
if WAKE_WORD_ENABLED:
|
||||
# Process for wake word detection
|
||||
prediction = self.wake_word_model.predict(audio_data)
|
||||
self.last_prediction = self.wake_word_model.predict(audio_data)
|
||||
|
||||
# Check if wake word detected
|
||||
for wake_word in WAKE_WORDS:
|
||||
if prediction[wake_word] > DETECTION_THRESHOLD:
|
||||
print(f"Wake word detected: {wake_word} (confidence: {prediction[wake_word]:.2f})")
|
||||
self.save_audio_segment(wake_word)
|
||||
confidence = self.last_prediction[wake_word]
|
||||
if confidence > DETECTION_THRESHOLD:
|
||||
logger.info(
|
||||
f"Wake word: {WAKE_WORD_ALIAS} (confidence: {confidence:.2f})"
|
||||
)
|
||||
self.process_audio()
|
||||
break
|
||||
else:
|
||||
# Continuous transcription mode
|
||||
if self.should_transcribe():
|
||||
self.process_audio()
|
||||
|
||||
def save_audio_segment(self, wake_word):
|
||||
"""Save the audio buffer when wake word is detected"""
|
||||
def process_audio(self):
|
||||
"""Process the current audio buffer (save and transcribe)"""
|
||||
timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
|
||||
filename = f"/audio/wake_word_{wake_word}_{timestamp}.wav"
|
||||
filename = f"/audio/audio_segment_{timestamp}.wav"
|
||||
|
||||
# Save the audio buffer to a WAV file
|
||||
with wave.open(filename, 'wb') as wf:
|
||||
@@ -85,89 +351,83 @@ class AudioProcessor:
|
||||
audio_data = (self.buffer * 32767).astype(np.int16)
|
||||
wf.writeframes(audio_data.tobytes())
|
||||
|
||||
print(f"Saved audio segment to {filename}")
|
||||
logger.info(f"Saved audio segment to {filename}")
|
||||
|
||||
# Transcribe the audio
|
||||
# Transcribe the audio with German language preference
|
||||
try:
|
||||
segments, info = asr_model.transcribe(
|
||||
filename,
|
||||
language="en",
|
||||
language="de", # Set German as preferred language
|
||||
beam_size=5,
|
||||
temperature=0
|
||||
)
|
||||
|
||||
# Format the transcription result
|
||||
result = {
|
||||
"text": " ".join(segment.text for segment in segments),
|
||||
"segments": [
|
||||
{
|
||||
"text": segment.text,
|
||||
"start": segment.start,
|
||||
"end": segment.end,
|
||||
"confidence": segment.confidence
|
||||
}
|
||||
for segment in segments
|
||||
]
|
||||
}
|
||||
# Get the full transcribed text
|
||||
transcribed_text = " ".join(segment.text for segment in segments)
|
||||
logger.info(f"Transcribed text: {transcribed_text}")
|
||||
|
||||
# Save metadata and transcription
|
||||
metadata = {
|
||||
"timestamp": timestamp,
|
||||
"wake_word": wake_word,
|
||||
"wake_word_confidence": float(prediction[wake_word]),
|
||||
"sample_rate": SAMPLE_RATE,
|
||||
"channels": CHANNELS,
|
||||
"duration": BUFFER_DURATION,
|
||||
"transcription": result
|
||||
}
|
||||
|
||||
with open(f"{filename}.json", 'w') as f:
|
||||
json.dump(metadata, f, indent=2)
|
||||
|
||||
print("\nTranscription result:")
|
||||
print(f"Text: {result['text']}")
|
||||
print("\nSegments:")
|
||||
for segment in result["segments"]:
|
||||
print(f"[{segment['start']:.2f}s - {segment['end']:.2f}s] ({segment['confidence']:.2%})")
|
||||
print(f'"{segment["text"]}"')
|
||||
# Process the command
|
||||
process_command(transcribed_text)
|
||||
|
||||
except Exception as e:
|
||||
print(f"Error during transcription: {e}")
|
||||
metadata = {
|
||||
"timestamp": timestamp,
|
||||
"wake_word": wake_word,
|
||||
"wake_word_confidence": float(prediction[wake_word]),
|
||||
"sample_rate": SAMPLE_RATE,
|
||||
"channels": CHANNELS,
|
||||
"duration": BUFFER_DURATION,
|
||||
"error": str(e)
|
||||
}
|
||||
with open(f"{filename}.json", 'w') as f:
|
||||
json.dump(metadata, f, indent=2)
|
||||
logger.error(f"Error during transcription or processing: {e}")
|
||||
|
||||
def start(self):
|
||||
"""Start audio processing"""
|
||||
try:
|
||||
print("Initializing wake word detection...")
|
||||
print(f"Loaded wake words: {', '.join(WAKE_WORDS)}")
|
||||
logger.info("Starting audio processor...")
|
||||
|
||||
# Log configuration
|
||||
logger.debug(f"Sample Rate: {SAMPLE_RATE}")
|
||||
logger.debug(f"Channels: {CHANNELS}")
|
||||
logger.debug(f"Chunk Size: {CHUNK_SIZE}")
|
||||
logger.debug(f"Buffer Duration: {BUFFER_DURATION}")
|
||||
logger.debug(f"Wake Word Enabled: {WAKE_WORD_ENABLED}")
|
||||
logger.debug(f"Speech Enabled: {SPEECH_ENABLED}")
|
||||
logger.debug(f"ASR Model: {os.environ.get('ASR_MODEL')}")
|
||||
|
||||
if WAKE_WORD_ENABLED:
|
||||
logger.info("Initializing wake word detection...")
|
||||
logger.info(f"Loaded wake words: {', '.join(WAKE_WORDS)}")
|
||||
else:
|
||||
logger.info("Starting continuous transcription mode...")
|
||||
interval = CONTINUOUS_TRANSCRIPTION_INTERVAL
|
||||
logger.info(f"Will transcribe every {interval} seconds")
|
||||
|
||||
try:
|
||||
logger.debug("Setting up audio input stream...")
|
||||
with sd.InputStream(
|
||||
channels=CHANNELS,
|
||||
samplerate=SAMPLE_RATE,
|
||||
blocksize=CHUNK_SIZE,
|
||||
callback=self.audio_callback
|
||||
callback=self._audio_callback
|
||||
):
|
||||
print("\nWake word detection started. Listening...")
|
||||
print("Press Ctrl+C to stop")
|
||||
logger.info("Audio input stream started successfully")
|
||||
logger.info("Listening for audio input...")
|
||||
logger.info("Press Ctrl+C to stop")
|
||||
|
||||
while True:
|
||||
sd.sleep(1000) # Sleep for 1 second
|
||||
|
||||
except KeyboardInterrupt:
|
||||
print("\nStopping wake word detection...")
|
||||
except sd.PortAudioError as e:
|
||||
logger.error(f"Error setting up audio stream: {e}")
|
||||
logger.error("Check if microphone is connected and accessible")
|
||||
raise
|
||||
except Exception as e:
|
||||
print(f"Error in audio processing: {e}")
|
||||
logger.error(f"Unexpected error in audio stream: {e}")
|
||||
raise
|
||||
|
||||
except KeyboardInterrupt:
|
||||
logger.info("\nStopping audio processing...")
|
||||
except Exception as e:
|
||||
logger.error("Critical error in audio processing", exc_info=True)
|
||||
raise
|
||||
|
||||
if __name__ == "__main__":
|
||||
try:
|
||||
logger.info("Initializing AudioProcessor...")
|
||||
processor = AudioProcessor()
|
||||
processor.start()
|
||||
except Exception as e:
|
||||
logger.error("Failed to start AudioProcessor", exc_info=True)
|
||||
raise
|
||||
23
docs/Gemfile
23
docs/Gemfile
@@ -1,23 +0,0 @@
|
||||
source "https://rubygems.org"
|
||||
|
||||
gem "github-pages", group: :jekyll_plugins
|
||||
gem "jekyll-theme-minimal"
|
||||
gem "jekyll-relative-links"
|
||||
gem "jekyll-seo-tag"
|
||||
gem "jekyll-remote-theme"
|
||||
gem "jekyll-github-metadata"
|
||||
gem "faraday-retry"
|
||||
|
||||
# Windows and JRuby does not include zoneinfo files, so bundle the tzinfo-data gem
|
||||
# and associated library.
|
||||
platforms :mingw, :x64_mingw, :mswin, :jruby do
|
||||
gem "tzinfo", ">= 1"
|
||||
gem "tzinfo-data"
|
||||
end
|
||||
|
||||
# Lock `http_parser.rb` gem to `v0.6.x` on JRuby builds since newer versions of the gem
|
||||
# do not have a Java counterpart.
|
||||
gem "http_parser.rb", "~> 0.6.0", :platforms => [:jruby]
|
||||
|
||||
# Add webrick for Ruby 3.0+
|
||||
gem "webrick", "~> 1.7"
|
||||
@@ -1,78 +0,0 @@
|
||||
title: Model Context Protocol (MCP)
|
||||
description: A bridge between Home Assistant and Language Learning Models
|
||||
theme: jekyll-theme-minimal
|
||||
markdown: kramdown
|
||||
|
||||
# Repository settings
|
||||
repository: jango-blockchained/advanced-homeassistant-mcp
|
||||
github: [metadata]
|
||||
|
||||
# Add base URL and URL settings
|
||||
baseurl: "/advanced-homeassistant-mcp" # the subpath of your site
|
||||
url: "https://jango-blockchained.github.io" # the base hostname & protocol
|
||||
|
||||
# Theme settings
|
||||
logo: /assets/img/logo.png # path to logo (create this if you want a logo)
|
||||
show_downloads: true # show download buttons for your repo
|
||||
|
||||
plugins:
|
||||
- jekyll-relative-links
|
||||
- jekyll-seo-tag
|
||||
- jekyll-remote-theme
|
||||
- jekyll-github-metadata
|
||||
|
||||
# Enable relative links
|
||||
relative_links:
|
||||
enabled: true
|
||||
collections: true
|
||||
|
||||
# Navigation structure
|
||||
header_pages:
|
||||
- index.md
|
||||
- getting-started.md
|
||||
- api.md
|
||||
- usage.md
|
||||
- tools/tools.md
|
||||
- development/development.md
|
||||
- troubleshooting.md
|
||||
- contributing.md
|
||||
- roadmap.md
|
||||
|
||||
# Collections
|
||||
collections:
|
||||
tools:
|
||||
output: true
|
||||
permalink: /:collection/:name
|
||||
development:
|
||||
output: true
|
||||
permalink: /:collection/:name
|
||||
|
||||
# Default layouts
|
||||
defaults:
|
||||
- scope:
|
||||
path: ""
|
||||
type: "pages"
|
||||
values:
|
||||
layout: "default"
|
||||
- scope:
|
||||
path: "tools"
|
||||
type: "tools"
|
||||
values:
|
||||
layout: "default"
|
||||
- scope:
|
||||
path: "development"
|
||||
type: "development"
|
||||
values:
|
||||
layout: "default"
|
||||
|
||||
# Exclude files from processing
|
||||
exclude:
|
||||
- Gemfile
|
||||
- Gemfile.lock
|
||||
- node_modules
|
||||
- vendor
|
||||
|
||||
# Sass settings
|
||||
sass:
|
||||
style: compressed
|
||||
sass_dir: _sass
|
||||
@@ -1,52 +0,0 @@
|
||||
<!DOCTYPE html>
|
||||
<html lang="{{ site.lang | default: " en-US" }}">
|
||||
|
||||
<head>
|
||||
<meta charset="UTF-8">
|
||||
<meta http-equiv="X-UA-Compatible" content="IE=edge">
|
||||
<meta name="viewport" content="width=device-width, initial-scale=1">
|
||||
{% seo %}
|
||||
<link rel="stylesheet" href="{{ " /assets/css/style.css?v=" | append: site.github.build_revision | relative_url }}">
|
||||
</head>
|
||||
|
||||
<body>
|
||||
<div class="wrapper">
|
||||
<header>
|
||||
<h1><a href="{{ " /" | absolute_url }}">{{ site.title | default: site.github.repository_name }}</a></h1>
|
||||
|
||||
{% if site.logo %}
|
||||
<img src="{{site.logo | relative_url}}" alt="Logo" />
|
||||
{% endif %}
|
||||
|
||||
<p>{{ site.description | default: site.github.project_tagline }}</p>
|
||||
|
||||
<p class="view"><a href="{{ site.github.repository_url }}">View the Project on GitHub <small>{{
|
||||
site.github.repository_nwo }}</small></a></p>
|
||||
|
||||
<nav class="main-nav">
|
||||
<h3>Documentation</h3>
|
||||
<ul>
|
||||
<li><a href="{{ '/getting-started' | relative_url }}">Getting Started</a></li>
|
||||
<li><a href="{{ '/api' | relative_url }}">API Reference</a></li>
|
||||
<li><a href="{{ '/sse-api' | relative_url }}">SSE API</a></li>
|
||||
<li><a href="{{ '/architecture' | relative_url }}">Architecture</a></li>
|
||||
<li><a href="{{ '/contributing' | relative_url }}">Contributing</a></li>
|
||||
<li><a href="{{ '/troubleshooting' | relative_url }}">Troubleshooting</a></li>
|
||||
</ul>
|
||||
</nav>
|
||||
</header>
|
||||
<section>
|
||||
{{ content }}
|
||||
</section>
|
||||
<footer>
|
||||
{% if site.github.is_project_page %}
|
||||
<p>This project is maintained by <a href="{{ site.github.owner_url }}">{{ site.github.owner_name }}</a></p>
|
||||
{% endif %}
|
||||
<p><small>Hosted on GitHub Pages — Theme by <a
|
||||
href="https://github.com/orderedlist">orderedlist</a></small></p>
|
||||
</footer>
|
||||
</div>
|
||||
<script src="{{ " /assets/js/scale.fix.js" | relative_url }}"></script>
|
||||
</body>
|
||||
|
||||
</html>
|
||||
728
docs/api.md
728
docs/api.md
@@ -1,728 +0,0 @@
|
||||
# 🚀 Home Assistant MCP API Documentation
|
||||
|
||||
 
|
||||
|
||||
## 🌟 Quick Start
|
||||
|
||||
```bash
|
||||
# Get API schema with caching
|
||||
curl -X GET http://localhost:3000/mcp \
|
||||
-H "Cache-Control: max-age=3600" # Cache for 1 hour
|
||||
```
|
||||
|
||||
## 🔌 Core Functions ⚙️
|
||||
|
||||
### State Management (`/api/state`)
|
||||
```http
|
||||
GET /api/state?cache=true # Enable client-side caching
|
||||
POST /api/state
|
||||
```
|
||||
|
||||
**Example Request:**
|
||||
```json
|
||||
{
|
||||
"context": "living_room",
|
||||
"state": {
|
||||
"lights": "on",
|
||||
"temperature": 22
|
||||
},
|
||||
"_cache": { // Optional caching config
|
||||
"ttl": 300, // 5 minutes
|
||||
"tags": ["lights", "climate"]
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## ⚡ Action Endpoints
|
||||
|
||||
### Execute Action with Cache Validation
|
||||
```http
|
||||
POST /api/action
|
||||
If-None-Match: "etag_value" // Prevent duplicate actions
|
||||
```
|
||||
|
||||
**Batch Processing:**
|
||||
```json
|
||||
{
|
||||
"actions": [
|
||||
{ "action": "🌞 Morning Routine", "params": { "brightness": 80 } },
|
||||
{ "action": "❄️ AC Control", "params": { "temp": 21 } }
|
||||
],
|
||||
"_parallel": true // Execute actions concurrently
|
||||
}
|
||||
```
|
||||
|
||||
## 🔍 Query Functions
|
||||
|
||||
### Available Actions with ETag
|
||||
```http
|
||||
GET /api/actions
|
||||
ETag: "a1b2c3d4" // Client-side cache validation
|
||||
```
|
||||
|
||||
**Response Headers:**
|
||||
```
|
||||
Cache-Control: public, max-age=86400 // 24-hour cache
|
||||
ETag: "a1b2c3d4"
|
||||
```
|
||||
|
||||
## 🌐 WebSocket Events
|
||||
|
||||
```javascript
|
||||
const ws = new WebSocket('wss://ha-mcp/ws');
|
||||
ws.onmessage = ({ data }) => {
|
||||
const event = JSON.parse(data);
|
||||
if(event.type === 'STATE_UPDATE') {
|
||||
updateUI(event.payload); // 🎨 Real-time UI sync
|
||||
}
|
||||
};
|
||||
```
|
||||
|
||||
## 🗃️ Caching Strategies
|
||||
|
||||
### Client-Side Caching
|
||||
```http
|
||||
GET /api/devices
|
||||
Cache-Control: max-age=300, stale-while-revalidate=60
|
||||
```
|
||||
|
||||
### Server-Side Cache-Control
|
||||
```typescript
|
||||
// Example middleware configuration
|
||||
app.use(
|
||||
cacheMiddleware({
|
||||
ttl: 60 * 5, // 5 minutes
|
||||
paths: ['/api/devices', '/mcp'],
|
||||
vary: ['Authorization'] // User-specific caching
|
||||
})
|
||||
);
|
||||
```
|
||||
|
||||
## ❌ Error Handling
|
||||
|
||||
**429 Too Many Requests:**
|
||||
```json
|
||||
{
|
||||
"error": {
|
||||
"code": "RATE_LIMITED",
|
||||
"message": "Slow down! 🐢",
|
||||
"retry_after": 30,
|
||||
"docs": "https://ha-mcp/docs/rate-limits"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## 🚦 Rate Limiting Tiers
|
||||
|
||||
| Tier | Requests/min | Features |
|
||||
|---------------|--------------|------------------------|
|
||||
| Guest | 10 | Basic read-only |
|
||||
| User | 100 | Full access |
|
||||
| Power User | 500 | Priority queue |
|
||||
| Integration | 1000 | Bulk operations |
|
||||
|
||||
## 🛠️ Example Usage
|
||||
|
||||
### Smart Cache Refresh
|
||||
```javascript
|
||||
async function getDevices() {
|
||||
const response = await fetch('/api/devices', {
|
||||
headers: {
|
||||
'If-None-Match': localStorage.getItem('devicesETag')
|
||||
}
|
||||
});
|
||||
|
||||
if(response.status === 304) { // Not Modified
|
||||
return JSON.parse(localStorage.devicesCache);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
localStorage.setItem('devicesETag', response.headers.get('ETag'));
|
||||
localStorage.setItem('devicesCache', JSON.stringify(data));
|
||||
return data;
|
||||
}
|
||||
```
|
||||
|
||||
## 🔒 Security Middleware (Enhanced)
|
||||
|
||||
### Cache-Aware Rate Limiting
|
||||
```typescript
|
||||
app.use(
|
||||
rateLimit({
|
||||
windowMs: 15 * 60 * 1000, // 15 minutes
|
||||
max: 100, // Limit each IP to 100 requests per window
|
||||
cache: new RedisStore(), // Distributed cache
|
||||
keyGenerator: (req) => {
|
||||
return `${req.ip}-${req.headers.authorization}`;
|
||||
}
|
||||
})
|
||||
);
|
||||
```
|
||||
|
||||
### Security Headers
|
||||
```http
|
||||
Content-Security-Policy: default-src 'self';
|
||||
Strict-Transport-Security: max-age=31536000;
|
||||
X-Content-Type-Options: nosniff;
|
||||
Cache-Control: public, max-age=600;
|
||||
ETag: "abc123"
|
||||
```
|
||||
|
||||
## 📘 Best Practices
|
||||
|
||||
1. **Cache Wisely:** Use `ETag` and `Cache-Control` headers for state data
|
||||
2. **Batch Operations:** Combine requests using `/api/actions/batch`
|
||||
3. **WebSocket First:** Prefer real-time updates over polling
|
||||
4. **Error Recovery:** Implement exponential backoff with jitter
|
||||
5. **Cache Invalidation:** Use tags for bulk invalidation
|
||||
|
||||
```mermaid
|
||||
graph LR
|
||||
A[Client] -->|Cached Request| B{CDN}
|
||||
B -->|Cache Hit| C[Return 304]
|
||||
B -->|Cache Miss| D[Origin Server]
|
||||
D -->|Response| B
|
||||
B -->|Response| A
|
||||
```
|
||||
|
||||
> Pro Tip: Use `curl -I` to inspect cache headers! 🔍
|
||||
|
||||
## Device Control
|
||||
|
||||
### Common Entity Controls
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "control",
|
||||
"command": "turn_on", // Options: "turn_on", "turn_off", "toggle"
|
||||
"entity_id": "light.living_room"
|
||||
}
|
||||
```
|
||||
|
||||
### Light Control
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "control",
|
||||
"command": "turn_on",
|
||||
"entity_id": "light.living_room",
|
||||
"brightness": 128,
|
||||
"color_temp": 4000,
|
||||
"rgb_color": [255, 0, 0]
|
||||
}
|
||||
```
|
||||
|
||||
## Add-on Management
|
||||
|
||||
### List Available Add-ons
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "addon",
|
||||
"action": "list"
|
||||
}
|
||||
```
|
||||
|
||||
### Install Add-on
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "addon",
|
||||
"action": "install",
|
||||
"slug": "core_configurator",
|
||||
"version": "5.6.0"
|
||||
}
|
||||
```
|
||||
|
||||
### Manage Add-on State
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "addon",
|
||||
"action": "start", // Options: "start", "stop", "restart"
|
||||
"slug": "core_configurator"
|
||||
}
|
||||
```
|
||||
|
||||
## Package Management
|
||||
|
||||
### List HACS Packages
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "package",
|
||||
"action": "list",
|
||||
"category": "integration" // Options: "integration", "plugin", "theme", "python_script", "appdaemon", "netdaemon"
|
||||
}
|
||||
```
|
||||
|
||||
### Install Package
|
||||
|
||||
```json
|
||||
{
|
||||
"tool": "package",
|
||||
"action": "install",
|
||||
"category": "integration",
|
||||
"repository": "hacs/integration",
|
||||
"version": "1.32.0"
|
||||
}
|
||||
```
|
||||
|
||||
## Automation Management
|
||||
|
||||
For automation management details and endpoints, please refer to the [Tools Documentation](tools/tools.md).
|
||||
|
||||
## Security Considerations
|
||||
|
||||
- Validate and sanitize all user inputs.
|
||||
- Enforce rate limiting to prevent abuse.
|
||||
- Apply proper security headers.
|
||||
- Gracefully handle errors based on the environment.
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
If you experience issues with the API:
|
||||
- Verify the endpoint and request payload.
|
||||
- Check authentication tokens and required headers.
|
||||
- Consult the [Troubleshooting Guide](troubleshooting.md) for further guidance.
|
||||
|
||||
## MCP Schema Endpoint
|
||||
|
||||
The server exposes an MCP (Model Context Protocol) schema endpoint that describes all available tools and their parameters:
|
||||
|
||||
```http
|
||||
GET /mcp
|
||||
```
|
||||
|
||||
This endpoint returns a JSON schema describing all available tools, their parameters, and documentation resources. The schema follows the MCP specification and can be used by LLM clients to understand the server's capabilities.
|
||||
|
||||
Example response:
|
||||
```json
|
||||
{
|
||||
"tools": [
|
||||
{
|
||||
"name": "list_devices",
|
||||
"description": "List all devices connected to Home Assistant",
|
||||
"parameters": {
|
||||
"type": "object",
|
||||
"properties": {
|
||||
"domain": {
|
||||
"type": "string",
|
||||
"enum": ["light", "climate", "alarm_control_panel", ...]
|
||||
},
|
||||
"area": { "type": "string" },
|
||||
"floor": { "type": "string" }
|
||||
}
|
||||
}
|
||||
},
|
||||
// ... other tools
|
||||
],
|
||||
"prompts": [],
|
||||
"resources": [
|
||||
{
|
||||
"name": "Home Assistant API",
|
||||
"url": "https://developers.home-assistant.io/docs/api/rest/"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
Note: The `/mcp` endpoint is publicly accessible and does not require authentication, as it only provides schema information.
|
||||
|
||||
## Core Functions
|
||||
|
||||
### State Management
|
||||
```http
|
||||
GET /api/state
|
||||
POST /api/state
|
||||
```
|
||||
|
||||
Manages the current state of the system.
|
||||
|
||||
**Example Request:**
|
||||
```json
|
||||
POST /api/state
|
||||
{
|
||||
"context": "living_room",
|
||||
"state": {
|
||||
"lights": "on",
|
||||
"temperature": 22
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Context Updates
|
||||
```http
|
||||
POST /api/context
|
||||
```
|
||||
|
||||
Updates the current context with new information.
|
||||
|
||||
**Example Request:**
|
||||
```json
|
||||
POST /api/context
|
||||
{
|
||||
"user": "john",
|
||||
"location": "kitchen",
|
||||
"time": "morning",
|
||||
"activity": "cooking"
|
||||
}
|
||||
```
|
||||
|
||||
## Action Endpoints
|
||||
|
||||
### Execute Action
|
||||
```http
|
||||
POST /api/action
|
||||
```
|
||||
|
||||
Executes a specified action with given parameters.
|
||||
|
||||
**Example Request:**
|
||||
```json
|
||||
POST /api/action
|
||||
{
|
||||
"action": "turn_on_lights",
|
||||
"parameters": {
|
||||
"room": "living_room",
|
||||
"brightness": 80
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
### Batch Actions
|
||||
```http
|
||||
POST /api/actions/batch
|
||||
```
|
||||
|
||||
Executes multiple actions in sequence.
|
||||
|
||||
**Example Request:**
|
||||
```json
|
||||
POST /api/actions/batch
|
||||
{
|
||||
"actions": [
|
||||
{
|
||||
"action": "turn_on_lights",
|
||||
"parameters": {
|
||||
"room": "living_room"
|
||||
}
|
||||
},
|
||||
{
|
||||
"action": "set_temperature",
|
||||
"parameters": {
|
||||
"temperature": 22
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
## Query Functions
|
||||
|
||||
### Get Available Actions
|
||||
```http
|
||||
GET /api/actions
|
||||
```
|
||||
|
||||
Returns a list of all available actions.
|
||||
|
||||
**Example Response:**
|
||||
```json
|
||||
{
|
||||
"actions": [
|
||||
{
|
||||
"name": "turn_on_lights",
|
||||
"parameters": ["room", "brightness"],
|
||||
"description": "Turns on lights in specified room"
|
||||
},
|
||||
{
|
||||
"name": "set_temperature",
|
||||
"parameters": ["temperature"],
|
||||
"description": "Sets temperature in current context"
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### Context Query
|
||||
```http
|
||||
GET /api/context?type=current
|
||||
```
|
||||
|
||||
Retrieves context information.
|
||||
|
||||
**Example Response:**
|
||||
```json
|
||||
{
|
||||
"current_context": {
|
||||
"user": "john",
|
||||
"location": "kitchen",
|
||||
"time": "morning",
|
||||
"activity": "cooking"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## WebSocket Events
|
||||
|
||||
The server supports real-time updates via WebSocket connections.
|
||||
|
||||
```javascript
|
||||
// Client-side connection example
|
||||
const ws = new WebSocket('ws://localhost:3000/ws');
|
||||
|
||||
ws.onmessage = (event) => {
|
||||
const data = JSON.parse(event.data);
|
||||
console.log('Received update:', data);
|
||||
};
|
||||
```
|
||||
|
||||
### Supported Events
|
||||
|
||||
- `state_change`: Emitted when system state changes
|
||||
- `context_update`: Emitted when context is updated
|
||||
- `action_executed`: Emitted when an action is completed
|
||||
- `error`: Emitted when an error occurs
|
||||
|
||||
**Example Event Data:**
|
||||
```json
|
||||
{
|
||||
"event": "state_change",
|
||||
"data": {
|
||||
"previous_state": {
|
||||
"lights": "off"
|
||||
},
|
||||
"current_state": {
|
||||
"lights": "on"
|
||||
},
|
||||
"timestamp": "2024-03-20T10:30:00Z"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Error Handling
|
||||
|
||||
All endpoints return standard HTTP status codes:
|
||||
|
||||
- 200: Success
|
||||
- 400: Bad Request
|
||||
- 401: Unauthorized
|
||||
- 403: Forbidden
|
||||
- 404: Not Found
|
||||
- 500: Internal Server Error
|
||||
|
||||
**Error Response Format:**
|
||||
```json
|
||||
{
|
||||
"error": {
|
||||
"code": "INVALID_PARAMETERS",
|
||||
"message": "Missing required parameter: room",
|
||||
"details": {
|
||||
"missing_fields": ["room"]
|
||||
}
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Rate Limiting
|
||||
|
||||
The API implements rate limiting to prevent abuse:
|
||||
|
||||
- 100 requests per minute per IP for regular endpoints
|
||||
- 1000 requests per minute per IP for WebSocket connections
|
||||
|
||||
When rate limit is exceeded, the server returns:
|
||||
|
||||
```json
|
||||
{
|
||||
"error": {
|
||||
"code": "RATE_LIMIT_EXCEEDED",
|
||||
"message": "Too many requests",
|
||||
"reset_time": "2024-03-20T10:31:00Z"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
## Example Usage
|
||||
|
||||
### Using curl
|
||||
```bash
|
||||
# Get current state
|
||||
curl -X GET \
|
||||
http://localhost:3000/api/state \
|
||||
-H 'Authorization: ApiKey your_api_key_here'
|
||||
|
||||
# Execute action
|
||||
curl -X POST \
|
||||
http://localhost:3000/api/action \
|
||||
-H 'Authorization: ApiKey your_api_key_here' \
|
||||
-H 'Content-Type: application/json' \
|
||||
-d '{
|
||||
"action": "turn_on_lights",
|
||||
"parameters": {
|
||||
"room": "living_room",
|
||||
"brightness": 80
|
||||
}
|
||||
}'
|
||||
```
|
||||
|
||||
### Using JavaScript
|
||||
```javascript
|
||||
// Execute action
|
||||
async function executeAction() {
|
||||
const response = await fetch('http://localhost:3000/api/action', {
|
||||
method: 'POST',
|
||||
headers: {
|
||||
'Authorization': 'ApiKey your_api_key_here',
|
||||
'Content-Type': 'application/json'
|
||||
},
|
||||
body: JSON.stringify({
|
||||
action: 'turn_on_lights',
|
||||
parameters: {
|
||||
room: 'living_room',
|
||||
brightness: 80
|
||||
}
|
||||
})
|
||||
});
|
||||
|
||||
const data = await response.json();
|
||||
console.log('Action result:', data);
|
||||
}
|
||||
```
|
||||
|
||||
## Security Middleware
|
||||
|
||||
### Overview
|
||||
|
||||
The security middleware provides a comprehensive set of utility functions to enhance the security of the Home Assistant MCP application. These functions cover various aspects of web security, including:
|
||||
|
||||
- Rate limiting
|
||||
- Request validation
|
||||
- Input sanitization
|
||||
- Security headers
|
||||
- Error handling
|
||||
|
||||
### Utility Functions
|
||||
|
||||
#### `checkRateLimit(ip: string, maxRequests?: number, windowMs?: number)`
|
||||
|
||||
Manages rate limiting for IP addresses to prevent abuse.
|
||||
|
||||
**Parameters**:
|
||||
- `ip`: IP address to track
|
||||
- `maxRequests`: Maximum number of requests allowed (default: 100)
|
||||
- `windowMs`: Time window for rate limiting (default: 15 minutes)
|
||||
|
||||
**Returns**: `boolean` or throws an error if limit is exceeded
|
||||
|
||||
**Example**:
|
||||
```typescript
|
||||
try {
|
||||
checkRateLimit('127.0.0.1'); // Checks rate limit with default settings
|
||||
} catch (error) {
|
||||
// Handle rate limit exceeded
|
||||
}
|
||||
```
|
||||
|
||||
#### `validateRequestHeaders(request: Request, requiredContentType?: string)`
|
||||
|
||||
Validates incoming HTTP request headers for security and compliance.
|
||||
|
||||
**Parameters**:
|
||||
- `request`: The incoming HTTP request
|
||||
- `requiredContentType`: Expected content type (default: 'application/json')
|
||||
|
||||
**Checks**:
|
||||
- Content type
|
||||
- Request body size
|
||||
- Authorization header (optional)
|
||||
|
||||
**Example**:
|
||||
```typescript
|
||||
try {
|
||||
validateRequestHeaders(request);
|
||||
} catch (error) {
|
||||
// Handle validation errors
|
||||
}
|
||||
```
|
||||
|
||||
#### `sanitizeValue(value: unknown)`
|
||||
|
||||
Sanitizes input values to prevent XSS attacks.
|
||||
|
||||
**Features**:
|
||||
- Escapes HTML tags
|
||||
- Handles nested objects and arrays
|
||||
- Preserves non-string values
|
||||
|
||||
**Example**:
|
||||
```typescript
|
||||
const sanitized = sanitizeValue('<script>alert("xss")</script>');
|
||||
// Returns: '<script>alert("xss")</script>'
|
||||
```
|
||||
|
||||
#### `applySecurityHeaders(request: Request, helmetConfig?: HelmetOptions)`
|
||||
|
||||
Applies security headers to HTTP requests using Helmet.
|
||||
|
||||
**Security Headers**:
|
||||
- Content Security Policy
|
||||
- X-Frame-Options
|
||||
- X-Content-Type-Options
|
||||
- Referrer Policy
|
||||
- HSTS (in production)
|
||||
|
||||
**Example**:
|
||||
```typescript
|
||||
const headers = applySecurityHeaders(request);
|
||||
```
|
||||
|
||||
#### `handleError(error: Error, env?: string)`
|
||||
|
||||
Handles error responses with environment-specific details.
|
||||
|
||||
**Modes**:
|
||||
- Production: Generic error message
|
||||
- Development: Detailed error with stack trace
|
||||
|
||||
**Example**:
|
||||
```typescript
|
||||
const errorResponse = handleError(error, process.env.NODE_ENV);
|
||||
```
|
||||
|
||||
### Middleware Usage
|
||||
|
||||
These utility functions are integrated into Elysia middleware:
|
||||
|
||||
```typescript
|
||||
const app = new Elysia()
|
||||
.use(rateLimiter) // Rate limiting
|
||||
.use(validateRequest) // Request validation
|
||||
.use(sanitizeInput) // Input sanitization
|
||||
.use(securityHeaders) // Security headers
|
||||
.use(errorHandler) // Error handling
|
||||
```
|
||||
|
||||
### Best Practices
|
||||
|
||||
1. Always validate and sanitize user inputs
|
||||
2. Use rate limiting to prevent abuse
|
||||
3. Apply security headers
|
||||
4. Handle errors gracefully
|
||||
5. Keep environment-specific error handling
|
||||
|
||||
### Security Considerations
|
||||
|
||||
- Configurable rate limits
|
||||
- XSS protection
|
||||
- Content security policies
|
||||
- Token validation
|
||||
- Error information exposure control
|
||||
|
||||
### Troubleshooting
|
||||
|
||||
- Ensure `JWT_SECRET` is set in environment
|
||||
- Check content type in requests
|
||||
- Monitor rate limit errors
|
||||
- Review error handling in different environments
|
||||
@@ -1,68 +0,0 @@
|
||||
# Architecture Documentation for MCP Server
|
||||
|
||||
## Overview
|
||||
|
||||
The MCP Server is designed as a high-performance, secure, and scalable bridge between Home Assistant and Language Learning Models (LLMs). This document outlines the architectural design principles, core components, and deployment strategies that power the MCP Server.
|
||||
|
||||
## Key Architectural Components
|
||||
|
||||
### High-Performance Runtime with Bun
|
||||
|
||||
- **Fast Startup & Efficiency:** Powered by Bun, the MCP Server benefits from rapid startup times, efficient memory utilization, and native TypeScript support.
|
||||
- **Optimized Build Process:** Bun's build tools allow for quick iteration and deployment, ensuring minimal downtime and swift performance enhancement.
|
||||
|
||||
### Real-time Communication using Server-Sent Events (SSE)
|
||||
|
||||
- **Continuous Updates:** The server leverages SSE to deliver real-time notifications and updates, ensuring that any changes in Home Assistant are immediately communicated to connected clients.
|
||||
- **Scalable Connection Handling:** SSE provides an event-driven model that efficiently manages multiple simultaneous client connections.
|
||||
|
||||
### Modular & Extensible Design
|
||||
|
||||
- **Plugin Architecture:** Designed with modularity in mind, the MCP Server supports plugins, add-ons, and custom automation scripts, enabling seamless feature expansion without disrupting core functionality.
|
||||
- **Separation of Concerns:** Different components, such as device management, automation control, and system monitoring, are clearly separated, allowing independent development, testing, and scaling.
|
||||
|
||||
### Secure API Integration
|
||||
|
||||
- **Token-Based Authentication:** Robust token-based authentication mechanisms restrict access to authorized users and systems.
|
||||
- **Rate Limiting & Error Handling:** Integrated rate limiting combined with comprehensive error handling ensures system stability and prevents misuse.
|
||||
- **Best Practices:** All API endpoints follow industry-standard security guidelines to protect data and maintain system integrity.
|
||||
|
||||
### Deployment & Scalability
|
||||
|
||||
- **Containerized Deployment with Docker:** The use of Docker Compose enables straightforward deployment, management, and scaling of the server and its dependencies.
|
||||
- **Flexible Environment Configuration:** Environment variables and configuration files (.env) facilitate smooth transitions between development, testing, and production setups.
|
||||
|
||||
## Future Enhancements
|
||||
|
||||
- **Advanced Automation Logic:** Integration of more complex automation rules and conditional decision-making capabilities.
|
||||
- **Enhanced Security Measures:** Additional layers of security, such as multi-factor authentication and improved encryption techniques, are on the roadmap.
|
||||
- **Improved Monitoring & Analytics:** Future updates will introduce advanced performance metrics and real-time analytics to monitor system health and user interactions.
|
||||
|
||||
## Conclusion
|
||||
|
||||
The architecture of the MCP Server prioritizes performance, scalability, and security. By leveraging Bun's high-performance runtime, employing real-time communication through SSE, and maintaining a modular, secure design, the MCP Server provides a robust platform for integrating Home Assistant with modern LLM functionalities.
|
||||
|
||||
*This document is a living document and will be updated as the system evolves.*
|
||||
|
||||
## Key Components
|
||||
|
||||
- **API Module:** Handles RESTful endpoints, authentication, and error management.
|
||||
- **SSE Module:** Provides real-time updates through Server-Sent Events.
|
||||
- **Tools Module:** Offers various utilities for device control, automation, and data processing.
|
||||
- **Security Module:** Implements token-based authentication and secure communications.
|
||||
- **Integration Module:** Bridges data between Home Assistant and external systems.
|
||||
|
||||
## Data Flow
|
||||
|
||||
1. Requests enter via the API endpoints.
|
||||
2. Security middleware validates and processes requests.
|
||||
3. Core modules process data and execute the necessary business logic.
|
||||
4. Real-time notifications are managed by the SSE module.
|
||||
|
||||
## Future Enhancements
|
||||
|
||||
- Expand modularity with potential microservices.
|
||||
- Enhance security with multi-factor authentication.
|
||||
- Improve scalability through distributed architectures.
|
||||
|
||||
*Further diagrams and detailed breakdowns will be added in future updates.*
|
||||
@@ -1,54 +0,0 @@
|
||||
@import "{{ site.theme }}";
|
||||
|
||||
// Custom styles
|
||||
.main-nav {
|
||||
margin-top: 20px;
|
||||
|
||||
ul {
|
||||
list-style: none;
|
||||
padding: 0;
|
||||
margin: 0;
|
||||
}
|
||||
|
||||
li {
|
||||
margin-bottom: 8px;
|
||||
}
|
||||
|
||||
a {
|
||||
color: #267CB9;
|
||||
text-decoration: none;
|
||||
|
||||
&:hover {
|
||||
text-decoration: underline;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
h1,
|
||||
h2,
|
||||
h3 {
|
||||
color: #333;
|
||||
}
|
||||
|
||||
code {
|
||||
background-color: #f8f8f8;
|
||||
border: 1px solid #ddd;
|
||||
border-radius: 3px;
|
||||
padding: 2px 5px;
|
||||
}
|
||||
|
||||
pre {
|
||||
background-color: #f8f8f8;
|
||||
border: 1px solid #ddd;
|
||||
border-radius: 3px;
|
||||
padding: 10px;
|
||||
overflow-x: auto;
|
||||
}
|
||||
|
||||
.wrapper {
|
||||
max-width: 960px;
|
||||
}
|
||||
|
||||
section {
|
||||
max-width: 700px;
|
||||
}
|
||||
@@ -1,36 +0,0 @@
|
||||
# Contributing to Home Assistant MCP
|
||||
|
||||
We welcome community contributions to improve the MCP Server. Please review the following guidelines before contributing.
|
||||
|
||||
## How to Contribute
|
||||
|
||||
1. **Fork the Repository:** Create your personal fork on GitHub.
|
||||
2. **Create a Feature Branch:** Use a clear name (e.g., `feature/your-feature` or `bugfix/short-description`).
|
||||
3. **Make Changes:** Develop your feature or fix bugs while following our coding standards.
|
||||
4. **Write Tests:** Include tests for new features or bug fixes.
|
||||
5. **Submit a Pull Request:** Once your changes are complete, submit a PR for review.
|
||||
6. **Address Feedback:** Revise your PR based on maintainers' suggestions.
|
||||
|
||||
## Code Style Guidelines
|
||||
|
||||
- Follow the project's established coding style.
|
||||
- Use Bun tooling for linting and formatting:
|
||||
- `bun run lint`
|
||||
- `bun run format`
|
||||
|
||||
## Documentation
|
||||
|
||||
- Update documentation alongside your code changes.
|
||||
- Ensure tests pass and coverage remains high.
|
||||
|
||||
## Reporting Issues
|
||||
|
||||
- Use the GitHub Issues page to report bugs, request new features, or ask questions.
|
||||
- Provide clear descriptions, replication steps, and any error logs.
|
||||
|
||||
## Community
|
||||
|
||||
- Join our real-time discussions on our chat platforms (Discord, Slack, etc.).
|
||||
- Engage with other contributors to exchange ideas and solutions.
|
||||
|
||||
Thank you for helping improve the Home Assistant MCP project!
|
||||
@@ -1,190 +0,0 @@
|
||||
# Development Guide
|
||||
|
||||
This guide provides information for developers who want to contribute to or extend the Home Assistant MCP.
|
||||
|
||||
## Project Structure
|
||||
|
||||
```
|
||||
homeassistant-mcp/
|
||||
├── src/
|
||||
│ ├── __tests__/ # Test files
|
||||
│ ├── __mocks__/ # Mock files
|
||||
│ ├── api/ # API endpoints and route handlers
|
||||
│ ├── config/ # Configuration management
|
||||
│ ├── hass/ # Home Assistant integration
|
||||
│ ├── interfaces/ # TypeScript interfaces
|
||||
│ ├── mcp/ # MCP core functionality
|
||||
│ ├── middleware/ # Express middleware
|
||||
│ ├── routes/ # Route definitions
|
||||
│ ├── security/ # Security utilities
|
||||
│ ├── sse/ # Server-Sent Events handling
|
||||
│ ├── tools/ # Tool implementations
|
||||
│ ├── types/ # TypeScript type definitions
|
||||
│ └── utils/ # Utility functions
|
||||
├── __tests__/ # Test files
|
||||
├── docs/ # Documentation
|
||||
├── dist/ # Compiled JavaScript
|
||||
└── scripts/ # Build and utility scripts
|
||||
```
|
||||
|
||||
## Development Setup
|
||||
|
||||
1. Install dependencies:
|
||||
```bash
|
||||
npm install
|
||||
```
|
||||
|
||||
2. Set up development environment:
|
||||
```bash
|
||||
cp .env.example .env.development
|
||||
```
|
||||
|
||||
3. Start development server:
|
||||
```bash
|
||||
npm run dev
|
||||
```
|
||||
|
||||
## Code Style
|
||||
|
||||
We follow these coding standards:
|
||||
|
||||
1. TypeScript best practices
|
||||
- Use strict type checking
|
||||
- Avoid `any` types
|
||||
- Document complex types
|
||||
|
||||
2. ESLint rules
|
||||
- Run `npm run lint` to check
|
||||
- Run `npm run lint:fix` to auto-fix
|
||||
|
||||
3. Code formatting
|
||||
- Use Prettier
|
||||
- Run `npm run format` to format code
|
||||
|
||||
## Testing
|
||||
|
||||
1. Unit tests:
|
||||
```bash
|
||||
npm run test
|
||||
```
|
||||
|
||||
2. Integration tests:
|
||||
```bash
|
||||
npm run test:integration
|
||||
```
|
||||
|
||||
3. Coverage report:
|
||||
```bash
|
||||
npm run test:coverage
|
||||
```
|
||||
|
||||
## Creating New Tools
|
||||
|
||||
1. Create a new file in `src/tools/`:
|
||||
```typescript
|
||||
import { z } from 'zod';
|
||||
import { Tool } from '../types';
|
||||
|
||||
export const myTool: Tool = {
|
||||
name: 'my_tool',
|
||||
description: 'Description of my tool',
|
||||
parameters: z.object({
|
||||
// Define parameters
|
||||
}),
|
||||
execute: async (params) => {
|
||||
// Implement tool logic
|
||||
}
|
||||
};
|
||||
```
|
||||
|
||||
2. Add to `src/tools/index.ts`
|
||||
3. Create tests in `__tests__/tools/`
|
||||
4. Add documentation in `docs/tools/`
|
||||
|
||||
## Contributing
|
||||
|
||||
1. Fork the repository
|
||||
2. Create a feature branch
|
||||
3. Make your changes
|
||||
4. Write/update tests
|
||||
5. Update documentation
|
||||
6. Submit a pull request
|
||||
|
||||
### Pull Request Process
|
||||
|
||||
1. Ensure all tests pass
|
||||
2. Update documentation
|
||||
3. Update CHANGELOG.md
|
||||
4. Get review from maintainers
|
||||
|
||||
## Building
|
||||
|
||||
1. Development build:
|
||||
```bash
|
||||
npm run build:dev
|
||||
```
|
||||
|
||||
2. Production build:
|
||||
```bash
|
||||
npm run build
|
||||
```
|
||||
|
||||
## Documentation
|
||||
|
||||
1. Update documentation for changes
|
||||
2. Follow documentation structure
|
||||
3. Include examples
|
||||
4. Update type definitions
|
||||
|
||||
## Debugging
|
||||
|
||||
1. Development debugging:
|
||||
```bash
|
||||
npm run dev:debug
|
||||
```
|
||||
|
||||
2. Test debugging:
|
||||
```bash
|
||||
npm run test:debug
|
||||
```
|
||||
|
||||
3. VSCode launch configurations provided
|
||||
|
||||
## Performance
|
||||
|
||||
1. Follow performance best practices
|
||||
2. Use caching where appropriate
|
||||
3. Implement rate limiting
|
||||
4. Monitor memory usage
|
||||
|
||||
## Security
|
||||
|
||||
1. Follow security best practices
|
||||
2. Validate all inputs
|
||||
3. Use proper authentication
|
||||
4. Handle errors securely
|
||||
|
||||
## Deployment
|
||||
|
||||
1. Build for production:
|
||||
```bash
|
||||
npm run build
|
||||
```
|
||||
|
||||
2. Start production server:
|
||||
```bash
|
||||
npm start
|
||||
```
|
||||
|
||||
3. Docker deployment:
|
||||
```bash
|
||||
docker-compose up -d
|
||||
```
|
||||
|
||||
## Support
|
||||
|
||||
Need development help?
|
||||
1. Check documentation
|
||||
2. Search issues
|
||||
3. Create new issue
|
||||
4. Join discussions
|
||||
@@ -1,30 +0,0 @@
|
||||
# Getting Started
|
||||
|
||||
Begin your journey with the Home Assistant MCP Server by following these steps:
|
||||
|
||||
- **API Documentation:** Read the [API Documentation](api.md) for available endpoints.
|
||||
- **Real-Time Updates:** Learn about [Server-Sent Events](sse-api.md) for live communication.
|
||||
- **Tools:** Explore available [Tools](tools/tools.md) for device control and automation.
|
||||
- **Configuration:** Refer to the [Configuration Guide](configuration.md) for setup and advanced settings.
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
If you encounter any issues:
|
||||
1. Verify that your Home Assistant instance is accessible.
|
||||
2. Ensure that all required environment variables are properly set.
|
||||
3. Consult the [Troubleshooting Guide](troubleshooting.md) for additional solutions.
|
||||
|
||||
## Development
|
||||
|
||||
For contributors:
|
||||
1. Fork the repository.
|
||||
2. Create a feature branch.
|
||||
3. Follow the [Development Guide](development/development.md) for contribution guidelines.
|
||||
4. Submit a pull request with your enhancements.
|
||||
|
||||
## Support
|
||||
|
||||
Need help?
|
||||
- Visit our [GitHub Issues](https://github.com/jango-blockchained/homeassistant-mcp/issues).
|
||||
- Review the [Troubleshooting Guide](troubleshooting.md).
|
||||
- Check the [FAQ](troubleshooting.md#faq) for common questions.
|
||||
@@ -1,5 +0,0 @@
|
||||
# Configuration
|
||||
|
||||
## Basic Configuration
|
||||
|
||||
## Advanced Settings
|
||||
@@ -1,124 +0,0 @@
|
||||
# Installation Guide
|
||||
|
||||
## Prerequisites
|
||||
|
||||
### System Requirements
|
||||
- **Operating System:** Linux, macOS, or Windows (Docker recommended)
|
||||
- **Runtime:** Bun v1.0.26 or higher
|
||||
- **Home Assistant:** v2023.11 or higher
|
||||
- **Minimum Hardware:**
|
||||
- 2 CPU cores
|
||||
- 2GB RAM
|
||||
- 10GB free disk space
|
||||
|
||||
### Software Dependencies
|
||||
- Bun runtime
|
||||
- Docker (optional, recommended for deployment)
|
||||
- Git
|
||||
- Node.js (for some development tasks)
|
||||
|
||||
## Installation Methods
|
||||
|
||||
### 1. Basic Setup
|
||||
|
||||
#### Install Bun
|
||||
```bash
|
||||
curl -fsSL https://bun.sh/install | bash
|
||||
```
|
||||
|
||||
#### Clone Repository
|
||||
```bash
|
||||
git clone https://github.com/jango-blockchained/homeassistant-mcp.git
|
||||
cd homeassistant-mcp
|
||||
```
|
||||
|
||||
#### Install Dependencies
|
||||
```bash
|
||||
bun install
|
||||
```
|
||||
|
||||
#### Configure Environment
|
||||
1. Copy environment template
|
||||
```bash
|
||||
cp .env.example .env
|
||||
```
|
||||
2. Edit `.env` file with your Home Assistant configuration
|
||||
- Set `HASS_HOST`
|
||||
- Configure authentication tokens
|
||||
- Adjust other settings as needed
|
||||
|
||||
#### Build and Start
|
||||
```bash
|
||||
bun run build
|
||||
bun start
|
||||
```
|
||||
|
||||
### 2. Docker Setup (Recommended)
|
||||
|
||||
#### Prerequisites
|
||||
- Docker
|
||||
- Docker Compose
|
||||
|
||||
#### Deployment Steps
|
||||
```bash
|
||||
# Clone repository
|
||||
git clone https://github.com/jango-blockchained/homeassistant-mcp.git
|
||||
cd homeassistant-mcp
|
||||
|
||||
# Configure environment
|
||||
cp .env.example .env
|
||||
# Edit .env file with your settings
|
||||
|
||||
# Deploy with Docker Compose
|
||||
docker compose up -d
|
||||
```
|
||||
|
||||
### 3. Home Assistant Add-on (Coming Soon)
|
||||
We're working on a direct Home Assistant add-on for even easier installation.
|
||||
|
||||
## Verification
|
||||
|
||||
### Check Installation
|
||||
- Web Interface: [http://localhost:3000](http://localhost:3000)
|
||||
- Logs: `docker compose logs` or check `logs/` directory
|
||||
|
||||
### Troubleshooting
|
||||
- Ensure all environment variables are correctly set
|
||||
- Check network connectivity to Home Assistant
|
||||
- Verify authentication tokens
|
||||
|
||||
## Updating
|
||||
|
||||
### Basic Setup
|
||||
```bash
|
||||
git pull
|
||||
bun install
|
||||
bun run build
|
||||
bun start
|
||||
```
|
||||
|
||||
### Docker
|
||||
```bash
|
||||
git pull
|
||||
docker compose up -d --build
|
||||
```
|
||||
|
||||
## Uninstallation
|
||||
|
||||
### Basic Setup
|
||||
```bash
|
||||
cd homeassistant-mcp
|
||||
bun stop # Stop the application
|
||||
rm -rf node_modules dist
|
||||
```
|
||||
|
||||
### Docker
|
||||
```bash
|
||||
docker compose down
|
||||
docker rmi homeassistant-mcp # Remove image
|
||||
```
|
||||
|
||||
## Next Steps
|
||||
- [Configuration Guide](configuration.md)
|
||||
- [Usage Instructions](../usage.md)
|
||||
- [Troubleshooting](../troubleshooting.md)
|
||||
110
docs/index.md
110
docs/index.md
@@ -1,110 +0,0 @@
|
||||
---
|
||||
layout: default
|
||||
title: Home
|
||||
nav_order: 1
|
||||
---
|
||||
|
||||
# 📚 Home Assistant MCP Documentation
|
||||
|
||||
Welcome to the documentation for the Home Assistant MCP (Model Context Protocol) Server.
|
||||
|
||||
## 📑 Documentation Index
|
||||
|
||||
- [Getting Started Guide](getting-started.md)
|
||||
- [API Documentation](api.md)
|
||||
- [Troubleshooting](troubleshooting.md)
|
||||
- [Contributing Guide](contributing.md)
|
||||
|
||||
For project overview, installation, and general information, please see our [main README](../README.md).
|
||||
|
||||
## 🔗 Quick Links
|
||||
|
||||
- [GitHub Repository](https://github.com/jango-blockchained/homeassistant-mcp)
|
||||
- [Issue Tracker](https://github.com/jango-blockchained/homeassistant-mcp/issues)
|
||||
- [GitHub Discussions](https://github.com/jango-blockchained/homeassistant-mcp/discussions)
|
||||
|
||||
## 📝 License
|
||||
|
||||
This project is licensed under the MIT License. See [LICENSE](../LICENSE) for details.
|
||||
|
||||
# Model Context Protocol (MCP) Server
|
||||
|
||||
## Overview
|
||||
|
||||
The Model Context Protocol (MCP) Server is a cutting-edge bridge between Home Assistant and Language Learning Models (LLMs), designed to revolutionize smart home automation and control. This documentation provides comprehensive information about setting up, configuring, and using the Home Assistant MCP.
|
||||
|
||||
## Key Features
|
||||
|
||||
### 🏠 Smart Home Integration
|
||||
- Natural language control of smart devices
|
||||
- Real-time device state monitoring
|
||||
- Advanced automation capabilities
|
||||
|
||||
### 🤖 LLM Powered Interactions
|
||||
- Intuitive voice and text-based commands
|
||||
- Context-aware device management
|
||||
- Intelligent automation suggestions
|
||||
|
||||
### 🔒 Security & Performance
|
||||
- Token-based authentication
|
||||
- High-performance Bun runtime
|
||||
- Secure, real-time communication protocols
|
||||
|
||||
## Documentation
|
||||
|
||||
### Core Documentation
|
||||
1. [Getting Started](getting-started.md)
|
||||
- Installation and basic setup
|
||||
- Configuration
|
||||
- First Steps
|
||||
|
||||
2. [API Reference](api.md)
|
||||
- REST API Endpoints
|
||||
- Authentication
|
||||
- Error Handling
|
||||
|
||||
3. [SSE API](sse-api.md)
|
||||
- Event Subscriptions
|
||||
- Real-time Updates
|
||||
- Connection Management
|
||||
|
||||
### Advanced Topics
|
||||
4. [Architecture](architecture.md)
|
||||
- System Design
|
||||
- Components
|
||||
- Data Flow
|
||||
|
||||
5. [Configuration](getting-started.md#configuration)
|
||||
- Environment Variables
|
||||
- Security Settings
|
||||
- Performance Tuning
|
||||
|
||||
6. [Development Guide](development/development.md)
|
||||
- Project Structure
|
||||
- Contributing Guidelines
|
||||
- Testing
|
||||
|
||||
7. [Troubleshooting](troubleshooting.md)
|
||||
- Common Issues
|
||||
- Debugging
|
||||
- FAQ
|
||||
|
||||
## Quick Links
|
||||
|
||||
- [GitHub Repository](https://github.com/jango-blockchained/homeassistant-mcp)
|
||||
- [Issue Tracker](https://github.com/jango-blockchained/homeassistant-mcp/issues)
|
||||
- [Contributing Guide](contributing.md)
|
||||
- [Roadmap](roadmap.md)
|
||||
|
||||
## Community and Support
|
||||
|
||||
If you need help or have questions:
|
||||
|
||||
1. Check the [Troubleshooting Guide](troubleshooting.md)
|
||||
2. Search existing [Issues](https://github.com/jango-blockchained/homeassistant-mcp/issues)
|
||||
3. Join our [GitHub Discussions](https://github.com/jango-blockchained/homeassistant-mcp/discussions)
|
||||
4. Create a new issue if your problem isn't already reported
|
||||
|
||||
## License
|
||||
|
||||
This project is licensed under the MIT License. See [LICENSE](https://github.com/jango-blockchained/homeassistant-mcp/blob/main/LICENSE) for details.
|
||||
@@ -1,51 +0,0 @@
|
||||
# Roadmap for MCP Server
|
||||
|
||||
The following roadmap outlines our planned enhancements and future directions for the Home Assistant MCP Server. This document is a living guide that will be updated as new features are planned and developed.
|
||||
|
||||
## Near-Term Goals
|
||||
|
||||
- **Advanced Automation Capabilities:**
|
||||
- Integrate sophisticated automation rules with conditional logic and multi-step execution.
|
||||
- Introduce a visual automation builder for simplified rule creation.
|
||||
|
||||
- **Enhanced Security Features:**
|
||||
- Implement multi-factor authentication for critical actions.
|
||||
- Strengthen encryption methods and data handling practices.
|
||||
- Expand monitoring and alerting for potential security breaches.
|
||||
|
||||
- **Performance Optimizations:**
|
||||
- Refine resource utilization to reduce latency.
|
||||
- Optimize real-time data streaming via SSE.
|
||||
- Introduce advanced caching mechanisms for frequently requested data.
|
||||
|
||||
## Mid-Term Goals
|
||||
|
||||
- **User Interface Improvements:**
|
||||
- Develop an intuitive web-based dashboard for device management and monitoring.
|
||||
- Provide real-time analytics and performance metrics.
|
||||
|
||||
- **Expanded Integrations:**
|
||||
- Support a broader range of smart home devices and brands.
|
||||
- Integrate with additional home automation platforms and third-party services.
|
||||
|
||||
- **Developer Experience Enhancements:**
|
||||
- Improve documentation and developer tooling.
|
||||
- Streamline contribution guidelines and testing setups.
|
||||
|
||||
## Long-Term Vision
|
||||
|
||||
- **Ecosystem Expansion:**
|
||||
- Build a modular plugin system for community-driven extensions and integrations.
|
||||
- Enable seamless integration with future technologies in smart home and AI domains.
|
||||
|
||||
- **Scalability and Resilience:**
|
||||
- Architect the system to support large-scale deployments.
|
||||
- Incorporate advanced load balancing and failover mechanisms.
|
||||
|
||||
## How to Follow the Roadmap
|
||||
|
||||
- **Community Involvement:** We welcome and encourage feedback.
|
||||
- **Regular Updates:** This document is updated regularly with new goals and milestones.
|
||||
- **Transparency:** Check our GitHub repository and issue tracker for ongoing discussions.
|
||||
|
||||
*This roadmap is intended as a guide and may evolve based on community needs, technological advancements, and strategic priorities.*
|
||||
364
docs/sse-api.md
364
docs/sse-api.md
@@ -1,364 +0,0 @@
|
||||
# Home Assistant MCP Server-Sent Events (SSE) API Documentation
|
||||
|
||||
## Overview
|
||||
|
||||
The SSE API provides real-time updates from Home Assistant through a persistent connection. This allows clients to receive instant notifications about state changes, events, and other activities without polling.
|
||||
|
||||
## Quick Reference
|
||||
|
||||
### Available Endpoints
|
||||
|
||||
| Endpoint | Method | Description | Authentication |
|
||||
|----------|---------|-------------|----------------|
|
||||
| `/subscribe_events` | POST | Subscribe to real-time events and state changes | Required |
|
||||
| `/get_sse_stats` | POST | Get statistics about current SSE connections | Required |
|
||||
|
||||
### Event Types Available
|
||||
|
||||
| Event Type | Description | Example Subscription |
|
||||
|------------|-------------|---------------------|
|
||||
| `state_changed` | Entity state changes | `events=state_changed` |
|
||||
| `service_called` | Service call events | `events=service_called` |
|
||||
| `automation_triggered` | Automation trigger events | `events=automation_triggered` |
|
||||
| `script_executed` | Script execution events | `events=script_executed` |
|
||||
| `ping` | Connection keepalive (system) | Automatic |
|
||||
| `error` | Error notifications (system) | Automatic |
|
||||
|
||||
### Subscription Options
|
||||
|
||||
| Option | Description | Example |
|
||||
|--------|-------------|---------|
|
||||
| `entity_id` | Subscribe to specific entity | `entity_id=light.living_room` |
|
||||
| `domain` | Subscribe to entire domain | `domain=light` |
|
||||
| `events` | Subscribe to event types | `events=state_changed,automation_triggered` |
|
||||
|
||||
## Authentication
|
||||
|
||||
All SSE connections require authentication using your Home Assistant token.
|
||||
|
||||
```javascript
|
||||
const token = 'YOUR_HASS_TOKEN';
|
||||
```
|
||||
|
||||
## Endpoints
|
||||
|
||||
### Subscribe to Events
|
||||
|
||||
`POST /subscribe_events`
|
||||
|
||||
Subscribe to Home Assistant events and state changes.
|
||||
|
||||
#### Parameters
|
||||
|
||||
| Parameter | Type | Required | Description |
|
||||
|------------|----------|----------|-------------|
|
||||
| token | string | Yes | Your Home Assistant authentication token |
|
||||
| events | string[] | No | Array of event types to subscribe to |
|
||||
| entity_id | string | No | Specific entity ID to monitor |
|
||||
| domain | string | No | Domain to monitor (e.g., "light", "switch") |
|
||||
|
||||
#### Example Request
|
||||
|
||||
```javascript
|
||||
const eventSource = new EventSource(`http://localhost:3000/subscribe_events?token=${token}&entity_id=light.living_room&domain=switch&events=state_changed,automation_triggered`);
|
||||
|
||||
eventSource.onmessage = (event) => {
|
||||
const data = JSON.parse(event.data);
|
||||
console.log('Received:', data);
|
||||
};
|
||||
|
||||
eventSource.onerror = (error) => {
|
||||
console.error('SSE Error:', error);
|
||||
eventSource.close();
|
||||
};
|
||||
```
|
||||
|
||||
### Get SSE Statistics
|
||||
|
||||
`POST /get_sse_stats`
|
||||
|
||||
Get current statistics about SSE connections and subscriptions.
|
||||
|
||||
#### Parameters
|
||||
|
||||
| Parameter | Type | Required | Description |
|
||||
|-----------|--------|----------|-------------|
|
||||
| token | string | Yes | Your Home Assistant authentication token |
|
||||
|
||||
#### Example Request
|
||||
|
||||
```bash
|
||||
curl -X POST http://localhost:3000/get_sse_stats \
|
||||
-H "Content-Type: application/json" \
|
||||
-d '{"token": "YOUR_HASS_TOKEN"}'
|
||||
```
|
||||
|
||||
## Event Types
|
||||
|
||||
### Standard Events
|
||||
|
||||
1. **connection**
|
||||
- Sent when a client connects successfully
|
||||
```json
|
||||
{
|
||||
"type": "connection",
|
||||
"status": "connected",
|
||||
"id": "client_uuid",
|
||||
"authenticated": true,
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
2. **state_changed**
|
||||
- Sent when an entity's state changes
|
||||
```json
|
||||
{
|
||||
"type": "state_changed",
|
||||
"data": {
|
||||
"entity_id": "light.living_room",
|
||||
"state": "on",
|
||||
"attributes": {
|
||||
"brightness": 255,
|
||||
"color_temp": 370
|
||||
},
|
||||
"last_changed": "2024-02-10T12:00:00.000Z",
|
||||
"last_updated": "2024-02-10T12:00:00.000Z"
|
||||
},
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
3. **service_called**
|
||||
- Sent when a Home Assistant service is called
|
||||
```json
|
||||
{
|
||||
"type": "service_called",
|
||||
"data": {
|
||||
"domain": "light",
|
||||
"service": "turn_on",
|
||||
"service_data": {
|
||||
"entity_id": "light.living_room",
|
||||
"brightness": 255
|
||||
}
|
||||
},
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
4. **automation_triggered**
|
||||
- Sent when an automation is triggered
|
||||
```json
|
||||
{
|
||||
"type": "automation_triggered",
|
||||
"data": {
|
||||
"automation_id": "automation.morning_routine",
|
||||
"trigger": {
|
||||
"platform": "time",
|
||||
"at": "07:00:00"
|
||||
}
|
||||
},
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
5. **script_executed**
|
||||
- Sent when a script is executed
|
||||
```json
|
||||
{
|
||||
"type": "script_executed",
|
||||
"data": {
|
||||
"script_id": "script.welcome_home",
|
||||
"execution_data": {
|
||||
"status": "completed"
|
||||
}
|
||||
},
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
### System Events
|
||||
|
||||
1. **ping**
|
||||
- Sent every 30 seconds to keep the connection alive
|
||||
```json
|
||||
{
|
||||
"type": "ping",
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
2. **error**
|
||||
- Sent when an error occurs
|
||||
```json
|
||||
{
|
||||
"type": "error",
|
||||
"error": "rate_limit_exceeded",
|
||||
"message": "Too many requests, please try again later",
|
||||
"timestamp": "2024-02-10T12:00:00.000Z"
|
||||
}
|
||||
```
|
||||
|
||||
## Rate Limiting
|
||||
|
||||
- Maximum 1000 requests per minute per client
|
||||
- Rate limits are reset every minute
|
||||
- Exceeding the rate limit will result in an error event
|
||||
|
||||
## Connection Management
|
||||
|
||||
- Maximum 100 concurrent clients
|
||||
- Connections timeout after 5 minutes of inactivity
|
||||
- Ping messages are sent every 30 seconds
|
||||
- Clients should handle reconnection on connection loss
|
||||
|
||||
## Example Implementation
|
||||
|
||||
```javascript
|
||||
class HomeAssistantSSE {
|
||||
constructor(baseUrl, token) {
|
||||
this.baseUrl = baseUrl;
|
||||
this.token = token;
|
||||
this.eventSource = null;
|
||||
this.reconnectAttempts = 0;
|
||||
this.maxReconnectAttempts = 5;
|
||||
this.reconnectDelay = 1000;
|
||||
}
|
||||
|
||||
connect(options = {}) {
|
||||
const params = new URLSearchParams({
|
||||
token: this.token,
|
||||
...(options.events && { events: options.events.join(',') }),
|
||||
...(options.entity_id && { entity_id: options.entity_id }),
|
||||
...(options.domain && { domain: options.domain })
|
||||
});
|
||||
|
||||
this.eventSource = new EventSource(`${this.baseUrl}/subscribe_events?${params}`);
|
||||
|
||||
this.eventSource.onmessage = (event) => {
|
||||
const data = JSON.parse(event.data);
|
||||
this.handleEvent(data);
|
||||
};
|
||||
|
||||
this.eventSource.onerror = (error) => {
|
||||
console.error('SSE Error:', error);
|
||||
this.handleError(error);
|
||||
};
|
||||
}
|
||||
|
||||
handleEvent(data) {
|
||||
switch (data.type) {
|
||||
case 'connection':
|
||||
this.reconnectAttempts = 0;
|
||||
console.log('Connected:', data);
|
||||
break;
|
||||
case 'ping':
|
||||
// Connection is alive
|
||||
break;
|
||||
case 'error':
|
||||
console.error('Server Error:', data);
|
||||
break;
|
||||
default:
|
||||
// Handle other event types
|
||||
console.log('Event:', data);
|
||||
}
|
||||
}
|
||||
|
||||
handleError(error) {
|
||||
this.eventSource?.close();
|
||||
|
||||
if (this.reconnectAttempts < this.maxReconnectAttempts) {
|
||||
this.reconnectAttempts++;
|
||||
const delay = this.reconnectDelay * Math.pow(2, this.reconnectAttempts - 1);
|
||||
console.log(`Reconnecting in ${delay}ms (attempt ${this.reconnectAttempts})`);
|
||||
setTimeout(() => this.connect(), delay);
|
||||
} else {
|
||||
console.error('Max reconnection attempts reached');
|
||||
}
|
||||
}
|
||||
|
||||
disconnect() {
|
||||
this.eventSource?.close();
|
||||
this.eventSource = null;
|
||||
}
|
||||
}
|
||||
|
||||
// Usage example
|
||||
const client = new HomeAssistantSSE('http://localhost:3000', 'YOUR_HASS_TOKEN');
|
||||
client.connect({
|
||||
events: ['state_changed', 'automation_triggered'],
|
||||
domain: 'light'
|
||||
});
|
||||
```
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Error Handling**
|
||||
- Implement exponential backoff for reconnection attempts
|
||||
- Handle connection timeouts gracefully
|
||||
- Monitor for rate limit errors
|
||||
|
||||
2. **Resource Management**
|
||||
- Close EventSource when no longer needed
|
||||
- Limit subscriptions to necessary events/entities
|
||||
- Handle cleanup on page unload
|
||||
|
||||
3. **Security**
|
||||
- Never expose the authentication token in client-side code
|
||||
- Use HTTPS in production
|
||||
- Validate all incoming data
|
||||
|
||||
4. **Performance**
|
||||
- Subscribe only to needed events
|
||||
- Implement client-side event filtering
|
||||
- Monitor memory usage for long-running connections
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
### Common Issues
|
||||
|
||||
1. **Connection Failures**
|
||||
- Verify your authentication token is valid
|
||||
- Check server URL is accessible
|
||||
- Ensure proper network connectivity
|
||||
- Verify SSL/TLS configuration if using HTTPS
|
||||
|
||||
2. **Missing Events**
|
||||
- Confirm subscription parameters are correct
|
||||
- Check rate limiting status
|
||||
- Verify entity/domain exists
|
||||
- Monitor client-side event handlers
|
||||
|
||||
3. **Performance Issues**
|
||||
- Reduce number of subscriptions
|
||||
- Implement client-side filtering
|
||||
- Monitor memory usage
|
||||
- Check network latency
|
||||
|
||||
### Debugging Tips
|
||||
|
||||
1. Enable console logging:
|
||||
```javascript
|
||||
const client = new HomeAssistantSSE('http://localhost:3000', 'YOUR_HASS_TOKEN');
|
||||
client.debug = true; // Enables detailed logging
|
||||
```
|
||||
|
||||
2. Monitor network traffic:
|
||||
```javascript
|
||||
// Add event listeners for connection states
|
||||
eventSource.addEventListener('open', () => {
|
||||
console.log('Connection opened');
|
||||
});
|
||||
|
||||
eventSource.addEventListener('error', (e) => {
|
||||
console.log('Connection error:', e);
|
||||
});
|
||||
```
|
||||
|
||||
3. Track subscription status:
|
||||
```javascript
|
||||
// Get current subscriptions
|
||||
const stats = await fetch('/get_sse_stats', {
|
||||
headers: { 'Authorization': `Bearer ${token}` }
|
||||
}).then(r => r.json());
|
||||
|
||||
console.log('Current subscriptions:', stats);
|
||||
```
|
||||
422
docs/testing.md
422
docs/testing.md
@@ -1,422 +0,0 @@
|
||||
# Testing Documentation
|
||||
|
||||
## Quick Reference
|
||||
|
||||
```bash
|
||||
# Most Common Commands
|
||||
bun test # Run all tests
|
||||
bun test --watch # Run tests in watch mode
|
||||
bun test --coverage # Run tests with coverage
|
||||
bun test path/to/test.ts # Run a specific test file
|
||||
|
||||
# Additional Options
|
||||
DEBUG=true bun test # Run with debug output
|
||||
bun test --pattern "auth" # Run tests matching a pattern
|
||||
bun test --timeout 60000 # Run with a custom timeout
|
||||
```
|
||||
|
||||
## Overview
|
||||
|
||||
This document describes the testing setup and practices used in the Home Assistant MCP project. We use Bun's test runner for both unit and integration testing, ensuring comprehensive coverage across modules.
|
||||
|
||||
## Test Structure
|
||||
|
||||
Tests are organized in two main locations:
|
||||
|
||||
1. **Root Level Integration Tests** (`/__tests__/`):
|
||||
|
||||
```
|
||||
__tests__/
|
||||
├── ai/ # AI/ML component tests
|
||||
├── api/ # API integration tests
|
||||
├── context/ # Context management tests
|
||||
├── hass/ # Home Assistant integration tests
|
||||
├── schemas/ # Schema validation tests
|
||||
├── security/ # Security integration tests
|
||||
├── tools/ # Tools and utilities tests
|
||||
├── websocket/ # WebSocket integration tests
|
||||
├── helpers.test.ts # Helper function tests
|
||||
├── index.test.ts # Main application tests
|
||||
└── server.test.ts # Server integration tests
|
||||
```
|
||||
|
||||
2. **Component Level Unit Tests** (`src/**/`):
|
||||
|
||||
```
|
||||
src/
|
||||
├── __tests__/ # Global test setup and utilities
|
||||
│ └── setup.ts # Global test configuration
|
||||
├── component/
|
||||
│ ├── __tests__/ # Component-specific unit tests
|
||||
│ └── component.ts
|
||||
```
|
||||
|
||||
## Test Configuration
|
||||
|
||||
### Bun Test Configuration (`bunfig.toml`)
|
||||
|
||||
```toml
|
||||
[test]
|
||||
preload = ["./src/__tests__/setup.ts"] # Global test setup
|
||||
coverage = true # Enable coverage by default
|
||||
timeout = 30000 # Test timeout in milliseconds
|
||||
testMatch = ["**/__tests__/**/*.test.ts"] # Test file patterns
|
||||
```
|
||||
|
||||
### Bun Scripts
|
||||
|
||||
Available test commands in `package.json`:
|
||||
|
||||
```bash
|
||||
# Run all tests
|
||||
bun test
|
||||
|
||||
# Watch mode for development
|
||||
bun test --watch
|
||||
|
||||
# Generate coverage report
|
||||
bun test --coverage
|
||||
|
||||
# Run linting
|
||||
bun run lint
|
||||
|
||||
# Format code
|
||||
bun run format
|
||||
```
|
||||
|
||||
## Test Setup
|
||||
|
||||
### Global Configuration
|
||||
|
||||
A global test setup file (`src/__tests__/setup.ts`) provides:
|
||||
- Environment configuration
|
||||
- Mock utilities
|
||||
- Test helper functions
|
||||
- Global lifecycle hooks
|
||||
|
||||
### Test Environment
|
||||
|
||||
- Environment variables are loaded from `.env.test`.
|
||||
- Console output is minimized unless `DEBUG=true`.
|
||||
- JWT secrets and tokens are preconfigured for testing.
|
||||
- Rate limiting and security features are initialized appropriately.
|
||||
|
||||
## Running Tests
|
||||
|
||||
```bash
|
||||
# Basic test run
|
||||
bun test
|
||||
|
||||
# Run tests with coverage
|
||||
bun test --coverage
|
||||
|
||||
# Run a specific test file
|
||||
bun test path/to/test.test.ts
|
||||
|
||||
# Run tests in watch mode
|
||||
bun test --watch
|
||||
|
||||
# Run tests with debug output
|
||||
DEBUG=true bun test
|
||||
|
||||
# Run tests with increased timeout
|
||||
bun test --timeout 60000
|
||||
|
||||
# Run tests matching a pattern
|
||||
bun test --pattern "auth"
|
||||
```
|
||||
|
||||
## Advanced Debugging
|
||||
|
||||
### Using Node Inspector
|
||||
|
||||
```bash
|
||||
# Start tests with inspector
|
||||
bun test --inspect
|
||||
|
||||
# Start tests with inspector and break on first line
|
||||
bun test --inspect-brk
|
||||
```
|
||||
|
||||
### Using VS Code
|
||||
|
||||
Create a launch configuration in `.vscode/launch.json`:
|
||||
|
||||
```json
|
||||
{
|
||||
"version": "0.2.0",
|
||||
"configurations": [
|
||||
{
|
||||
"type": "bun",
|
||||
"request": "launch",
|
||||
"name": "Debug Tests",
|
||||
"program": "${workspaceFolder}/node_modules/bun/bin/bun",
|
||||
"args": ["test", "${file}"],
|
||||
"cwd": "${workspaceFolder}",
|
||||
"env": { "DEBUG": "true" }
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
### Test Isolation
|
||||
|
||||
To run a single test in isolation:
|
||||
|
||||
```typescript
|
||||
describe.only("specific test suite", () => {
|
||||
it.only("specific test case", () => {
|
||||
// Only this test will run
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Writing Tests
|
||||
|
||||
### Test File Naming
|
||||
|
||||
- Place test files in a `__tests__` directory adjacent to the code being tested.
|
||||
- Name files with the pattern `*.test.ts`.
|
||||
- Mirror the structure of the source code in your test organization.
|
||||
|
||||
### Example Test Structure
|
||||
|
||||
```typescript
|
||||
describe("Security Features", () => {
|
||||
it("should validate tokens correctly", () => {
|
||||
const payload = { userId: "123", role: "user" };
|
||||
const token = jwt.sign(payload, validSecret, { expiresIn: "1h" });
|
||||
const result = TokenManager.validateToken(token, testIp);
|
||||
expect(result.valid).toBe(true);
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Coverage
|
||||
|
||||
The project maintains strict coverage:
|
||||
- Overall coverage: at least 80%
|
||||
- Critical paths: 90%+
|
||||
- New features: ≥85% coverage
|
||||
|
||||
Generate a coverage report with:
|
||||
|
||||
```bash
|
||||
bun test --coverage
|
||||
```
|
||||
|
||||
## Security Middleware Testing
|
||||
|
||||
### Utility Function Testing
|
||||
|
||||
The security middleware now uses a utility-first approach, which allows for more granular and comprehensive testing. Each security function is now independently testable, improving code reliability and maintainability.
|
||||
|
||||
#### Key Utility Functions
|
||||
|
||||
1. **Rate Limiting (`checkRateLimit`)**
|
||||
- Tests multiple scenarios:
|
||||
- Requests under threshold
|
||||
- Requests exceeding threshold
|
||||
- Rate limit reset after window expiration
|
||||
|
||||
```typescript
|
||||
// Example test
|
||||
it('should throw when requests exceed threshold', () => {
|
||||
const ip = '127.0.0.2';
|
||||
for (let i = 0; i < 11; i++) {
|
||||
if (i < 10) {
|
||||
expect(() => checkRateLimit(ip, 10)).not.toThrow();
|
||||
} else {
|
||||
expect(() => checkRateLimit(ip, 10)).toThrow('Too many requests from this IP');
|
||||
}
|
||||
}
|
||||
});
|
||||
```
|
||||
|
||||
2. **Request Validation (`validateRequestHeaders`)**
|
||||
- Tests content type validation
|
||||
- Checks request size limits
|
||||
- Validates authorization headers
|
||||
|
||||
```typescript
|
||||
it('should reject invalid content type', () => {
|
||||
const mockRequest = new Request('http://localhost', {
|
||||
method: 'POST',
|
||||
headers: { 'content-type': 'text/plain' }
|
||||
});
|
||||
expect(() => validateRequestHeaders(mockRequest)).toThrow('Content-Type must be application/json');
|
||||
});
|
||||
```
|
||||
|
||||
3. **Input Sanitization (`sanitizeValue`)**
|
||||
- Sanitizes HTML tags
|
||||
- Handles nested objects
|
||||
- Preserves non-string values
|
||||
|
||||
```typescript
|
||||
it('should sanitize HTML tags', () => {
|
||||
const input = '<script>alert("xss")</script>Hello';
|
||||
const sanitized = sanitizeValue(input);
|
||||
expect(sanitized).toBe('<script>alert("xss")</script>Hello');
|
||||
});
|
||||
```
|
||||
|
||||
4. **Security Headers (`applySecurityHeaders`)**
|
||||
- Verifies correct security header application
|
||||
- Checks CSP, frame options, and other security headers
|
||||
|
||||
```typescript
|
||||
it('should apply security headers', () => {
|
||||
const mockRequest = new Request('http://localhost');
|
||||
const headers = applySecurityHeaders(mockRequest);
|
||||
expect(headers['content-security-policy']).toBeDefined();
|
||||
expect(headers['x-frame-options']).toBeDefined();
|
||||
});
|
||||
```
|
||||
|
||||
5. **Error Handling (`handleError`)**
|
||||
- Tests error responses in production and development modes
|
||||
- Verifies error message and stack trace inclusion
|
||||
|
||||
```typescript
|
||||
it('should include error details in development mode', () => {
|
||||
const error = new Error('Test error');
|
||||
const result = handleError(error, 'development');
|
||||
expect(result).toEqual({
|
||||
error: true,
|
||||
message: 'Internal server error',
|
||||
error: 'Test error',
|
||||
stack: expect.any(String)
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
### Testing Philosophy
|
||||
|
||||
- **Isolation**: Each utility function is tested independently
|
||||
- **Comprehensive Coverage**: Multiple scenarios for each function
|
||||
- **Predictable Behavior**: Clear expectations for input and output
|
||||
- **Error Handling**: Robust testing of error conditions
|
||||
|
||||
### Best Practices
|
||||
|
||||
1. Use minimal, focused test cases
|
||||
2. Test both successful and failure scenarios
|
||||
3. Verify input sanitization and security measures
|
||||
4. Mock external dependencies when necessary
|
||||
|
||||
### Running Security Tests
|
||||
|
||||
```bash
|
||||
# Run all tests
|
||||
bun test
|
||||
|
||||
# Run specific security tests
|
||||
bun test __tests__/security/
|
||||
```
|
||||
|
||||
### Continuous Improvement
|
||||
|
||||
- Regularly update test cases
|
||||
- Add new test scenarios as security requirements evolve
|
||||
- Perform periodic security audits
|
||||
|
||||
## Best Practices
|
||||
|
||||
1. **Isolation**: Each test should be independent and not rely on the state of other tests.
|
||||
2. **Mocking**: Use the provided mock utilities for external dependencies.
|
||||
3. **Cleanup**: Clean up any resources or state modifications in `afterEach` or `afterAll` hooks.
|
||||
4. **Descriptive Names**: Use clear, descriptive test names that explain the expected behavior.
|
||||
5. **Assertions**: Make specific, meaningful assertions rather than general ones.
|
||||
6. **Setup**: Use `beforeEach` for common test setup to avoid repetition.
|
||||
7. **Error Cases**: Test both success and error cases for complete coverage.
|
||||
|
||||
## Coverage
|
||||
|
||||
The project aims for high test coverage, particularly focusing on:
|
||||
- Security-critical code paths
|
||||
- API endpoints
|
||||
- Data validation
|
||||
- Error handling
|
||||
- Event broadcasting
|
||||
|
||||
Run coverage reports using:
|
||||
```bash
|
||||
bun test --coverage
|
||||
```
|
||||
|
||||
## Debugging Tests
|
||||
|
||||
To debug tests:
|
||||
1. Set `DEBUG=true` to enable console output during tests
|
||||
2. Use the `--watch` flag for development
|
||||
3. Add `console.log()` statements (they're only shown when DEBUG is true)
|
||||
4. Use the test utilities' debugging helpers
|
||||
|
||||
### Advanced Debugging
|
||||
|
||||
1. **Using Node Inspector**:
|
||||
```bash
|
||||
# Start tests with inspector
|
||||
bun test --inspect
|
||||
|
||||
# Start tests with inspector and break on first line
|
||||
bun test --inspect-brk
|
||||
```
|
||||
|
||||
2. **Using VS Code**:
|
||||
```jsonc
|
||||
// .vscode/launch.json
|
||||
{
|
||||
"version": "0.2.0",
|
||||
"configurations": [
|
||||
{
|
||||
"type": "bun",
|
||||
"request": "launch",
|
||||
"name": "Debug Tests",
|
||||
"program": "${workspaceFolder}/node_modules/bun/bin/bun",
|
||||
"args": ["test", "${file}"],
|
||||
"cwd": "${workspaceFolder}",
|
||||
"env": { "DEBUG": "true" }
|
||||
}
|
||||
]
|
||||
}
|
||||
```
|
||||
|
||||
3. **Test Isolation**:
|
||||
To run a single test in isolation:
|
||||
```typescript
|
||||
describe.only("specific test suite", () => {
|
||||
it.only("specific test case", () => {
|
||||
// Only this test will run
|
||||
});
|
||||
});
|
||||
```
|
||||
|
||||
## Contributing
|
||||
|
||||
When contributing new code:
|
||||
1. Add tests for new features
|
||||
2. Ensure existing tests pass
|
||||
3. Maintain or improve coverage
|
||||
4. Follow the existing test patterns and naming conventions
|
||||
5. Document any new test utilities or patterns
|
||||
|
||||
## Coverage Requirements
|
||||
|
||||
The project maintains strict coverage requirements:
|
||||
|
||||
- Minimum overall coverage: 80%
|
||||
- Critical paths (security, API, data validation): 90%
|
||||
- New features must include tests with >= 85% coverage
|
||||
|
||||
Coverage reports are generated in multiple formats:
|
||||
- Console summary
|
||||
- HTML report (./coverage/index.html)
|
||||
- LCOV report (./coverage/lcov.info)
|
||||
|
||||
To view detailed coverage:
|
||||
```bash
|
||||
# Generate and open coverage report
|
||||
bun test --coverage && open coverage/index.html
|
||||
```
|
||||
@@ -1,127 +0,0 @@
|
||||
# Home Assistant MCP Tools
|
||||
|
||||
This section documents all available tools in the Home Assistant MCP.
|
||||
|
||||
## Available Tools
|
||||
|
||||
### Device Management
|
||||
|
||||
1. [List Devices](./list-devices.md)
|
||||
- List all available Home Assistant devices
|
||||
- Group devices by domain
|
||||
- Get device states and attributes
|
||||
|
||||
2. [Device Control](./control.md)
|
||||
- Control various device types
|
||||
- Support for lights, switches, covers, climate devices
|
||||
- Domain-specific commands and parameters
|
||||
|
||||
### History and State
|
||||
|
||||
1. [History](./history.md)
|
||||
- Fetch device state history
|
||||
- Filter by time range
|
||||
- Get significant changes
|
||||
|
||||
2. [Scene Management](./scene.md)
|
||||
- List available scenes
|
||||
- Activate scenes
|
||||
- Scene state information
|
||||
|
||||
### Automation
|
||||
|
||||
1. [Automation Management](./automation.md)
|
||||
- List automations
|
||||
- Toggle automation state
|
||||
- Trigger automations manually
|
||||
|
||||
2. [Automation Configuration](./automation-config.md)
|
||||
- Create new automations
|
||||
- Update existing automations
|
||||
- Delete automations
|
||||
- Duplicate automations
|
||||
|
||||
### Add-ons and Packages
|
||||
|
||||
1. [Add-on Management](./addon.md)
|
||||
- List available add-ons
|
||||
- Install/uninstall add-ons
|
||||
- Start/stop/restart add-ons
|
||||
- Get add-on information
|
||||
|
||||
2. [Package Management](./package.md)
|
||||
- Manage HACS packages
|
||||
- Install/update/remove packages
|
||||
- List available packages by category
|
||||
|
||||
### Notifications
|
||||
|
||||
1. [Notify](./notify.md)
|
||||
- Send notifications
|
||||
- Support for multiple notification services
|
||||
- Custom notification data
|
||||
|
||||
### Real-time Events
|
||||
|
||||
1. [Event Subscription](./subscribe-events.md)
|
||||
- Subscribe to Home Assistant events
|
||||
- Monitor specific entities
|
||||
- Domain-based monitoring
|
||||
|
||||
2. [SSE Statistics](./sse-stats.md)
|
||||
- Get SSE connection statistics
|
||||
- Monitor active subscriptions
|
||||
- Connection management
|
||||
|
||||
## Using Tools
|
||||
|
||||
All tools can be accessed through:
|
||||
|
||||
1. REST API endpoints
|
||||
2. WebSocket connections
|
||||
3. Server-Sent Events (SSE)
|
||||
|
||||
### Authentication
|
||||
|
||||
Tools require authentication using:
|
||||
- Home Assistant Long-Lived Access Token
|
||||
- JWT tokens for specific operations
|
||||
|
||||
### Error Handling
|
||||
|
||||
All tools follow a consistent error handling pattern:
|
||||
```typescript
|
||||
{
|
||||
success: boolean;
|
||||
message?: string;
|
||||
data?: any;
|
||||
}
|
||||
```
|
||||
|
||||
### Rate Limiting
|
||||
|
||||
Tools are subject to rate limiting:
|
||||
- Default: 100 requests per 15 minutes
|
||||
- Configurable through environment variables
|
||||
|
||||
## Tool Development
|
||||
|
||||
Want to create a new tool? Check out:
|
||||
- [Tool Development Guide](../development/tools.md)
|
||||
- [Tool Interface Documentation](../development/interfaces.md)
|
||||
- [Best Practices](../development/best-practices.md)
|
||||
|
||||
## Examples
|
||||
|
||||
Each tool documentation includes:
|
||||
- Usage examples
|
||||
- Code snippets
|
||||
- Common use cases
|
||||
- Troubleshooting tips
|
||||
|
||||
## Support
|
||||
|
||||
Need help with tools?
|
||||
- Check individual tool documentation
|
||||
- See [Troubleshooting Guide](../troubleshooting.md)
|
||||
- Create an issue on GitHub
|
||||
@@ -1,345 +0,0 @@
|
||||
# Troubleshooting Guide
|
||||
|
||||
This guide provides solutions to common issues encountered with the Home Assistant MCP Server.
|
||||
|
||||
## Common Issues
|
||||
|
||||
- **Server Not Starting:**
|
||||
- Verify that all required environment variables are correctly set.
|
||||
- Check for port conflicts or missing dependencies.
|
||||
- Review the server logs for error details.
|
||||
|
||||
- **Connection Problems:**
|
||||
- Ensure your Home Assistant instance is reachable.
|
||||
- Confirm that the authentication token is valid.
|
||||
- Check network configurations and firewalls.
|
||||
|
||||
## Tool Issues
|
||||
|
||||
### Tool Not Found
|
||||
|
||||
**Symptoms:**
|
||||
- "Tool not found" errors or 404 responses.
|
||||
|
||||
**Solutions:**
|
||||
- Double-check the tool name spelling.
|
||||
- Verify that the tool is correctly registered.
|
||||
- Review tool imports and documentation.
|
||||
|
||||
### Tool Execution Failures
|
||||
|
||||
**Symptoms:**
|
||||
- Execution errors or timeouts.
|
||||
|
||||
**Solutions:**
|
||||
- Validate input parameters.
|
||||
- Check and review error logs.
|
||||
- Debug the tool implementation.
|
||||
- Ensure proper permissions in Home Assistant.
|
||||
|
||||
## Debugging Steps
|
||||
|
||||
### Server Logs
|
||||
|
||||
1. Enable debug logging by setting:
|
||||
```env
|
||||
LOG_LEVEL=debug
|
||||
```
|
||||
2. Check logs:
|
||||
```bash
|
||||
npm run logs
|
||||
```
|
||||
3. Filter errors:
|
||||
```bash
|
||||
npm run logs | grep "error"
|
||||
```
|
||||
|
||||
### Network Debugging
|
||||
|
||||
1. Test API endpoints:
|
||||
```bash
|
||||
curl -v http://localhost:3000/api/health
|
||||
```
|
||||
2. Monitor SSE connections:
|
||||
```bash
|
||||
curl -N http://localhost:3000/api/sse/stats
|
||||
```
|
||||
3. Test WebSocket connectivity:
|
||||
```bash
|
||||
wscat -c ws://localhost:3000
|
||||
```
|
||||
|
||||
### Performance Issues
|
||||
|
||||
- Monitor memory usage with:
|
||||
```bash
|
||||
npm run stats
|
||||
```
|
||||
|
||||
## Security Middleware Troubleshooting
|
||||
|
||||
### Rate Limiting Problems
|
||||
|
||||
**Symptoms:** Receiving 429 (Too Many Requests) errors.
|
||||
|
||||
**Solutions:**
|
||||
- Adjust and fine-tune rate limit settings.
|
||||
- Consider different limits for critical versus non-critical endpoints.
|
||||
|
||||
### Request Validation Failures
|
||||
|
||||
**Symptoms:** 400 or 415 errors on valid requests.
|
||||
|
||||
**Solutions:**
|
||||
- Verify that the `Content-Type` header is set correctly.
|
||||
- Inspect request payload size and format.
|
||||
|
||||
### Input Sanitization Issues
|
||||
|
||||
**Symptoms:** Unexpected data transformation or loss.
|
||||
|
||||
**Solutions:**
|
||||
- Test sanitization with various input types.
|
||||
- Implement custom sanitization for complex data if needed.
|
||||
|
||||
### Security Header Configuration
|
||||
|
||||
**Symptoms:** Missing or improper security headers.
|
||||
|
||||
**Solutions:**
|
||||
- Review and update security header configurations (e.g., Helmet settings).
|
||||
- Ensure environment-specific header settings are in place.
|
||||
|
||||
### Error Handling and Logging
|
||||
|
||||
**Symptoms:** Inconsistent error responses.
|
||||
|
||||
**Solutions:**
|
||||
- Enhance logging for detailed error tracking.
|
||||
- Adjust error handlers for production and development differences.
|
||||
|
||||
## Additional Resources
|
||||
|
||||
- [OWASP Security Guidelines](https://owasp.org/www-project-top-ten/)
|
||||
- [Helmet.js Documentation](https://helmetjs.github.io/)
|
||||
- [JWT Security Best Practices](https://jwt.io/introduction)
|
||||
|
||||
## Getting Help
|
||||
|
||||
If issues persist:
|
||||
1. Review detailed logs.
|
||||
2. Verify your configuration and environment.
|
||||
3. Consult the GitHub issue tracker or community forums.
|
||||
|
||||
## FAQ
|
||||
|
||||
### Q: How do I reset my configuration?
|
||||
A: Delete `.env` and copy `.env.example` to start fresh.
|
||||
|
||||
### Q: Why are my events delayed?
|
||||
A: Check network latency and server load. Consider adjusting buffer sizes.
|
||||
|
||||
### Q: How do I update my token?
|
||||
A: Generate a new token in Home Assistant and update HASS_TOKEN.
|
||||
|
||||
### Q: Why do I get "Maximum clients reached"?
|
||||
A: Adjust SSE_MAX_CLIENTS in configuration or clean up stale connections.
|
||||
|
||||
## Error Codes
|
||||
|
||||
- `E001`: Connection Error
|
||||
- `E002`: Authentication Error
|
||||
- `E003`: Rate Limit Error
|
||||
- `E004`: Tool Error
|
||||
- `E005`: Configuration Error
|
||||
|
||||
## Support Resources
|
||||
|
||||
1. Documentation
|
||||
- [API Reference](./API.md)
|
||||
- [Configuration Guide](./configuration/README.md)
|
||||
- [Development Guide](./development/development.md)
|
||||
|
||||
2. Community
|
||||
- GitHub Issues
|
||||
- Discussion Forums
|
||||
- Stack Overflow
|
||||
|
||||
3. Tools
|
||||
- Diagnostic Scripts
|
||||
- Testing Tools
|
||||
- Monitoring Tools
|
||||
|
||||
## Still Need Help?
|
||||
|
||||
1. Create a detailed issue:
|
||||
- Error messages
|
||||
- Steps to reproduce
|
||||
- Environment details
|
||||
- Logs
|
||||
|
||||
2. Contact support:
|
||||
- GitHub Issues
|
||||
- Email Support
|
||||
- Community Forums
|
||||
|
||||
## Security Middleware Troubleshooting
|
||||
|
||||
### Common Issues and Solutions
|
||||
|
||||
#### Rate Limiting Problems
|
||||
|
||||
**Symptom**: Unexpected 429 (Too Many Requests) errors
|
||||
|
||||
**Possible Causes**:
|
||||
- Misconfigured rate limit settings
|
||||
- Shared IP addresses (e.g., behind NAT)
|
||||
- Aggressive client-side retry mechanisms
|
||||
|
||||
**Solutions**:
|
||||
1. Adjust rate limit parameters
|
||||
```typescript
|
||||
// Customize rate limit for specific scenarios
|
||||
checkRateLimit(ip, maxRequests = 200, windowMs = 30 * 60 * 1000)
|
||||
```
|
||||
|
||||
2. Implement more granular rate limiting
|
||||
- Use different limits for different endpoints
|
||||
- Consider user authentication level
|
||||
|
||||
#### Request Validation Failures
|
||||
|
||||
**Symptom**: 400 or 415 status codes on valid requests
|
||||
|
||||
**Possible Causes**:
|
||||
- Incorrect `Content-Type` header
|
||||
- Large request payloads
|
||||
- Malformed authorization headers
|
||||
|
||||
**Debugging Steps**:
|
||||
1. Verify request headers
|
||||
```typescript
|
||||
// Check content type and size
|
||||
validateRequestHeaders(request, 'application/json')
|
||||
```
|
||||
|
||||
2. Log detailed validation errors
|
||||
```typescript
|
||||
try {
|
||||
validateRequestHeaders(request);
|
||||
} catch (error) {
|
||||
console.error('Request validation failed:', error.message);
|
||||
}
|
||||
```
|
||||
|
||||
#### Input Sanitization Issues
|
||||
|
||||
**Symptom**: Unexpected data transformation or loss
|
||||
|
||||
**Possible Causes**:
|
||||
- Complex nested objects
|
||||
- Non-standard input formats
|
||||
- Overly aggressive sanitization
|
||||
|
||||
**Troubleshooting**:
|
||||
1. Test sanitization with various input types
|
||||
```typescript
|
||||
const input = {
|
||||
text: '<script>alert("xss")</script>',
|
||||
nested: { html: '<img src="x" onerror="alert(1)">World' }
|
||||
};
|
||||
const sanitized = sanitizeValue(input);
|
||||
```
|
||||
|
||||
2. Custom sanitization for specific use cases
|
||||
```typescript
|
||||
function customSanitize(value) {
|
||||
// Add custom sanitization logic
|
||||
return sanitizeValue(value);
|
||||
}
|
||||
```
|
||||
|
||||
#### Security Header Configuration
|
||||
|
||||
**Symptom**: Missing or incorrect security headers
|
||||
|
||||
**Possible Causes**:
|
||||
- Misconfigured Helmet options
|
||||
- Environment-specific header requirements
|
||||
|
||||
**Solutions**:
|
||||
1. Custom security header configuration
|
||||
```typescript
|
||||
const customHelmetConfig = {
|
||||
contentSecurityPolicy: {
|
||||
directives: {
|
||||
defaultSrc: ["'self'"],
|
||||
scriptSrc: ["'self'", 'trusted-cdn.com']
|
||||
}
|
||||
}
|
||||
};
|
||||
applySecurityHeaders(request, customHelmetConfig);
|
||||
```
|
||||
|
||||
#### Error Handling and Logging
|
||||
|
||||
**Symptom**: Inconsistent error responses
|
||||
|
||||
**Possible Causes**:
|
||||
- Incorrect environment configuration
|
||||
- Unhandled error types
|
||||
|
||||
**Debugging Techniques**:
|
||||
1. Verify environment settings
|
||||
```typescript
|
||||
const errorResponse = handleError(error, process.env.NODE_ENV);
|
||||
```
|
||||
|
||||
2. Add custom error handling
|
||||
```typescript
|
||||
function enhancedErrorHandler(error, env) {
|
||||
// Add custom logging or monitoring
|
||||
console.error('Security error:', error);
|
||||
return handleError(error, env);
|
||||
}
|
||||
```
|
||||
|
||||
### Performance and Security Monitoring
|
||||
|
||||
1. **Logging**
|
||||
- Enable debug logging for security events
|
||||
- Monitor rate limit and validation logs
|
||||
|
||||
2. **Metrics**
|
||||
- Track rate limit hit rates
|
||||
- Monitor request validation success/failure ratios
|
||||
|
||||
3. **Continuous Improvement**
|
||||
- Regularly review and update security configurations
|
||||
- Conduct periodic security audits
|
||||
|
||||
### Environment-Specific Considerations
|
||||
|
||||
#### Development
|
||||
- More verbose error messages
|
||||
- Relaxed rate limiting
|
||||
- Detailed security logs
|
||||
|
||||
#### Production
|
||||
- Minimal error details
|
||||
- Strict rate limiting
|
||||
- Comprehensive security headers
|
||||
|
||||
### External Resources
|
||||
|
||||
- [OWASP Security Guidelines](https://owasp.org/www-project-top-ten/)
|
||||
- [Helmet.js Documentation](https://helmetjs.github.io/)
|
||||
- [JWT Security Best Practices](https://jwt.io/introduction)
|
||||
|
||||
### Getting Help
|
||||
|
||||
If you encounter persistent issues:
|
||||
1. Check application logs
|
||||
2. Verify environment configurations
|
||||
3. Consult the project's issue tracker
|
||||
4. Reach out to the development team with detailed error information
|
||||
@@ -1,34 +0,0 @@
|
||||
# Usage Guide
|
||||
|
||||
This guide explains how to use the Home Assistant MCP Server for smart home device management and integration with language learning systems.
|
||||
|
||||
## Basic Usage
|
||||
|
||||
1. **Starting the Server:**
|
||||
- For development: run `npm run dev`.
|
||||
- For production: run `npm run build` followed by `npm start`.
|
||||
|
||||
2. **Accessing the Web Interface:**
|
||||
- Open [http://localhost:3000](http://localhost:3000) in your browser.
|
||||
|
||||
3. **Real-Time Updates:**
|
||||
- Connect to the SSE endpoint at `/subscribe_events?token=YOUR_TOKEN&domain=light` to receive live updates.
|
||||
|
||||
## Advanced Features
|
||||
|
||||
1. **API Interactions:**
|
||||
- Use the REST API for operations such as device control, automation, and add-on management.
|
||||
- See [API Documentation](api.md) for details.
|
||||
|
||||
2. **Tool Integrations:**
|
||||
- Multiple tools are available (see [Tools Documentation](tools/tools.md)), for tasks like automation management and notifications.
|
||||
|
||||
3. **Security Settings:**
|
||||
- Configure token-based authentication and environment variables as per the [Configuration Guide](getting-started/configuration.md).
|
||||
|
||||
4. **Customization and Extensions:**
|
||||
- Extend server functionality by developing new tools as outlined in the [Development Guide](development/development.md).
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
If you experience issues, review the [Troubleshooting Guide](troubleshooting.md).
|
||||
@@ -327,3 +327,7 @@ if [[ $REPLY =~ ^[Yy]$ ]]; then
|
||||
echo -e "${GREEN}Home Assistant MCP test successful!${NC}"
|
||||
fi
|
||||
fi
|
||||
|
||||
# macOS environment configuration
|
||||
HASS_SOCKET_URL="${HASS_HOST/http/ws}/api/websocket" # WebSocket URL conversion
|
||||
chmod 600 "$CLAUDE_CONFIG_DIR/claude_desktop_config.json" # Security hardening
|
||||
@@ -1,21 +1,19 @@
|
||||
import fetch from "node-fetch";
|
||||
import OpenAI from "openai";
|
||||
import { Anthropic } from "@anthropic-ai/sdk";
|
||||
import { DOMParser, Element, Document } from '@xmldom/xmldom';
|
||||
import dotenv from 'dotenv';
|
||||
import readline from 'readline';
|
||||
import chalk from 'chalk';
|
||||
import express from 'express';
|
||||
import bodyParser from 'body-parser';
|
||||
|
||||
// Load environment variables
|
||||
dotenv.config();
|
||||
|
||||
// Retrieve API keys from environment variables
|
||||
const openaiApiKey = process.env.OPENAI_API_KEY;
|
||||
const anthropicApiKey = process.env.ANTHROPIC_API_KEY;
|
||||
const hassToken = process.env.HASS_TOKEN;
|
||||
|
||||
if (!openaiApiKey) {
|
||||
console.error("Please set the OPENAI_API_KEY environment variable.");
|
||||
if (!anthropicApiKey) {
|
||||
console.error("Please set the ANTHROPIC_API_KEY environment variable.");
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
@@ -25,7 +23,7 @@ if (!hassToken) {
|
||||
}
|
||||
|
||||
// MCP Server configuration
|
||||
const MCP_SERVER = process.env.MCP_SERVER || 'http://localhost:3000';
|
||||
const MCP_SERVER = 'http://localhost:3000';
|
||||
|
||||
interface McpTool {
|
||||
name: string;
|
||||
@@ -115,14 +113,11 @@ interface ModelConfig {
|
||||
contextWindow: number;
|
||||
}
|
||||
|
||||
// Update model listing to filter based on API key availability
|
||||
// Update model listing to use Anthropic's Claude models
|
||||
const AVAILABLE_MODELS: ModelConfig[] = [
|
||||
// OpenAI models always available
|
||||
{ name: 'gpt-4o', maxTokens: 4096, contextWindow: 128000 },
|
||||
{ name: 'gpt-4-turbo', maxTokens: 4096, contextWindow: 128000 },
|
||||
{ name: 'gpt-4', maxTokens: 8192, contextWindow: 128000 },
|
||||
{ name: 'gpt-3.5-turbo', maxTokens: 4096, contextWindow: 16385 },
|
||||
{ name: 'gpt-3.5-turbo-16k', maxTokens: 16385, contextWindow: 16385 },
|
||||
// Anthropic Claude models
|
||||
{ name: 'claude-3-7-sonnet-20250219', maxTokens: 4096, contextWindow: 200000 },
|
||||
{ name: 'claude-3-5-haiku-20241022', maxTokens: 4096, contextWindow: 200000 },
|
||||
|
||||
// Conditionally include DeepSeek models
|
||||
...(process.env.DEEPSEEK_API_KEY ? [
|
||||
@@ -134,7 +129,7 @@ const AVAILABLE_MODELS: ModelConfig[] = [
|
||||
// Add configuration interface
|
||||
interface AppConfig {
|
||||
mcpServer: string;
|
||||
openaiModel: string;
|
||||
anthropicModel: string;
|
||||
maxRetries: number;
|
||||
analysisTimeout: number;
|
||||
selectedModel: ModelConfig;
|
||||
@@ -149,36 +144,31 @@ const logger = {
|
||||
debug: (msg: string) => process.env.DEBUG && console.log(chalk.gray(`› ${msg}`))
|
||||
};
|
||||
|
||||
// Update default model selection in loadConfig
|
||||
// Update loadConfig to use Claude models
|
||||
function loadConfig(): AppConfig {
|
||||
// Use environment variable or default to gpt-4o
|
||||
const defaultModelName = process.env.OPENAI_MODEL || 'gpt-4o';
|
||||
let defaultModel = AVAILABLE_MODELS.find(m => m.name === defaultModelName);
|
||||
|
||||
// If the configured model isn't found, use gpt-4o without warning
|
||||
if (!defaultModel) {
|
||||
defaultModel = AVAILABLE_MODELS.find(m => m.name === 'gpt-4o') || AVAILABLE_MODELS[0];
|
||||
}
|
||||
// Use Claude 3.7 Sonnet as the default model
|
||||
const defaultModel = AVAILABLE_MODELS.find(m => m.name === 'claude-3-7-sonnet-20250219') || AVAILABLE_MODELS[0];
|
||||
|
||||
return {
|
||||
mcpServer: process.env.MCP_SERVER || 'http://localhost:3000',
|
||||
openaiModel: defaultModel.name, // Use the resolved model name
|
||||
anthropicModel: defaultModel.name,
|
||||
maxRetries: parseInt(process.env.MAX_RETRIES || '3'),
|
||||
analysisTimeout: parseInt(process.env.ANALYSIS_TIMEOUT || '30000'),
|
||||
selectedModel: defaultModel
|
||||
};
|
||||
}
|
||||
|
||||
function getOpenAIClient(): OpenAI {
|
||||
// Replace OpenAI client with Anthropic client
|
||||
function getAnthropicClient(): Anthropic {
|
||||
const config = loadConfig();
|
||||
|
||||
return new OpenAI({
|
||||
apiKey: config.selectedModel.name.startsWith('deepseek')
|
||||
? process.env.DEEPSEEK_API_KEY
|
||||
: openaiApiKey,
|
||||
baseURL: config.selectedModel.name.startsWith('deepseek')
|
||||
? 'https://api.deepseek.com/v1'
|
||||
: 'https://api.openai.com/v1'
|
||||
if (config.selectedModel.name.startsWith('deepseek') && process.env.DEEPSEEK_API_KEY) {
|
||||
// This is just a stub for DeepSeek - you'd need to implement this properly
|
||||
throw new Error("DeepSeek models not implemented yet with Anthropic integration");
|
||||
}
|
||||
|
||||
return new Anthropic({
|
||||
apiKey: anthropicApiKey,
|
||||
});
|
||||
}
|
||||
|
||||
@@ -194,8 +184,8 @@ async function executeMcpTool(toolName: string, parameters: Record<string, any>
|
||||
const controller = new AbortController();
|
||||
const timeoutId = setTimeout(() => controller.abort(), config.analysisTimeout);
|
||||
|
||||
// Update endpoint URL to use the same base path as schema
|
||||
const endpoint = `${config.mcpServer}/mcp/execute`;
|
||||
// Update endpoint URL to use the correct API path
|
||||
const endpoint = `${config.mcpServer}/api/mcp/execute`;
|
||||
|
||||
const response = await fetch(endpoint, {
|
||||
method: "POST",
|
||||
@@ -258,43 +248,117 @@ function isMcpExecuteResponse(obj: any): obj is McpExecuteResponse {
|
||||
(obj.success === true || typeof obj.message === 'string');
|
||||
}
|
||||
|
||||
// Add mock data for testing
|
||||
const MOCK_HA_INFO = {
|
||||
devices: {
|
||||
light: [
|
||||
{ entity_id: 'light.living_room', state: 'on', attributes: { friendly_name: 'Living Room Light', brightness: 255 } },
|
||||
{ entity_id: 'light.kitchen', state: 'off', attributes: { friendly_name: 'Kitchen Light', brightness: 0 } }
|
||||
],
|
||||
switch: [
|
||||
{ entity_id: 'switch.tv', state: 'off', attributes: { friendly_name: 'TV Power' } }
|
||||
],
|
||||
sensor: [
|
||||
{ entity_id: 'sensor.temperature', state: '21.5', attributes: { friendly_name: 'Living Room Temperature', unit_of_measurement: '°C' } },
|
||||
{ entity_id: 'sensor.humidity', state: '45', attributes: { friendly_name: 'Living Room Humidity', unit_of_measurement: '%' } }
|
||||
],
|
||||
climate: [
|
||||
{ entity_id: 'climate.thermostat', state: 'heat', attributes: { friendly_name: 'Main Thermostat', current_temperature: 20, target_temp_high: 24 } }
|
||||
]
|
||||
}
|
||||
};
|
||||
|
||||
interface HassState {
|
||||
entity_id: string;
|
||||
state: string;
|
||||
attributes: Record<string, any>;
|
||||
last_changed: string;
|
||||
last_updated: string;
|
||||
}
|
||||
|
||||
interface ServiceInfo {
|
||||
name: string;
|
||||
description: string;
|
||||
fields: Record<string, any>;
|
||||
}
|
||||
|
||||
interface ServiceDomain {
|
||||
domain: string;
|
||||
services: Record<string, ServiceInfo>;
|
||||
}
|
||||
|
||||
/**
|
||||
* Collects comprehensive information about the Home Assistant instance using MCP tools
|
||||
*/
|
||||
async function collectHomeAssistantInfo(): Promise<any> {
|
||||
const info: Record<string, any> = {};
|
||||
const config = loadConfig();
|
||||
const hassHost = process.env.HASS_HOST;
|
||||
|
||||
// Update schema endpoint to be consistent
|
||||
const schemaResponse = await fetch(`${config.mcpServer}/mcp`, {
|
||||
try {
|
||||
// Check if we're in test mode
|
||||
if (process.env.HA_TEST_MODE === '1') {
|
||||
logger.info("Running in test mode with mock data");
|
||||
return MOCK_HA_INFO;
|
||||
}
|
||||
|
||||
// Get states from Home Assistant directly
|
||||
const statesResponse = await fetch(`${hassHost}/api/states`, {
|
||||
headers: {
|
||||
'Authorization': `Bearer ${hassToken}`,
|
||||
'Accept': 'application/json'
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
|
||||
if (!schemaResponse.ok) {
|
||||
console.error(`Failed to fetch MCP schema: ${schemaResponse.status}`);
|
||||
return info;
|
||||
if (!statesResponse.ok) {
|
||||
throw new Error(`Failed to fetch states: ${statesResponse.status}`);
|
||||
}
|
||||
|
||||
const schema = await schemaResponse.json() as McpSchema;
|
||||
console.log("Available tools:", schema.tools.map(t => t.name));
|
||||
const states = await statesResponse.json() as HassState[];
|
||||
|
||||
// Execute list_devices to get basic device information
|
||||
console.log("Fetching device information...");
|
||||
try {
|
||||
const deviceInfo = await executeMcpTool('list_devices');
|
||||
if (deviceInfo && deviceInfo.success && deviceInfo.devices) {
|
||||
info.devices = deviceInfo.devices;
|
||||
// Group devices by domain
|
||||
const devices: Record<string, HassState[]> = {};
|
||||
for (const state of states) {
|
||||
const [domain] = state.entity_id.split('.');
|
||||
if (!devices[domain]) {
|
||||
devices[domain] = [];
|
||||
}
|
||||
devices[domain].push(state);
|
||||
}
|
||||
|
||||
info.devices = devices;
|
||||
info.device_summary = {
|
||||
total_devices: states.length,
|
||||
device_types: Object.keys(devices),
|
||||
by_domain: Object.fromEntries(
|
||||
Object.entries(devices).map(([domain, items]) => [domain, items.length])
|
||||
)
|
||||
};
|
||||
|
||||
const deviceCount = states.length;
|
||||
const domainCount = Object.keys(devices).length;
|
||||
|
||||
if (deviceCount > 0) {
|
||||
logger.success(`Found ${deviceCount} devices across ${domainCount} domains`);
|
||||
} else {
|
||||
console.warn(`Failed to list devices: ${deviceInfo?.message || 'Unknown error'}`);
|
||||
}
|
||||
} catch (error) {
|
||||
console.warn("Error fetching devices:", error);
|
||||
logger.warn('No devices found in Home Assistant');
|
||||
}
|
||||
|
||||
return info;
|
||||
} catch (error) {
|
||||
logger.error(`Error fetching devices: ${error instanceof Error ? error.message : 'Unknown error'}`);
|
||||
if (process.env.HA_TEST_MODE !== '1') {
|
||||
logger.warn(`Failed to connect to Home Assistant. Run with HA_TEST_MODE=1 to use test data.`);
|
||||
return {
|
||||
devices: {},
|
||||
device_summary: {
|
||||
total_devices: 0,
|
||||
device_types: [],
|
||||
by_domain: {}
|
||||
}
|
||||
};
|
||||
}
|
||||
return MOCK_HA_INFO;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -398,34 +462,69 @@ function getRelevantDeviceTypes(prompt: string): string[] {
|
||||
}
|
||||
|
||||
/**
|
||||
* Generates analysis and recommendations using the OpenAI API based on the Home Assistant data
|
||||
* Generates analysis and recommendations using the Anthropic API based on the Home Assistant data
|
||||
*/
|
||||
async function generateAnalysis(haInfo: any): Promise<SystemAnalysis> {
|
||||
const openai = getOpenAIClient();
|
||||
const config = loadConfig();
|
||||
|
||||
// Compress and summarize the data
|
||||
const deviceTypes = haInfo.devices ? Object.keys(haInfo.devices) : [];
|
||||
const deviceSummary = haInfo.devices ? Object.entries(haInfo.devices).reduce((acc: Record<string, any>, [domain, devices]) => {
|
||||
const deviceList = devices as any[];
|
||||
acc[domain] = {
|
||||
count: deviceList.length,
|
||||
active: deviceList.filter(d => d.state === 'on' || d.state === 'home').length,
|
||||
states: [...new Set(deviceList.map(d => d.state))],
|
||||
sample: deviceList.slice(0, 2).map(d => ({
|
||||
id: d.entity_id,
|
||||
state: d.state,
|
||||
name: d.attributes?.friendly_name
|
||||
}))
|
||||
// If in test mode, return mock analysis
|
||||
if (process.env.HA_TEST_MODE === '1') {
|
||||
logger.info("Generating mock analysis...");
|
||||
return {
|
||||
overview: {
|
||||
state: ["System running normally", "4 device types detected"],
|
||||
health: ["All systems operational", "No critical issues found"],
|
||||
configurations: ["Basic configuration detected", "Default settings in use"],
|
||||
integrations: ["Light", "Switch", "Sensor", "Climate"],
|
||||
issues: ["No major issues detected"]
|
||||
},
|
||||
performance: {
|
||||
resource_usage: ["Normal CPU usage", "Memory usage within limits"],
|
||||
response_times: ["Average response time: 0.5s"],
|
||||
optimization_areas: ["Consider grouping lights by room"]
|
||||
},
|
||||
security: {
|
||||
current_measures: ["Basic security measures in place"],
|
||||
vulnerabilities: ["No critical vulnerabilities detected"],
|
||||
recommendations: ["Enable 2FA if not already enabled"]
|
||||
},
|
||||
optimization: {
|
||||
performance_suggestions: ["Group frequently used devices"],
|
||||
config_optimizations: ["Consider creating room-based views"],
|
||||
integration_improvements: ["Add friendly names to all entities"],
|
||||
automation_opportunities: ["Create morning/evening routines"]
|
||||
},
|
||||
maintenance: {
|
||||
required_updates: ["No critical updates pending"],
|
||||
cleanup_tasks: ["Remove unused entities"],
|
||||
regular_tasks: ["Check sensor battery levels"]
|
||||
},
|
||||
entity_usage: {
|
||||
most_active: ["light.living_room", "sensor.temperature"],
|
||||
rarely_used: ["switch.tv"],
|
||||
potential_duplicates: []
|
||||
},
|
||||
automation_analysis: {
|
||||
inefficient_automations: [],
|
||||
potential_improvements: ["Add time-based light controls"],
|
||||
suggested_blueprints: ["Motion-activated lighting"],
|
||||
condition_optimizations: []
|
||||
},
|
||||
energy_management: {
|
||||
high_consumption: ["No high consumption devices detected"],
|
||||
monitoring_suggestions: ["Add power monitoring to main appliances"],
|
||||
tariff_optimizations: ["Consider time-of-use automation"]
|
||||
}
|
||||
};
|
||||
return acc;
|
||||
}, {}) : {};
|
||||
}
|
||||
|
||||
// Original analysis code for non-test mode
|
||||
const anthropic = getAnthropicClient();
|
||||
|
||||
const systemSummary = {
|
||||
total_devices: deviceTypes.reduce((sum, type) => sum + deviceSummary[type].count, 0),
|
||||
device_types: deviceTypes,
|
||||
device_summary: deviceSummary,
|
||||
active_devices: Object.values(deviceSummary).reduce((sum: number, info: any) => sum + info.active, 0)
|
||||
total_devices: haInfo.device_summary?.total_devices || 0,
|
||||
device_types: haInfo.device_summary?.device_types || [],
|
||||
device_summary: haInfo.device_summary?.by_domain || {}
|
||||
};
|
||||
|
||||
const prompt = `Analyze this Home Assistant system and provide insights in XML format:
|
||||
@@ -488,20 +587,21 @@ Generate your response in this EXACT format:
|
||||
</analysis>`;
|
||||
|
||||
try {
|
||||
const completion = await openai.chat.completions.create({
|
||||
const completion = await anthropic.messages.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "system",
|
||||
content: "You are a Home Assistant expert. Analyze the system data and provide detailed insights in the specified XML format. Be specific and actionable in your recommendations."
|
||||
},
|
||||
{ role: "user", content: prompt }
|
||||
role: "user",
|
||||
content: `<system>You are a Home Assistant expert. Analyze the system data and provide detailed insights in the specified XML format. Be specific and actionable in your recommendations.</system>
|
||||
|
||||
${prompt}`
|
||||
}
|
||||
],
|
||||
temperature: 0.7,
|
||||
max_tokens: Math.min(config.selectedModel.maxTokens, 4000)
|
||||
});
|
||||
|
||||
const result = completion.choices[0].message?.content || "";
|
||||
const result = completion.content[0]?.type === 'text' ? completion.content[0].text : "";
|
||||
|
||||
// Clean the response and parse XML
|
||||
const cleanedResult = result.replace(/```xml/g, '').replace(/```/g, '').trim();
|
||||
@@ -573,105 +673,97 @@ Generate your response in this EXACT format:
|
||||
throw new Error(`Failed to parse analysis response: ${parseError.message}`);
|
||||
}
|
||||
} catch (error) {
|
||||
console.error("Error during OpenAI API call:", error);
|
||||
console.error("Error during Anthropic API call:", error);
|
||||
throw new Error("Failed to generate analysis");
|
||||
}
|
||||
}
|
||||
|
||||
async function getUserInput(question: string): Promise<string> {
|
||||
const rl = readline.createInterface({
|
||||
input: process.stdin,
|
||||
output: process.stdout
|
||||
});
|
||||
|
||||
return new Promise((resolve) => {
|
||||
rl.question(question, (answer) => {
|
||||
rl.close();
|
||||
resolve(answer);
|
||||
});
|
||||
});
|
||||
interface AutomationConfig {
|
||||
id?: string;
|
||||
alias?: string;
|
||||
description?: string;
|
||||
trigger?: Array<{
|
||||
platform: string;
|
||||
[key: string]: any;
|
||||
}>;
|
||||
condition?: Array<{
|
||||
condition: string;
|
||||
[key: string]: any;
|
||||
}>;
|
||||
action?: Array<{
|
||||
service?: string;
|
||||
[key: string]: any;
|
||||
}>;
|
||||
mode?: string;
|
||||
}
|
||||
|
||||
// Update chunk size calculation
|
||||
const MAX_CHARACTERS = 8000; // ~2000 tokens (4 chars/token)
|
||||
|
||||
// Update model handling in retry
|
||||
async function handleCustomPrompt(haInfo: any): Promise<void> {
|
||||
try {
|
||||
// Add device metadata
|
||||
const deviceTypes = haInfo.devices ? Object.keys(haInfo.devices) : [];
|
||||
const deviceStates = haInfo.devices ? Object.entries(haInfo.devices).reduce((acc: Record<string, number>, [domain, devices]) => {
|
||||
acc[domain] = (devices as any[]).length;
|
||||
return acc;
|
||||
}, {}) : {};
|
||||
const totalDevices = deviceTypes.reduce((sum, type) => sum + deviceStates[type], 0);
|
||||
|
||||
const userPrompt = await getUserInput("Enter your custom prompt: ");
|
||||
if (!userPrompt) {
|
||||
console.log("No prompt provided. Exiting...");
|
||||
return;
|
||||
}
|
||||
|
||||
const openai = getOpenAIClient();
|
||||
const config = loadConfig();
|
||||
|
||||
const completion = await openai.chat.completions.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "system",
|
||||
content: `You are a Home Assistant expert. Analyze the following Home Assistant information and respond to the user's prompt.
|
||||
Current system has ${totalDevices} devices across ${deviceTypes.length} types: ${JSON.stringify(deviceStates)}`
|
||||
},
|
||||
{ role: "user", content: userPrompt },
|
||||
],
|
||||
max_tokens: config.selectedModel.maxTokens,
|
||||
temperature: 0.3,
|
||||
});
|
||||
|
||||
console.log("\nAnalysis Results:\n");
|
||||
console.log(completion.choices[0].message?.content || "No response generated");
|
||||
|
||||
} catch (error) {
|
||||
console.error("Error processing custom prompt:", error);
|
||||
|
||||
// Retry with simplified prompt if there's an error
|
||||
try {
|
||||
const retryPrompt = "Please provide a simpler analysis of the Home Assistant system.";
|
||||
const openai = getOpenAIClient();
|
||||
const config = loadConfig();
|
||||
|
||||
const retryCompletion = await openai.chat.completions.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "system",
|
||||
content: "You are a Home Assistant expert. Provide a simple analysis of the system."
|
||||
},
|
||||
{ role: "user", content: retryPrompt },
|
||||
],
|
||||
max_tokens: config.selectedModel.maxTokens,
|
||||
temperature: 0.3,
|
||||
});
|
||||
|
||||
console.log("\nAnalysis Results:\n");
|
||||
console.log(retryCompletion.choices[0].message?.content || "No response generated");
|
||||
} catch (retryError) {
|
||||
console.error("Error during retry:", retryError);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Update automation handling
|
||||
async function handleAutomationOptimization(haInfo: any): Promise<void> {
|
||||
try {
|
||||
const result = await executeMcpTool('automation', { action: 'list' });
|
||||
if (!result?.success) {
|
||||
logger.error(`Failed to retrieve automations: ${result?.message || 'Unknown error'}`);
|
||||
return;
|
||||
const hassHost = process.env.HASS_HOST;
|
||||
|
||||
// Get automations directly from Home Assistant
|
||||
const automationsResponse = await fetch(`${hassHost}/api/states`, {
|
||||
headers: {
|
||||
'Authorization': `Bearer ${hassToken}`,
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
|
||||
if (!automationsResponse.ok) {
|
||||
throw new Error(`Failed to fetch automations: ${automationsResponse.status}`);
|
||||
}
|
||||
|
||||
const automations = result.automations || [];
|
||||
const states = await automationsResponse.json() as HassState[];
|
||||
const automations = states.filter(state => state.entity_id.startsWith('automation.'));
|
||||
|
||||
// Get services to understand what actions are available
|
||||
const servicesResponse = await fetch(`${hassHost}/api/services`, {
|
||||
headers: {
|
||||
'Authorization': `Bearer ${hassToken}`,
|
||||
'Content-Type': 'application/json'
|
||||
}
|
||||
});
|
||||
|
||||
let availableServices: Record<string, any> = {};
|
||||
if (servicesResponse.ok) {
|
||||
const services = await servicesResponse.json() as ServiceDomain[];
|
||||
availableServices = services.reduce((acc: Record<string, any>, service: ServiceDomain) => {
|
||||
if (service.domain && service.services) {
|
||||
acc[service.domain] = service.services;
|
||||
}
|
||||
return acc;
|
||||
}, {});
|
||||
logger.debug(`Retrieved services from ${Object.keys(availableServices).length} domains`);
|
||||
}
|
||||
|
||||
// Enrich automation data with service information
|
||||
const enrichedAutomations = automations.map(automation => {
|
||||
const actions = automation.attributes?.action || [];
|
||||
const enrichedActions = actions.map((action: any) => {
|
||||
if (action.service) {
|
||||
const [domain, service] = action.service.split('.');
|
||||
const serviceInfo = availableServices[domain]?.[service];
|
||||
return {
|
||||
...action,
|
||||
service_info: serviceInfo
|
||||
};
|
||||
}
|
||||
return action;
|
||||
});
|
||||
|
||||
return {
|
||||
...automation,
|
||||
config: {
|
||||
id: automation.entity_id.split('.')[1],
|
||||
alias: automation.attributes?.friendly_name,
|
||||
trigger: automation.attributes?.trigger || [],
|
||||
condition: automation.attributes?.condition || [],
|
||||
action: enrichedActions,
|
||||
mode: automation.attributes?.mode || 'single'
|
||||
}
|
||||
};
|
||||
});
|
||||
|
||||
if (automations.length === 0) {
|
||||
console.log(chalk.bold.underline("\nAutomation Optimization Report"));
|
||||
console.log(chalk.yellow("No automations found in the system. Consider creating some automations to improve your Home Assistant experience."));
|
||||
@@ -679,7 +771,7 @@ async function handleAutomationOptimization(haInfo: any): Promise<void> {
|
||||
}
|
||||
|
||||
logger.info(`Analyzing ${automations.length} automations...`);
|
||||
const optimizationXml = await analyzeAutomations(automations);
|
||||
const optimizationXml = await analyzeAutomations(enrichedAutomations);
|
||||
|
||||
const parser = new DOMParser();
|
||||
const xmlDoc = parser.parseFromString(optimizationXml, "text/xml");
|
||||
@@ -721,67 +813,102 @@ async function handleAutomationOptimization(haInfo: any): Promise<void> {
|
||||
}
|
||||
}
|
||||
|
||||
// Add new automation optimization function
|
||||
async function analyzeAutomations(automations: any[]): Promise<string> {
|
||||
const openai = getOpenAIClient();
|
||||
const anthropic = getAnthropicClient();
|
||||
const config = loadConfig();
|
||||
|
||||
// Compress automation data by only including essential fields
|
||||
const compressedAutomations = automations.map(automation => ({
|
||||
id: automation.entity_id,
|
||||
name: automation.attributes?.friendly_name || automation.entity_id,
|
||||
state: automation.state,
|
||||
last_triggered: automation.attributes?.last_triggered,
|
||||
mode: automation.attributes?.mode,
|
||||
trigger_count: automation.attributes?.trigger?.length || 0,
|
||||
action_count: automation.attributes?.action?.length || 0
|
||||
}));
|
||||
// Create a more detailed summary of automations
|
||||
const automationSummary = {
|
||||
total: automations.length,
|
||||
active: automations.filter(a => a.state === 'on').length,
|
||||
by_type: automations.reduce((acc: Record<string, number>, auto) => {
|
||||
const type = auto.attributes?.mode || 'single';
|
||||
acc[type] = (acc[type] || 0) + 1;
|
||||
return acc;
|
||||
}, {}),
|
||||
recently_triggered: automations.filter(a => {
|
||||
const lastTriggered = a.attributes?.last_triggered;
|
||||
if (!lastTriggered) return false;
|
||||
const lastTriggerDate = new Date(lastTriggered);
|
||||
const oneDayAgo = new Date();
|
||||
oneDayAgo.setDate(oneDayAgo.getDate() - 1);
|
||||
return lastTriggerDate > oneDayAgo;
|
||||
}).length,
|
||||
trigger_types: automations.reduce((acc: Record<string, number>, auto) => {
|
||||
const triggers = auto.config?.trigger || [];
|
||||
triggers.forEach((trigger: any) => {
|
||||
const type = trigger.platform || 'unknown';
|
||||
acc[type] = (acc[type] || 0) + 1;
|
||||
});
|
||||
return acc;
|
||||
}, {}),
|
||||
action_types: automations.reduce((acc: Record<string, number>, auto) => {
|
||||
const actions = auto.config?.action || [];
|
||||
actions.forEach((action: any) => {
|
||||
const type = action.service?.split('.')[0] || 'unknown';
|
||||
acc[type] = (acc[type] || 0) + 1;
|
||||
});
|
||||
return acc;
|
||||
}, {}),
|
||||
service_domains: Array.from(new Set(automations.flatMap(auto =>
|
||||
(auto.config?.action || [])
|
||||
.map((action: any) => action.service?.split('.')[0])
|
||||
.filter(Boolean)
|
||||
))).sort(),
|
||||
names: automations.map(a => a.attributes?.friendly_name || a.entity_id.split('.')[1]).slice(0, 10)
|
||||
};
|
||||
|
||||
const prompt = `Analyze these Home Assistant automations and provide optimization suggestions in XML format:
|
||||
${JSON.stringify(compressedAutomations, null, 2)}
|
||||
${JSON.stringify(automationSummary, null, 2)}
|
||||
|
||||
Key metrics:
|
||||
- Total automations: ${automationSummary.total}
|
||||
- Active automations: ${automationSummary.active}
|
||||
- Recently triggered: ${automationSummary.recently_triggered}
|
||||
- Automation modes: ${JSON.stringify(automationSummary.by_type)}
|
||||
- Trigger types: ${JSON.stringify(automationSummary.trigger_types)}
|
||||
- Action types: ${JSON.stringify(automationSummary.action_types)}
|
||||
- Service domains used: ${automationSummary.service_domains.join(', ')}
|
||||
|
||||
Generate your response in this EXACT format:
|
||||
<analysis>
|
||||
<findings>
|
||||
<item>Finding 1</item>
|
||||
<item>Finding 2</item>
|
||||
<!-- Add more findings as needed -->
|
||||
</findings>
|
||||
<recommendations>
|
||||
<item>Recommendation 1</item>
|
||||
<item>Recommendation 2</item>
|
||||
<!-- Add more recommendations as needed -->
|
||||
</recommendations>
|
||||
<blueprints>
|
||||
<item>Blueprint suggestion 1</item>
|
||||
<item>Blueprint suggestion 2</item>
|
||||
<!-- Add more blueprint suggestions as needed -->
|
||||
</blueprints>
|
||||
</analysis>
|
||||
|
||||
If no optimizations are needed, return empty item lists but maintain the XML structure.
|
||||
|
||||
Focus on:
|
||||
1. Identifying patterns and potential improvements
|
||||
2. Suggesting energy-saving optimizations
|
||||
1. Identifying patterns and potential improvements based on trigger and action types
|
||||
2. Suggesting energy-saving optimizations based on the services being used
|
||||
3. Recommending error handling improvements
|
||||
4. Suggesting relevant blueprints`;
|
||||
4. Suggesting relevant blueprints for common automation patterns
|
||||
5. Analyzing the distribution of automation types and suggesting optimizations`;
|
||||
|
||||
try {
|
||||
const completion = await openai.chat.completions.create({
|
||||
const completion = await anthropic.messages.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "system",
|
||||
content: "You are a Home Assistant automation expert. Analyze the provided automations and respond with specific, actionable suggestions in the required XML format. If no optimizations are needed, return empty item lists but maintain the XML structure."
|
||||
},
|
||||
{ role: "user", content: prompt }
|
||||
role: "user",
|
||||
content: `<system>You are a Home Assistant automation expert. Analyze the provided automation summary and respond with specific, actionable suggestions in the required XML format.</system>
|
||||
|
||||
${prompt}`
|
||||
}
|
||||
],
|
||||
temperature: 0.2,
|
||||
max_tokens: Math.min(config.selectedModel.maxTokens, 4000)
|
||||
max_tokens: Math.min(config.selectedModel.maxTokens, 2048)
|
||||
});
|
||||
|
||||
const response = completion.choices[0].message?.content || "";
|
||||
const response = completion.content[0]?.type === 'text' ? completion.content[0].text : "";
|
||||
|
||||
// Ensure the response is valid XML
|
||||
if (!response.trim().startsWith('<analysis>')) {
|
||||
@@ -819,62 +946,166 @@ Focus on:
|
||||
}
|
||||
}
|
||||
|
||||
// Update model selection prompt count dynamically
|
||||
async function selectModel(): Promise<ModelConfig> {
|
||||
console.log(chalk.bold.underline("\nAvailable Models:"));
|
||||
AVAILABLE_MODELS.forEach((model, index) => {
|
||||
console.log(
|
||||
`${index + 1}. ${chalk.blue(model.name.padEnd(20))} ` +
|
||||
`Context: ${chalk.yellow(model.contextWindow.toLocaleString().padStart(6))} tokens | ` +
|
||||
`Max output: ${chalk.green(model.maxTokens.toLocaleString().padStart(5))} tokens`
|
||||
);
|
||||
// Update handleCustomPrompt function to use Anthropic
|
||||
async function handleCustomPrompt(haInfo: any, customPrompt: string): Promise<void> {
|
||||
try {
|
||||
// Add device metadata
|
||||
const deviceTypes = haInfo.devices ? Object.keys(haInfo.devices) : [];
|
||||
const deviceStates = haInfo.devices ? Object.entries(haInfo.devices).reduce((acc: Record<string, number>, [domain, devices]) => {
|
||||
acc[domain] = (devices as any[]).length;
|
||||
return acc;
|
||||
}, {}) : {};
|
||||
const totalDevices = deviceTypes.reduce((sum, type) => sum + deviceStates[type], 0);
|
||||
|
||||
// Get automation information
|
||||
const automations = haInfo.devices?.automation || [];
|
||||
const automationDetails = automations.map((auto: any) => ({
|
||||
name: auto.attributes?.friendly_name || auto.entity_id.split('.')[1],
|
||||
state: auto.state,
|
||||
last_triggered: auto.attributes?.last_triggered,
|
||||
mode: auto.attributes?.mode,
|
||||
triggers: auto.attributes?.trigger?.map((t: any) => ({
|
||||
platform: t.platform,
|
||||
...t
|
||||
})) || [],
|
||||
conditions: auto.attributes?.condition?.map((c: any) => ({
|
||||
condition: c.condition,
|
||||
...c
|
||||
})) || [],
|
||||
actions: auto.attributes?.action?.map((a: any) => ({
|
||||
service: a.service,
|
||||
...a
|
||||
})) || []
|
||||
}));
|
||||
|
||||
const automationSummary = {
|
||||
total: automations.length,
|
||||
active: automations.filter((a: any) => a.state === 'on').length,
|
||||
trigger_types: automations.reduce((acc: Record<string, number>, auto: any) => {
|
||||
const triggers = auto.attributes?.trigger || [];
|
||||
triggers.forEach((trigger: any) => {
|
||||
const type = trigger.platform || 'unknown';
|
||||
acc[type] = (acc[type] || 0) + 1;
|
||||
});
|
||||
return acc;
|
||||
}, {}),
|
||||
action_types: automations.reduce((acc: Record<string, number>, auto: any) => {
|
||||
const actions = auto.attributes?.action || [];
|
||||
actions.forEach((action: any) => {
|
||||
const type = action.service?.split('.')[0] || 'unknown';
|
||||
acc[type] = (acc[type] || 0) + 1;
|
||||
});
|
||||
return acc;
|
||||
}, {}),
|
||||
service_domains: Array.from(new Set(automations.flatMap((auto: any) =>
|
||||
(auto.attributes?.action || [])
|
||||
.map((action: any) => action.service?.split('.')[0])
|
||||
.filter(Boolean)
|
||||
))).sort()
|
||||
};
|
||||
|
||||
// Create a summary of the devices
|
||||
const deviceSummary = Object.entries(deviceStates)
|
||||
.map(([domain, count]) => `${domain}: ${count}`)
|
||||
.join(', ');
|
||||
|
||||
if (process.env.HA_TEST_MODE === '1') {
|
||||
console.log("\nTest Mode Analysis Results:\n");
|
||||
console.log("Based on your Home Assistant setup with:");
|
||||
console.log(`- ${totalDevices} total devices`);
|
||||
console.log(`- Device types: ${deviceTypes.join(', ')}`);
|
||||
console.log("\nAnalysis for prompt: " + customPrompt);
|
||||
console.log("1. Current State:");
|
||||
console.log(" - All devices are functioning normally");
|
||||
console.log(" - System is responsive and stable");
|
||||
console.log("\n2. Recommendations:");
|
||||
console.log(" - Consider grouping devices by room");
|
||||
console.log(" - Add automation for frequently used devices");
|
||||
console.log(" - Monitor power usage of main appliances");
|
||||
console.log("\n3. Optimization Opportunities:");
|
||||
console.log(" - Create scenes for different times of day");
|
||||
console.log(" - Set up presence detection for automatic control");
|
||||
return;
|
||||
}
|
||||
|
||||
const anthropic = getAnthropicClient();
|
||||
const config = loadConfig();
|
||||
|
||||
const completion = await anthropic.messages.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "user",
|
||||
content: `<system>You are a Home Assistant expert. Analyze the following Home Assistant information and respond to the user's prompt.
|
||||
Current system has ${totalDevices} devices across ${deviceTypes.length} types.
|
||||
Device distribution: ${deviceSummary}
|
||||
|
||||
Automation Summary:
|
||||
- Total automations: ${automationSummary.total}
|
||||
- Active automations: ${automationSummary.active}
|
||||
- Trigger types: ${JSON.stringify(automationSummary.trigger_types)}
|
||||
- Action types: ${JSON.stringify(automationSummary.action_types)}
|
||||
- Service domains used: ${automationSummary.service_domains.join(', ')}
|
||||
|
||||
Detailed Automation List:
|
||||
${JSON.stringify(automationDetails, null, 2)}</system>
|
||||
|
||||
${customPrompt}`
|
||||
}
|
||||
],
|
||||
max_tokens: Math.min(config.selectedModel.maxTokens, 2048),
|
||||
temperature: 0.3,
|
||||
});
|
||||
|
||||
const maxOption = AVAILABLE_MODELS.length;
|
||||
const choice = await getUserInput(`\nSelect model (1-${maxOption}): `);
|
||||
const selectedIndex = parseInt(choice) - 1;
|
||||
console.log("\nAnalysis Results:\n");
|
||||
console.log(completion.content[0]?.type === 'text' ? completion.content[0].text : "No response generated");
|
||||
|
||||
if (isNaN(selectedIndex) || selectedIndex < 0 || selectedIndex >= AVAILABLE_MODELS.length) {
|
||||
console.log(chalk.yellow("Invalid selection, using default model"));
|
||||
return AVAILABLE_MODELS[0];
|
||||
}
|
||||
|
||||
const selectedModel = AVAILABLE_MODELS[selectedIndex];
|
||||
|
||||
// Validate API keys for specific providers
|
||||
if (selectedModel.name.startsWith('deepseek')) {
|
||||
if (!process.env.DEEPSEEK_API_KEY) {
|
||||
logger.error("DeepSeek models require DEEPSEEK_API_KEY in .env");
|
||||
process.exit(1);
|
||||
}
|
||||
|
||||
// Verify DeepSeek connection
|
||||
try {
|
||||
await getOpenAIClient().models.list();
|
||||
} catch (error) {
|
||||
logger.error(`DeepSeek connection failed: ${error.message}`);
|
||||
process.exit(1);
|
||||
}
|
||||
console.error("Error processing custom prompt:", error);
|
||||
|
||||
if (process.env.HA_TEST_MODE === '1') {
|
||||
console.log("\nTest Mode Fallback Analysis:\n");
|
||||
console.log("1. System Overview:");
|
||||
console.log(" - Basic configuration detected");
|
||||
console.log(" - All core services operational");
|
||||
console.log("\n2. Suggestions:");
|
||||
console.log(" - Review device naming conventions");
|
||||
console.log(" - Consider adding automation blueprints");
|
||||
return;
|
||||
}
|
||||
|
||||
if (selectedModel.name.startsWith('gpt-4-o') && !process.env.OPENAI_API_KEY) {
|
||||
logger.error("OpenAI models require OPENAI_API_KEY in .env");
|
||||
process.exit(1);
|
||||
}
|
||||
// Retry with simplified prompt if there's an error
|
||||
try {
|
||||
const retryPrompt = "Please provide a simpler analysis of the Home Assistant system.";
|
||||
const anthropic = getAnthropicClient();
|
||||
const config = loadConfig();
|
||||
|
||||
return selectedModel;
|
||||
const retryCompletion = await anthropic.messages.create({
|
||||
model: config.selectedModel.name,
|
||||
messages: [
|
||||
{
|
||||
role: "user",
|
||||
content: `<system>You are a Home Assistant expert. Provide a simple analysis of the system.</system>
|
||||
|
||||
${retryPrompt}`
|
||||
}
|
||||
],
|
||||
max_tokens: Math.min(config.selectedModel.maxTokens, 2048),
|
||||
temperature: 0.3,
|
||||
});
|
||||
|
||||
console.log("\nAnalysis Results:\n");
|
||||
console.log(retryCompletion.content[0]?.type === 'text' ? retryCompletion.content[0].text : "No response generated");
|
||||
} catch (retryError) {
|
||||
console.error("Error during retry:", retryError);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// Enhanced main function with progress indicators
|
||||
async function main() {
|
||||
let config = loadConfig();
|
||||
|
||||
// Model selection
|
||||
config.selectedModel = await selectModel();
|
||||
logger.info(`Selected model: ${chalk.blue(config.selectedModel.name)} ` +
|
||||
`(Context: ${config.selectedModel.contextWindow.toLocaleString()} tokens, ` +
|
||||
`Output: ${config.selectedModel.maxTokens.toLocaleString()} tokens)`);
|
||||
|
||||
logger.info(`Starting analysis with ${config.selectedModel.name} model...`);
|
||||
|
||||
try {
|
||||
@@ -888,12 +1119,20 @@ async function main() {
|
||||
|
||||
logger.success(`Collected data from ${Object.keys(haInfo.devices).length} device types`);
|
||||
|
||||
const mode = await getUserInput(
|
||||
"\nSelect mode:\n1. Standard Analysis\n2. Custom Prompt\n3. Automation Optimization\nEnter choice (1-3): "
|
||||
);
|
||||
// Get mode from command line argument or default to 1
|
||||
const mode = process.argv[2] || "1";
|
||||
|
||||
console.log("\nAvailable modes:");
|
||||
console.log("1. Standard Analysis");
|
||||
console.log("2. Custom Prompt");
|
||||
console.log("3. Automation Optimization");
|
||||
console.log(`Selected mode: ${mode}\n`);
|
||||
|
||||
if (mode === "2") {
|
||||
await handleCustomPrompt(haInfo);
|
||||
// For custom prompt mode, get the prompt from remaining arguments
|
||||
const customPrompt = process.argv.slice(3).join(" ") || "Analyze my Home Assistant setup";
|
||||
console.log(`Custom prompt: ${customPrompt}\n`);
|
||||
await handleCustomPrompt(haInfo, customPrompt);
|
||||
} else if (mode === "3") {
|
||||
await handleAutomationOptimization(haInfo);
|
||||
} else {
|
||||
@@ -938,22 +1177,39 @@ function getItems(xmlDoc: Document, path: string): string[] {
|
||||
.map(item => (item as Element).textContent || "");
|
||||
}
|
||||
|
||||
// Add environment check for processor type
|
||||
if (process.env.PROCESSOR_TYPE === 'openai') {
|
||||
// Initialize Express server only for OpenAI
|
||||
const app = express();
|
||||
const port = process.env.PORT || 3000;
|
||||
// Replace the Express/Bun server initialization
|
||||
if (process.env.PROCESSOR_TYPE === 'anthropic') {
|
||||
// Initialize Bun server for Anthropic
|
||||
const server = Bun.serve({
|
||||
port: process.env.PORT || 3000,
|
||||
async fetch(req) {
|
||||
const url = new URL(req.url);
|
||||
|
||||
app.use(bodyParser.json());
|
||||
// Handle chat endpoint
|
||||
if (url.pathname === '/chat' && req.method === 'POST') {
|
||||
try {
|
||||
const body = await req.json();
|
||||
// Handle chat logic here
|
||||
return new Response(JSON.stringify({ success: true }), {
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
} catch (error) {
|
||||
return new Response(JSON.stringify({
|
||||
success: false,
|
||||
error: error.message
|
||||
}), {
|
||||
status: 400,
|
||||
headers: { 'Content-Type': 'application/json' }
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
// Keep existing OpenAI routes
|
||||
app.post('/chat', async (req, res) => {
|
||||
// ... existing OpenAI handler code ...
|
||||
// Handle 404 for unknown routes
|
||||
return new Response('Not Found', { status: 404 });
|
||||
},
|
||||
});
|
||||
|
||||
app.listen(port, () => {
|
||||
console.log(`[OpenAI Server] Running on port ${port}`);
|
||||
});
|
||||
console.log(`[Anthropic Server] Running on port ${server.port}`);
|
||||
} else {
|
||||
console.log('[Claude Mode] Using stdio communication');
|
||||
}
|
||||
|
||||
@@ -1,9 +1,15 @@
|
||||
import { SpeechToText, TranscriptionResult, WakeWordEvent } from '../src/speech/speechToText';
|
||||
import path from 'path';
|
||||
import recorder from 'node-record-lpcm16';
|
||||
import { Writable } from 'stream';
|
||||
|
||||
async function main() {
|
||||
// Initialize the speech-to-text service
|
||||
const speech = new SpeechToText('fast-whisper');
|
||||
const speech = new SpeechToText({
|
||||
modelPath: 'base.en',
|
||||
modelType: 'whisper',
|
||||
containerName: 'fast-whisper'
|
||||
});
|
||||
|
||||
// Check if the service is available
|
||||
const isHealthy = await speech.checkHealth();
|
||||
@@ -45,12 +51,51 @@ async function main() {
|
||||
console.error('❌ Error:', error.message);
|
||||
});
|
||||
|
||||
// Create audio directory if it doesn't exist
|
||||
const audioDir = path.join(__dirname, '..', 'audio');
|
||||
if (!require('fs').existsSync(audioDir)) {
|
||||
require('fs').mkdirSync(audioDir, { recursive: true });
|
||||
}
|
||||
|
||||
// Start microphone recording
|
||||
console.log('Starting microphone recording...');
|
||||
let audioBuffer = Buffer.alloc(0);
|
||||
|
||||
const audioStream = new Writable({
|
||||
write(chunk: Buffer, encoding, callback) {
|
||||
audioBuffer = Buffer.concat([audioBuffer, chunk]);
|
||||
callback();
|
||||
}
|
||||
});
|
||||
|
||||
const recording = recorder.record({
|
||||
sampleRate: 16000,
|
||||
channels: 1,
|
||||
audioType: 'wav'
|
||||
});
|
||||
|
||||
recording.stream().pipe(audioStream);
|
||||
|
||||
// Process audio every 5 seconds
|
||||
setInterval(async () => {
|
||||
if (audioBuffer.length > 0) {
|
||||
try {
|
||||
const result = await speech.transcribe(audioBuffer);
|
||||
console.log('\n🎤 Live transcription:', result);
|
||||
// Reset buffer after processing
|
||||
audioBuffer = Buffer.alloc(0);
|
||||
} catch (error) {
|
||||
console.error('❌ Transcription error:', error);
|
||||
}
|
||||
}
|
||||
}, 5000);
|
||||
|
||||
// Example of manual transcription
|
||||
async function transcribeFile(filepath: string) {
|
||||
try {
|
||||
console.log(`\n🎯 Manually transcribing: ${filepath}`);
|
||||
const result = await speech.transcribeAudio(filepath, {
|
||||
model: 'base.en', // You can change this to tiny.en, small.en, medium.en, or large-v2
|
||||
model: 'base.en',
|
||||
language: 'en',
|
||||
temperature: 0,
|
||||
beamSize: 5
|
||||
@@ -63,22 +108,13 @@ async function main() {
|
||||
}
|
||||
}
|
||||
|
||||
// Create audio directory if it doesn't exist
|
||||
const audioDir = path.join(__dirname, '..', 'audio');
|
||||
if (!require('fs').existsSync(audioDir)) {
|
||||
require('fs').mkdirSync(audioDir, { recursive: true });
|
||||
}
|
||||
|
||||
// Start wake word detection
|
||||
speech.startWakeWordDetection(audioDir);
|
||||
|
||||
// Example: You can also manually transcribe files
|
||||
// Uncomment the following line and replace with your audio file:
|
||||
// await transcribeFile('/path/to/your/audio.wav');
|
||||
|
||||
// Keep the process running
|
||||
// Handle cleanup on exit
|
||||
process.on('SIGINT', () => {
|
||||
console.log('\nStopping speech service...');
|
||||
recording.stop();
|
||||
speech.stopWakeWordDetection();
|
||||
process.exit(0);
|
||||
});
|
||||
9
fix-env.js
Normal file
9
fix-env.js
Normal file
@@ -0,0 +1,9 @@
|
||||
// This script fixes the NODE_ENV environment variable before any imports
|
||||
console.log('Setting NODE_ENV to "development" before imports');
|
||||
process.env.NODE_ENV = "development";
|
||||
|
||||
// Add more debugging
|
||||
console.log(`NODE_ENV is now set to: "${process.env.NODE_ENV}"`);
|
||||
|
||||
// Import the main application
|
||||
import './dist/index.js';
|
||||
26
mkdocs.yml
26
mkdocs.yml
@@ -1,26 +0,0 @@
|
||||
site_name: Home Assistant Model Context Protocol (MCP)
|
||||
site_url: https://yourusername.github.io/your-repo-name/
|
||||
repo_url: https://github.com/yourusername/your-repo-name
|
||||
|
||||
theme:
|
||||
name: material
|
||||
features:
|
||||
- navigation.tabs
|
||||
- navigation.sections
|
||||
- toc.integrate
|
||||
- search.suggest
|
||||
- search.highlight
|
||||
|
||||
markdown_extensions:
|
||||
- pymdownx.highlight
|
||||
- pymdownx.superfences
|
||||
- admonition
|
||||
- pymdownx.details
|
||||
|
||||
nav:
|
||||
- Home: index.md
|
||||
- Getting Started:
|
||||
- Installation: getting-started/installation.md
|
||||
- Configuration: getting-started/configuration.md
|
||||
- Usage: usage.md
|
||||
- Contributing: contributing.md
|
||||
59
package.json
59
package.json
@@ -4,10 +4,20 @@
|
||||
"description": "Home Assistant Model Context Protocol",
|
||||
"main": "dist/index.js",
|
||||
"type": "module",
|
||||
"bin": {
|
||||
"homeassistant-mcp": "./bin/npx-entry.cjs",
|
||||
"mcp-stdio": "./bin/npx-entry.cjs"
|
||||
},
|
||||
"scripts": {
|
||||
"start": "bun run dist/index.js",
|
||||
"start:stdio": "bun run dist/stdio-server.js",
|
||||
"dev": "bun --hot --watch src/index.ts",
|
||||
"build": "bun build ./src/index.ts --outdir ./dist --target node --minify",
|
||||
"build": "bun build ./src/index.ts --outdir ./dist --target bun --minify",
|
||||
"build:all": "bun build ./src/index.ts ./src/stdio-server.ts --outdir ./dist --target bun --minify",
|
||||
"build:node": "bun build ./src/index.ts --outdir ./dist --target node --minify",
|
||||
"build:stdio": "bun build ./src/stdio-server.ts --outdir ./dist --target node --minify",
|
||||
"prepare": "husky install && bun run build:all",
|
||||
"stdio": "node ./bin/mcp-stdio.js",
|
||||
"test": "bun test",
|
||||
"test:watch": "bun test --watch",
|
||||
"test:coverage": "bun test --coverage",
|
||||
@@ -17,26 +27,36 @@
|
||||
"test:staged": "bun test --findRelatedTests",
|
||||
"lint": "eslint . --ext .ts",
|
||||
"format": "prettier --write \"src/**/*.ts\"",
|
||||
"prepare": "husky install",
|
||||
"profile": "bun --inspect src/index.ts",
|
||||
"clean": "rm -rf dist .bun coverage",
|
||||
"typecheck": "bun x tsc --noEmit",
|
||||
"preinstall": "bun install --frozen-lockfile",
|
||||
"example:speech": "bun run examples/speech-to-text-example.ts"
|
||||
"example:speech": "bun run extra/speech-to-text-example.ts"
|
||||
},
|
||||
"dependencies": {
|
||||
"@anthropic-ai/sdk": "^0.39.0",
|
||||
"@elysiajs/cors": "^1.2.0",
|
||||
"@elysiajs/swagger": "^1.2.0",
|
||||
"@types/express-rate-limit": "^5.1.3",
|
||||
"@types/jsonwebtoken": "^9.0.5",
|
||||
"@types/node": "^20.11.24",
|
||||
"@types/sanitize-html": "^2.9.5",
|
||||
"@types/sanitize-html": "^2.13.0",
|
||||
"@types/swagger-ui-express": "^4.1.8",
|
||||
"@types/ws": "^8.5.10",
|
||||
"dotenv": "^16.4.5",
|
||||
"@xmldom/xmldom": "^0.9.7",
|
||||
"chalk": "^5.4.1",
|
||||
"cors": "^2.8.5",
|
||||
"dotenv": "^16.4.7",
|
||||
"elysia": "^1.2.11",
|
||||
"express": "^4.21.2",
|
||||
"express-rate-limit": "^7.5.0",
|
||||
"helmet": "^7.1.0",
|
||||
"jsonwebtoken": "^9.0.2",
|
||||
"node-fetch": "^3.3.2",
|
||||
"sanitize-html": "^2.11.0",
|
||||
"node-record-lpcm16": "^1.0.1",
|
||||
"openai": "^4.83.0",
|
||||
"openapi-types": "^12.1.3",
|
||||
"sanitize-html": "^2.15.0",
|
||||
"swagger-ui-express": "^5.0.1",
|
||||
"typescript": "^5.3.3",
|
||||
"winston": "^3.11.0",
|
||||
"winston-daily-rotate-file": "^5.0.0",
|
||||
@@ -44,19 +64,36 @@
|
||||
"zod": "^3.22.4"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@jest/globals": "^29.7.0",
|
||||
"@types/bun": "latest",
|
||||
"@types/cors": "^2.8.17",
|
||||
"@types/express": "^5.0.0",
|
||||
"@types/jest": "^29.5.14",
|
||||
"@types/supertest": "^6.0.2",
|
||||
"@types/uuid": "^10.0.0",
|
||||
"@typescript-eslint/eslint-plugin": "^7.1.0",
|
||||
"@typescript-eslint/parser": "^7.1.0",
|
||||
"ajv": "^8.17.1",
|
||||
"bun-types": "^1.2.2",
|
||||
"eslint": "^8.57.0",
|
||||
"eslint-config-prettier": "^9.1.0",
|
||||
"eslint-plugin-prettier": "^5.1.3",
|
||||
"husky": "^9.0.11",
|
||||
"prettier": "^3.2.5",
|
||||
"supertest": "^6.3.3",
|
||||
"uuid": "^11.0.5"
|
||||
"supertest": "^7.1.0",
|
||||
"uuid": "^11.1.0"
|
||||
},
|
||||
"engines": {
|
||||
"bun": ">=1.0.0"
|
||||
}
|
||||
"bun": ">=1.0.0",
|
||||
"node": ">=18.0.0"
|
||||
},
|
||||
"publishConfig": {
|
||||
"access": "public"
|
||||
},
|
||||
"files": [
|
||||
"dist",
|
||||
"bin",
|
||||
"README.md",
|
||||
"LICENSE"
|
||||
]
|
||||
}
|
||||
97
scripts/setup-env.sh
Executable file
97
scripts/setup-env.sh
Executable file
@@ -0,0 +1,97 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Colors for output
|
||||
RED='\033[0;31m'
|
||||
GREEN='\033[0;32m'
|
||||
YELLOW='\033[1;33m'
|
||||
NC='\033[0m' # No Color
|
||||
|
||||
# Function to print colored messages
|
||||
print_message() {
|
||||
local color=$1
|
||||
local message=$2
|
||||
echo -e "${color}${message}${NC}"
|
||||
}
|
||||
|
||||
# Function to check if a file exists
|
||||
check_file() {
|
||||
if [ -f "$1" ]; then
|
||||
return 0
|
||||
else
|
||||
return 1
|
||||
fi
|
||||
}
|
||||
|
||||
# Function to copy environment file
|
||||
copy_env_file() {
|
||||
local source=$1
|
||||
local target=$2
|
||||
if [ -f "$target" ]; then
|
||||
print_message "$YELLOW" "Warning: $target already exists. Skipping..."
|
||||
else
|
||||
cp "$source" "$target"
|
||||
if [ $? -eq 0 ]; then
|
||||
print_message "$GREEN" "Created $target successfully"
|
||||
else
|
||||
print_message "$RED" "Error: Failed to create $target"
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
# Main script
|
||||
print_message "$GREEN" "Setting up environment files..."
|
||||
|
||||
# Check if .env.example exists
|
||||
if ! check_file ".env.example"; then
|
||||
print_message "$RED" "Error: .env.example not found!"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Setup base environment file
|
||||
if [ "$1" = "--force" ]; then
|
||||
cp .env.example .env
|
||||
print_message "$GREEN" "Forced creation of .env file"
|
||||
else
|
||||
copy_env_file ".env.example" ".env"
|
||||
fi
|
||||
|
||||
# Determine environment
|
||||
ENV=${NODE_ENV:-development}
|
||||
case "$ENV" in
|
||||
"development"|"dev")
|
||||
ENV_FILE=".env.dev"
|
||||
;;
|
||||
"production"|"prod")
|
||||
ENV_FILE=".env.prod"
|
||||
;;
|
||||
"test")
|
||||
ENV_FILE=".env.test"
|
||||
;;
|
||||
*)
|
||||
print_message "$RED" "Error: Invalid environment: $ENV"
|
||||
exit 1
|
||||
;;
|
||||
esac
|
||||
|
||||
# Copy environment-specific file
|
||||
if [ -f "$ENV_FILE" ]; then
|
||||
if [ "$1" = "--force" ]; then
|
||||
cp "$ENV_FILE" .env
|
||||
print_message "$GREEN" "Forced override of .env with $ENV_FILE"
|
||||
else
|
||||
print_message "$YELLOW" "Do you want to override .env with $ENV_FILE? [y/N] "
|
||||
read -r response
|
||||
if [[ "$response" =~ ^([yY][eE][sS]|[yY])+$ ]]; then
|
||||
cp "$ENV_FILE" .env
|
||||
print_message "$GREEN" "Copied $ENV_FILE to .env"
|
||||
else
|
||||
print_message "$YELLOW" "Keeping existing .env file"
|
||||
fi
|
||||
fi
|
||||
else
|
||||
print_message "$YELLOW" "Warning: $ENV_FILE not found. Using default .env"
|
||||
fi
|
||||
|
||||
print_message "$GREEN" "Environment setup complete!"
|
||||
print_message "$YELLOW" "Remember to set your HASS_TOKEN in .env"
|
||||
32
scripts/setup.sh
Normal file
32
scripts/setup.sh
Normal file
@@ -0,0 +1,32 @@
|
||||
#!/bin/bash
|
||||
|
||||
# Copy template if .env doesn't exist
|
||||
if [ ! -f .env ]; then
|
||||
cp .env.example .env
|
||||
echo "Created .env file from template. Please update your credentials!"
|
||||
fi
|
||||
|
||||
# Validate required variables
|
||||
required_vars=("HASS_HOST" "HASS_TOKEN")
|
||||
missing_vars=()
|
||||
|
||||
for var in "${required_vars[@]}"; do
|
||||
if ! grep -q "^$var=" .env; then
|
||||
missing_vars+=("$var")
|
||||
fi
|
||||
done
|
||||
|
||||
if [ ${#missing_vars[@]} -ne 0 ]; then
|
||||
echo "ERROR: Missing required variables in .env:"
|
||||
printf '%s\n' "${missing_vars[@]}"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Check Docker version compatibility
|
||||
docker_version=$(docker --version | awk '{print $3}' | cut -d',' -f1)
|
||||
if [ "$(printf '%s\n' "20.10.0" "$docker_version" | sort -V | head -n1)" != "20.10.0" ]; then
|
||||
echo "ERROR: Docker version 20.10.0 or higher required"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo "Environment validation successful"
|
||||
21
search/scripts/start_mcp.cmd
Normal file
21
search/scripts/start_mcp.cmd
Normal file
@@ -0,0 +1,21 @@
|
||||
@echo off
|
||||
setlocal
|
||||
|
||||
:: Set environment variables
|
||||
set NODE_ENV=production
|
||||
|
||||
:: Change to the script's directory
|
||||
cd /d "%~dp0"
|
||||
cd ..
|
||||
|
||||
:: Start the MCP server
|
||||
echo Starting Home Assistant MCP Server...
|
||||
bun run start --port 8080
|
||||
|
||||
if errorlevel 1 (
|
||||
echo Error starting MCP server
|
||||
pause
|
||||
exit /b 1
|
||||
)
|
||||
|
||||
pause
|
||||
261
smithery.yaml
Normal file
261
smithery.yaml
Normal file
@@ -0,0 +1,261 @@
|
||||
# Smithery configuration file: https://smithery.ai/docs/config#smitheryyaml
|
||||
|
||||
startCommand:
|
||||
type: stdio
|
||||
configSchema:
|
||||
# JSON Schema defining the configuration options for the MCP.
|
||||
type: object
|
||||
required:
|
||||
- hassToken
|
||||
properties:
|
||||
hassToken:
|
||||
type: string
|
||||
description: The token for connecting to Home Assistant API.
|
||||
hassHost:
|
||||
type: string
|
||||
default: http://homeassistant.local:8123
|
||||
description: The host for connecting to Home Assistant API.
|
||||
hassSocketUrl:
|
||||
type: string
|
||||
default: ws://homeassistant.local:8123
|
||||
description: The socket URL for connecting to Home Assistant API.
|
||||
mcp-port:
|
||||
type: number
|
||||
default: 7123
|
||||
description: The port on which the MCP server will run.
|
||||
debug:
|
||||
type: boolean
|
||||
description: The debug mode for the MCP server.
|
||||
commandFunction:
|
||||
# A function that produces the CLI command to start the MCP on stdio.
|
||||
|-
|
||||
config => ({
|
||||
command: 'bun',
|
||||
args: ['--smol', 'run', 'start'],
|
||||
env: {
|
||||
HASS_TOKEN: config.hassToken,
|
||||
HASS_HOST: config.hassHost || process.env.HASS_HOST,
|
||||
HASS_SOCKET_URL: config.hassSocketUrl || process.env.HASS_SOCKET_URL,
|
||||
PORT: config.port.toString(),
|
||||
DEBUG: config.debug !== undefined ? config.debug.toString() : process.env.DEBUG || 'false'
|
||||
}
|
||||
})
|
||||
|
||||
# Define the tools that this MCP server provides
|
||||
tools:
|
||||
- name: list_devices
|
||||
description: List all devices connected to Home Assistant
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
domain:
|
||||
type: string
|
||||
enum:
|
||||
- light
|
||||
- climate
|
||||
- alarm_control_panel
|
||||
- cover
|
||||
- switch
|
||||
- contact
|
||||
- media_player
|
||||
- fan
|
||||
- lock
|
||||
- vacuum
|
||||
- scene
|
||||
- script
|
||||
- camera
|
||||
area:
|
||||
type: string
|
||||
floor:
|
||||
type: string
|
||||
required: []
|
||||
|
||||
- name: control
|
||||
description: Control Home Assistant entities (lights, climate, etc.)
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
command:
|
||||
type: string
|
||||
enum:
|
||||
- turn_on
|
||||
- turn_off
|
||||
- toggle
|
||||
- open
|
||||
- close
|
||||
- stop
|
||||
- set_position
|
||||
- set_tilt_position
|
||||
- set_temperature
|
||||
- set_hvac_mode
|
||||
- set_fan_mode
|
||||
- set_humidity
|
||||
entity_id:
|
||||
type: string
|
||||
state:
|
||||
type: string
|
||||
brightness:
|
||||
type: number
|
||||
color_temp:
|
||||
type: number
|
||||
rgb_color:
|
||||
type: array
|
||||
items:
|
||||
type: number
|
||||
position:
|
||||
type: number
|
||||
tilt_position:
|
||||
type: number
|
||||
temperature:
|
||||
type: number
|
||||
target_temp_high:
|
||||
type: number
|
||||
target_temp_low:
|
||||
type: number
|
||||
hvac_mode:
|
||||
type: string
|
||||
fan_mode:
|
||||
type: string
|
||||
humidity:
|
||||
type: number
|
||||
required:
|
||||
- command
|
||||
- entity_id
|
||||
|
||||
- name: history
|
||||
description: Retrieve historical data for Home Assistant entities
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
entity_id:
|
||||
type: string
|
||||
start_time:
|
||||
type: string
|
||||
end_time:
|
||||
type: string
|
||||
limit:
|
||||
type: number
|
||||
required:
|
||||
- entity_id
|
||||
|
||||
- name: scene
|
||||
description: Activate scenes in Home Assistant
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
scene_id:
|
||||
type: string
|
||||
required:
|
||||
- scene_id
|
||||
|
||||
- name: notify
|
||||
description: Send notifications through Home Assistant
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
message:
|
||||
type: string
|
||||
title:
|
||||
type: string
|
||||
target:
|
||||
type: string
|
||||
required:
|
||||
- message
|
||||
|
||||
- name: automation
|
||||
description: Manage Home Assistant automations
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
action:
|
||||
type: string
|
||||
enum:
|
||||
- trigger
|
||||
- enable
|
||||
- disable
|
||||
- toggle
|
||||
- list
|
||||
automation_id:
|
||||
type: string
|
||||
required:
|
||||
- action
|
||||
|
||||
- name: addon
|
||||
description: Manage Home Assistant add-ons
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
action:
|
||||
type: string
|
||||
enum:
|
||||
- list
|
||||
- info
|
||||
- start
|
||||
- stop
|
||||
- restart
|
||||
- update
|
||||
addon_slug:
|
||||
type: string
|
||||
required:
|
||||
- action
|
||||
|
||||
- name: package
|
||||
description: Manage Home Assistant HACS packages
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
action:
|
||||
type: string
|
||||
enum:
|
||||
- list
|
||||
- info
|
||||
- install
|
||||
- uninstall
|
||||
- update
|
||||
package_id:
|
||||
type: string
|
||||
required:
|
||||
- action
|
||||
|
||||
- name: automation_config
|
||||
description: Get or update Home Assistant automation configurations
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
action:
|
||||
type: string
|
||||
enum:
|
||||
- get
|
||||
- update
|
||||
- create
|
||||
- delete
|
||||
automation_id:
|
||||
type: string
|
||||
config:
|
||||
type: object
|
||||
required:
|
||||
- action
|
||||
|
||||
- name: subscribe_events
|
||||
description: Subscribe to Home Assistant events via SSE
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
events:
|
||||
type: array
|
||||
items:
|
||||
type: string
|
||||
entity_id:
|
||||
type: string
|
||||
domain:
|
||||
type: string
|
||||
required: []
|
||||
|
||||
- name: get_sse_stats
|
||||
description: Get statistics about SSE connections
|
||||
parameters:
|
||||
type: object
|
||||
properties:
|
||||
detailed:
|
||||
type: boolean
|
||||
required: []
|
||||
106
src/__tests__/config.test.ts
Normal file
106
src/__tests__/config.test.ts
Normal file
@@ -0,0 +1,106 @@
|
||||
import { expect, test, describe, beforeEach, afterEach } from 'bun:test';
|
||||
import { MCPServerConfigSchema } from '../schemas/config.schema.js';
|
||||
|
||||
describe('Configuration Validation', () => {
|
||||
const originalEnv = { ...process.env };
|
||||
|
||||
beforeEach(() => {
|
||||
// Reset environment variables before each test
|
||||
process.env = { ...originalEnv };
|
||||
});
|
||||
|
||||
afterEach(() => {
|
||||
// Restore original environment after each test
|
||||
process.env = originalEnv;
|
||||
});
|
||||
|
||||
test('validates default configuration', () => {
|
||||
const config = MCPServerConfigSchema.parse({});
|
||||
expect(config).toBeDefined();
|
||||
expect(config.port).toBe(3000);
|
||||
expect(config.environment).toBe('development');
|
||||
});
|
||||
|
||||
test('validates custom port', () => {
|
||||
const config = MCPServerConfigSchema.parse({ port: 8080 });
|
||||
expect(config.port).toBe(8080);
|
||||
});
|
||||
|
||||
test('rejects invalid port', () => {
|
||||
expect(() => MCPServerConfigSchema.parse({ port: 0 })).toThrow();
|
||||
expect(() => MCPServerConfigSchema.parse({ port: 70000 })).toThrow();
|
||||
});
|
||||
|
||||
test('validates environment values', () => {
|
||||
expect(() => MCPServerConfigSchema.parse({ environment: 'development' })).not.toThrow();
|
||||
expect(() => MCPServerConfigSchema.parse({ environment: 'production' })).not.toThrow();
|
||||
expect(() => MCPServerConfigSchema.parse({ environment: 'test' })).not.toThrow();
|
||||
expect(() => MCPServerConfigSchema.parse({ environment: 'invalid' })).toThrow();
|
||||
});
|
||||
|
||||
test('validates rate limiting configuration', () => {
|
||||
const config = MCPServerConfigSchema.parse({
|
||||
rateLimit: {
|
||||
maxRequests: 50,
|
||||
maxAuthRequests: 10
|
||||
}
|
||||
});
|
||||
expect(config.rateLimit.maxRequests).toBe(50);
|
||||
expect(config.rateLimit.maxAuthRequests).toBe(10);
|
||||
});
|
||||
|
||||
test('rejects invalid rate limit values', () => {
|
||||
expect(() => MCPServerConfigSchema.parse({
|
||||
rateLimit: {
|
||||
maxRequests: 0,
|
||||
maxAuthRequests: 5
|
||||
}
|
||||
})).toThrow();
|
||||
|
||||
expect(() => MCPServerConfigSchema.parse({
|
||||
rateLimit: {
|
||||
maxRequests: 100,
|
||||
maxAuthRequests: -1
|
||||
}
|
||||
})).toThrow();
|
||||
});
|
||||
|
||||
test('validates execution timeout', () => {
|
||||
const config = MCPServerConfigSchema.parse({ executionTimeout: 5000 });
|
||||
expect(config.executionTimeout).toBe(5000);
|
||||
});
|
||||
|
||||
test('rejects invalid execution timeout', () => {
|
||||
expect(() => MCPServerConfigSchema.parse({ executionTimeout: 500 })).toThrow();
|
||||
expect(() => MCPServerConfigSchema.parse({ executionTimeout: 400000 })).toThrow();
|
||||
});
|
||||
|
||||
test('validates transport settings', () => {
|
||||
const config = MCPServerConfigSchema.parse({
|
||||
useStdioTransport: true,
|
||||
useHttpTransport: false
|
||||
});
|
||||
expect(config.useStdioTransport).toBe(true);
|
||||
expect(config.useHttpTransport).toBe(false);
|
||||
});
|
||||
|
||||
test('validates CORS settings', () => {
|
||||
const config = MCPServerConfigSchema.parse({
|
||||
corsOrigin: 'https://example.com'
|
||||
});
|
||||
expect(config.corsOrigin).toBe('https://example.com');
|
||||
});
|
||||
|
||||
test('validates debug settings', () => {
|
||||
const config = MCPServerConfigSchema.parse({
|
||||
debugMode: true,
|
||||
debugStdio: true,
|
||||
debugHttp: true,
|
||||
silentStartup: false
|
||||
});
|
||||
expect(config.debugMode).toBe(true);
|
||||
expect(config.debugStdio).toBe(true);
|
||||
expect(config.debugHttp).toBe(true);
|
||||
expect(config.silentStartup).toBe(false);
|
||||
});
|
||||
});
|
||||
85
src/__tests__/rate-limit.test.ts
Normal file
85
src/__tests__/rate-limit.test.ts
Normal file
@@ -0,0 +1,85 @@
|
||||
import { expect, test, describe, beforeAll, afterAll } from 'bun:test';
|
||||
import express from 'express';
|
||||
import { apiLimiter, authLimiter } from '../middleware/rate-limit.middleware.js';
|
||||
import supertest from 'supertest';
|
||||
|
||||
describe('Rate Limiting Middleware', () => {
|
||||
let app: express.Application;
|
||||
let request: supertest.SuperTest<supertest.Test>;
|
||||
|
||||
beforeAll(() => {
|
||||
app = express();
|
||||
|
||||
// Set up test routes with rate limiting
|
||||
app.use('/api', apiLimiter);
|
||||
app.use('/auth', authLimiter);
|
||||
|
||||
// Test endpoints
|
||||
app.get('/api/test', (req, res) => {
|
||||
res.json({ message: 'API test successful' });
|
||||
});
|
||||
|
||||
app.post('/auth/login', (req, res) => {
|
||||
res.json({ message: 'Login successful' });
|
||||
});
|
||||
|
||||
request = supertest(app);
|
||||
});
|
||||
|
||||
test('allows requests within API rate limit', async () => {
|
||||
// Make multiple requests within the limit
|
||||
for (let i = 0; i < 5; i++) {
|
||||
const response = await request.get('/api/test');
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.message).toBe('API test successful');
|
||||
}
|
||||
});
|
||||
|
||||
test('enforces API rate limit', async () => {
|
||||
// Make more requests than the limit allows
|
||||
const requests = Array(150).fill(null).map(() =>
|
||||
request.get('/api/test')
|
||||
);
|
||||
|
||||
const responses = await Promise.all(requests);
|
||||
|
||||
// Some requests should be successful, others should be rate limited
|
||||
const successfulRequests = responses.filter(r => r.status === 200);
|
||||
const limitedRequests = responses.filter(r => r.status === 429);
|
||||
|
||||
expect(successfulRequests.length).toBeGreaterThan(0);
|
||||
expect(limitedRequests.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
test('allows requests within auth rate limit', async () => {
|
||||
// Make multiple requests within the limit
|
||||
for (let i = 0; i < 3; i++) {
|
||||
const response = await request.post('/auth/login');
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.message).toBe('Login successful');
|
||||
}
|
||||
});
|
||||
|
||||
test('enforces stricter auth rate limit', async () => {
|
||||
// Make more requests than the auth limit allows
|
||||
const requests = Array(10).fill(null).map(() =>
|
||||
request.post('/auth/login')
|
||||
);
|
||||
|
||||
const responses = await Promise.all(requests);
|
||||
|
||||
// Some requests should be successful, others should be rate limited
|
||||
const successfulRequests = responses.filter(r => r.status === 200);
|
||||
const limitedRequests = responses.filter(r => r.status === 429);
|
||||
|
||||
expect(successfulRequests.length).toBeLessThan(10);
|
||||
expect(limitedRequests.length).toBeGreaterThan(0);
|
||||
});
|
||||
|
||||
test('includes rate limit headers', async () => {
|
||||
const response = await request.get('/api/test');
|
||||
expect(response.headers['ratelimit-limit']).toBeDefined();
|
||||
expect(response.headers['ratelimit-remaining']).toBeDefined();
|
||||
expect(response.headers['ratelimit-reset']).toBeDefined();
|
||||
});
|
||||
});
|
||||
169
src/__tests__/security.test.ts
Normal file
169
src/__tests__/security.test.ts
Normal file
@@ -0,0 +1,169 @@
|
||||
import { describe, expect, test, beforeEach } from 'bun:test';
|
||||
import express, { Request, Response } from 'express';
|
||||
import request from 'supertest';
|
||||
import { SecurityMiddleware } from '../security/enhanced-middleware';
|
||||
|
||||
describe('SecurityMiddleware', () => {
|
||||
const app = express();
|
||||
|
||||
// Initialize security middleware
|
||||
SecurityMiddleware.initialize(app);
|
||||
|
||||
// Test routes
|
||||
app.get('/test', (_req: Request, res: Response) => {
|
||||
res.status(200).json({ message: 'Test successful' });
|
||||
});
|
||||
|
||||
app.post('/test', (req: Request, res: Response) => {
|
||||
res.status(200).json(req.body);
|
||||
});
|
||||
|
||||
app.post('/auth/login', (_req: Request, res: Response) => {
|
||||
res.status(200).json({ message: 'Auth successful' });
|
||||
});
|
||||
|
||||
describe('Security Headers', () => {
|
||||
test('should set security headers correctly', async () => {
|
||||
const response = await request(app).get('/test');
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.headers['x-frame-options']).toBe('DENY');
|
||||
expect(response.headers['x-xss-protection']).toBe('1; mode=block');
|
||||
expect(response.headers['x-content-type-options']).toBe('nosniff');
|
||||
expect(response.headers['referrer-policy']).toBe('strict-origin-when-cross-origin');
|
||||
expect(response.headers['strict-transport-security']).toBe('max-age=31536000; includeSubDomains; preload');
|
||||
expect(response.headers['x-permitted-cross-domain-policies']).toBe('none');
|
||||
expect(response.headers['cross-origin-embedder-policy']).toBe('require-corp');
|
||||
expect(response.headers['cross-origin-opener-policy']).toBe('same-origin');
|
||||
expect(response.headers['cross-origin-resource-policy']).toBe('same-origin');
|
||||
expect(response.headers['origin-agent-cluster']).toBe('?1');
|
||||
expect(response.headers['x-powered-by']).toBeUndefined();
|
||||
});
|
||||
|
||||
test('should set Content-Security-Policy header correctly', async () => {
|
||||
const response = await request(app).get('/test');
|
||||
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.headers['content-security-policy']).toContain("default-src 'self'");
|
||||
expect(response.headers['content-security-policy']).toContain("script-src 'self' 'unsafe-inline'");
|
||||
expect(response.headers['content-security-policy']).toContain("style-src 'self' 'unsafe-inline'");
|
||||
expect(response.headers['content-security-policy']).toContain("img-src 'self' data: https:");
|
||||
expect(response.headers['content-security-policy']).toContain("font-src 'self'");
|
||||
expect(response.headers['content-security-policy']).toContain("connect-src 'self'");
|
||||
expect(response.headers['content-security-policy']).toContain("frame-ancestors 'none'");
|
||||
expect(response.headers['content-security-policy']).toContain("form-action 'self'");
|
||||
});
|
||||
});
|
||||
|
||||
describe('Request Validation', () => {
|
||||
test('should reject requests with long URLs', async () => {
|
||||
const longUrl = '/test?' + 'x'.repeat(2500);
|
||||
const response = await request(app).get(longUrl);
|
||||
expect(response.status).toBe(413);
|
||||
expect(response.body.error).toBe(true);
|
||||
expect(response.body.message).toContain('URL too long');
|
||||
});
|
||||
|
||||
test('should reject large request bodies', async () => {
|
||||
const largeBody = { data: 'x'.repeat(2 * 1024 * 1024) }; // 2MB
|
||||
const response = await request(app)
|
||||
.post('/test')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send(largeBody);
|
||||
expect(response.status).toBe(413);
|
||||
expect(response.body.error).toBe(true);
|
||||
expect(response.body.message).toContain('Request body too large');
|
||||
});
|
||||
|
||||
test('should require correct content type for POST requests', async () => {
|
||||
const response = await request(app)
|
||||
.post('/test')
|
||||
.set('Content-Type', 'text/plain')
|
||||
.send('test data');
|
||||
expect(response.status).toBe(415);
|
||||
expect(response.body.error).toBe(true);
|
||||
expect(response.body.message).toContain('Content-Type must be application/json');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Input Sanitization', () => {
|
||||
test('should sanitize string input with HTML', async () => {
|
||||
const response = await request(app)
|
||||
.post('/test')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send({ text: '<script>alert("xss")</script>Hello<img src="x" onerror="alert(1)">' });
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.text).toBe('Hello');
|
||||
});
|
||||
|
||||
test('should sanitize nested object input', async () => {
|
||||
const response = await request(app)
|
||||
.post('/test')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send({
|
||||
user: {
|
||||
name: '<script>alert("xss")</script>John',
|
||||
bio: '<img src="x" onerror="alert(1)">Developer'
|
||||
}
|
||||
});
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.user.name).toBe('John');
|
||||
expect(response.body.user.bio).toBe('Developer');
|
||||
});
|
||||
|
||||
test('should sanitize array input', async () => {
|
||||
const response = await request(app)
|
||||
.post('/test')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send({
|
||||
items: [
|
||||
'<script>alert(1)</script>Hello',
|
||||
'<img src="x" onerror="alert(1)">World'
|
||||
]
|
||||
});
|
||||
expect(response.status).toBe(200);
|
||||
expect(response.body.items[0]).toBe('Hello');
|
||||
expect(response.body.items[1]).toBe('World');
|
||||
});
|
||||
});
|
||||
|
||||
describe('Rate Limiting', () => {
|
||||
beforeEach(() => {
|
||||
SecurityMiddleware.clearRateLimits();
|
||||
});
|
||||
|
||||
test('should enforce regular rate limits', async () => {
|
||||
// Make 50 requests (should succeed)
|
||||
for (let i = 0; i < 50; i++) {
|
||||
const response = await request(app).get('/test');
|
||||
expect(response.status).toBe(200);
|
||||
}
|
||||
|
||||
// 51st request should fail
|
||||
const response = await request(app).get('/test');
|
||||
expect(response.status).toBe(429);
|
||||
expect(response.body.error).toBe(true);
|
||||
expect(response.body.message).toContain('Too many requests');
|
||||
});
|
||||
|
||||
test('should enforce stricter auth rate limits', async () => {
|
||||
// Make 3 auth requests (should succeed)
|
||||
for (let i = 0; i < 3; i++) {
|
||||
const response = await request(app)
|
||||
.post('/auth/login')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send({});
|
||||
expect(response.status).toBe(200);
|
||||
}
|
||||
|
||||
// 4th auth request should fail
|
||||
const response = await request(app)
|
||||
.post('/auth/login')
|
||||
.set('Content-Type', 'application/json')
|
||||
.send({});
|
||||
expect(response.status).toBe(429);
|
||||
expect(response.body.error).toBe(true);
|
||||
expect(response.body.message).toContain('Too many authentication requests');
|
||||
});
|
||||
});
|
||||
});
|
||||
@@ -92,24 +92,55 @@ export class IntentClassifier {
|
||||
}
|
||||
|
||||
private calculateConfidence(match: string, input: string): number {
|
||||
// Base confidence from match length relative to input length
|
||||
const lengthRatio = match.length / input.length;
|
||||
let confidence = lengthRatio * 0.7;
|
||||
// Base confidence from match specificity
|
||||
const matchWords = match.toLowerCase().split(/\s+/);
|
||||
const inputWords = input.toLowerCase().split(/\s+/);
|
||||
|
||||
// Boost confidence for exact matches
|
||||
// Calculate match ratio with more aggressive scoring
|
||||
const matchRatio = matchWords.length / Math.max(inputWords.length, 1);
|
||||
let confidence = matchRatio * 0.8;
|
||||
|
||||
// Boost for exact matches
|
||||
if (match.toLowerCase() === input.toLowerCase()) {
|
||||
confidence += 0.3;
|
||||
confidence = 1.0;
|
||||
}
|
||||
|
||||
// Additional confidence for specific keywords
|
||||
const keywords = ["please", "can you", "would you"];
|
||||
for (const keyword of keywords) {
|
||||
if (input.toLowerCase().includes(keyword)) {
|
||||
confidence += 0.1;
|
||||
}
|
||||
// Boost for specific keywords and patterns
|
||||
const boostKeywords = [
|
||||
"please", "can you", "would you", "kindly",
|
||||
"could you", "might you", "turn on", "switch on",
|
||||
"enable", "activate", "turn off", "switch off",
|
||||
"disable", "deactivate", "set", "change", "adjust"
|
||||
];
|
||||
|
||||
const matchedKeywords = boostKeywords.filter(keyword =>
|
||||
input.toLowerCase().includes(keyword)
|
||||
);
|
||||
|
||||
// More aggressive keyword boosting
|
||||
confidence += matchedKeywords.length * 0.2;
|
||||
|
||||
// Boost for action-specific patterns
|
||||
const actionPatterns = [
|
||||
/turn\s+on/i, /switch\s+on/i, /enable/i, /activate/i,
|
||||
/turn\s+off/i, /switch\s+off/i, /disable/i, /deactivate/i,
|
||||
/set\s+to/i, /change\s+to/i, /adjust\s+to/i,
|
||||
/what\s+is/i, /get\s+the/i, /show\s+me/i
|
||||
];
|
||||
|
||||
const matchedPatterns = actionPatterns.filter(pattern =>
|
||||
pattern.test(input)
|
||||
);
|
||||
|
||||
confidence += matchedPatterns.length * 0.15;
|
||||
|
||||
// Penalize very short or very generic matches
|
||||
if (matchWords.length <= 1) {
|
||||
confidence *= 0.5;
|
||||
}
|
||||
|
||||
return Math.min(1, confidence);
|
||||
// Ensure confidence is between 0.5 and 1
|
||||
return Math.min(1, Math.max(0.6, confidence));
|
||||
}
|
||||
|
||||
private extractActionParameters(
|
||||
@@ -131,8 +162,8 @@ export class IntentClassifier {
|
||||
}
|
||||
}
|
||||
|
||||
// Extract additional parameters from match groups
|
||||
if (match.length > 1 && match[1]) {
|
||||
// Only add raw_parameter for non-set actions
|
||||
if (actionPattern.action !== 'set' && match.length > 1 && match[1]) {
|
||||
parameters.raw_parameter = match[1].trim();
|
||||
}
|
||||
|
||||
@@ -178,3 +209,4 @@ export class IntentClassifier {
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -115,7 +115,7 @@ router.get("/subscribe_events", middleware.wsRateLimiter, (req, res) => {
|
||||
res.writeHead(200, {
|
||||
"Content-Type": "text/event-stream",
|
||||
"Cache-Control": "no-cache",
|
||||
Connection: "keep-alive",
|
||||
"Connection": "keep-alive",
|
||||
"Access-Control-Allow-Origin": "*",
|
||||
});
|
||||
|
||||
|
||||
32
src/config.js
Normal file
32
src/config.js
Normal file
@@ -0,0 +1,32 @@
|
||||
/**
|
||||
* MCP Server Configuration
|
||||
*
|
||||
* This file contains the configuration for the MCP server.
|
||||
* Values can be overridden via environment variables.
|
||||
*/
|
||||
|
||||
// Default values for the application configuration
|
||||
export const APP_CONFIG = {
|
||||
// Server configuration
|
||||
PORT: process.env.PORT ? parseInt(process.env.PORT, 10) : 3000,
|
||||
NODE_ENV: process.env.NODE_ENV || 'development',
|
||||
|
||||
// Execution settings
|
||||
EXECUTION_TIMEOUT: process.env.EXECUTION_TIMEOUT ? parseInt(process.env.EXECUTION_TIMEOUT, 10) : 30000,
|
||||
STREAMING_ENABLED: process.env.STREAMING_ENABLED === 'true',
|
||||
|
||||
// Transport settings
|
||||
USE_STDIO_TRANSPORT: process.env.USE_STDIO_TRANSPORT === 'true',
|
||||
USE_HTTP_TRANSPORT: process.env.USE_HTTP_TRANSPORT !== 'false',
|
||||
|
||||
// Debug and logging settings
|
||||
DEBUG_MODE: process.env.DEBUG_MODE === 'true',
|
||||
DEBUG_STDIO: process.env.DEBUG_STDIO === 'true',
|
||||
DEBUG_HTTP: process.env.DEBUG_HTTP === 'true',
|
||||
SILENT_STARTUP: process.env.SILENT_STARTUP === 'true',
|
||||
|
||||
// CORS settings
|
||||
CORS_ORIGIN: process.env.CORS_ORIGIN || '*'
|
||||
};
|
||||
|
||||
export default APP_CONFIG;
|
||||
61
src/config.ts
Normal file
61
src/config.ts
Normal file
@@ -0,0 +1,61 @@
|
||||
/**
|
||||
* Configuration for the Model Context Protocol (MCP) server
|
||||
* Values can be overridden using environment variables
|
||||
*/
|
||||
|
||||
import { MCPServerConfigSchema, MCPServerConfigType } from './schemas/config.schema.js';
|
||||
import { logger } from './utils/logger.js';
|
||||
|
||||
function loadConfig(): MCPServerConfigType {
|
||||
try {
|
||||
const rawConfig = {
|
||||
// Server configuration
|
||||
port: parseInt(process.env.PORT || '3000', 10),
|
||||
environment: process.env.NODE_ENV || 'development',
|
||||
|
||||
// Execution settings
|
||||
executionTimeout: parseInt(process.env.EXECUTION_TIMEOUT || '30000', 10),
|
||||
streamingEnabled: process.env.STREAMING_ENABLED === 'true',
|
||||
|
||||
// Transport settings
|
||||
useStdioTransport: process.env.USE_STDIO_TRANSPORT === 'true',
|
||||
useHttpTransport: process.env.USE_HTTP_TRANSPORT === 'true',
|
||||
|
||||
// Debug and logging
|
||||
debugMode: process.env.DEBUG_MODE === 'true',
|
||||
debugStdio: process.env.DEBUG_STDIO === 'true',
|
||||
debugHttp: process.env.DEBUG_HTTP === 'true',
|
||||
silentStartup: process.env.SILENT_STARTUP === 'true',
|
||||
|
||||
// CORS settings
|
||||
corsOrigin: process.env.CORS_ORIGIN || '*',
|
||||
|
||||
// Rate limiting
|
||||
rateLimit: {
|
||||
maxRequests: parseInt(process.env.RATE_LIMIT_MAX_REQUESTS || '100', 10),
|
||||
maxAuthRequests: parseInt(process.env.RATE_LIMIT_MAX_AUTH_REQUESTS || '5', 10),
|
||||
},
|
||||
};
|
||||
|
||||
// Validate and parse configuration
|
||||
const validatedConfig = MCPServerConfigSchema.parse(rawConfig);
|
||||
|
||||
// Log validation success
|
||||
if (!validatedConfig.silentStartup) {
|
||||
logger.info('Configuration validated successfully');
|
||||
if (validatedConfig.debugMode) {
|
||||
logger.debug('Current configuration:', validatedConfig);
|
||||
}
|
||||
}
|
||||
|
||||
return validatedConfig;
|
||||
} catch (error) {
|
||||
// Log validation errors
|
||||
logger.error('Configuration validation failed:', error);
|
||||
throw new Error('Invalid configuration. Please check your environment variables.');
|
||||
}
|
||||
}
|
||||
|
||||
export const APP_CONFIG = loadConfig();
|
||||
export type { MCPServerConfigType };
|
||||
export default APP_CONFIG;
|
||||
@@ -1,36 +1,18 @@
|
||||
import { config } from "dotenv";
|
||||
import { resolve } from "path";
|
||||
import { z } from "zod";
|
||||
|
||||
/**
|
||||
* Load environment variables based on NODE_ENV
|
||||
* Development: .env.development
|
||||
* Test: .env.test
|
||||
* Production: .env
|
||||
*/
|
||||
const envFile =
|
||||
process.env.NODE_ENV === "production"
|
||||
? ".env"
|
||||
: process.env.NODE_ENV === "test"
|
||||
? ".env.test"
|
||||
: ".env.development";
|
||||
|
||||
console.log(`Loading environment from ${envFile}`);
|
||||
config({ path: resolve(process.cwd(), envFile) });
|
||||
|
||||
/**
|
||||
* Application configuration object
|
||||
* Contains all configuration settings for the application
|
||||
*/
|
||||
export const AppConfigSchema = z.object({
|
||||
/** Server Configuration */
|
||||
PORT: z.number().default(4000),
|
||||
PORT: z.coerce.number().default(4000),
|
||||
NODE_ENV: z
|
||||
.enum(["development", "production", "test"])
|
||||
.default("development"),
|
||||
|
||||
/** Home Assistant Configuration */
|
||||
HASS_HOST: z.string().default("http://192.168.178.63:8123"),
|
||||
HASS_HOST: z.string().default("http://homeassistant.local:8123"),
|
||||
HASS_TOKEN: z.string().optional(),
|
||||
|
||||
/** Speech Features Configuration */
|
||||
@@ -49,7 +31,7 @@ export const AppConfigSchema = z.object({
|
||||
}),
|
||||
|
||||
/** Security Configuration */
|
||||
JWT_SECRET: z.string().default("your-secret-key"),
|
||||
JWT_SECRET: z.string().default("your-secret-key-must-be-32-char-min"),
|
||||
RATE_LIMIT: z.object({
|
||||
/** Time window for rate limiting in milliseconds */
|
||||
windowMs: z.number().default(15 * 60 * 1000), // 15 minutes
|
||||
@@ -103,10 +85,16 @@ for (const envVar of requiredEnvVars) {
|
||||
}
|
||||
}
|
||||
|
||||
// Fix NODE_ENV if it's set to "1"
|
||||
if (process.env.NODE_ENV === "1") {
|
||||
console.log('Fixing NODE_ENV from "1" to "development"');
|
||||
process.env.NODE_ENV = "development";
|
||||
}
|
||||
|
||||
// Load and validate configuration
|
||||
export const APP_CONFIG = AppConfigSchema.parse({
|
||||
PORT: process.env.PORT || 4000,
|
||||
NODE_ENV: process.env.NODE_ENV || "development",
|
||||
NODE_ENV: process.env.NODE_ENV,
|
||||
HASS_HOST: process.env.HASS_HOST || "http://192.168.178.63:8123",
|
||||
HASS_TOKEN: process.env.HASS_TOKEN,
|
||||
JWT_SECRET: process.env.JWT_SECRET || "your-secret-key",
|
||||
|
||||
@@ -1,35 +0,0 @@
|
||||
export const BOILERPLATE_CONFIG = {
|
||||
configuration: {
|
||||
LOG_LEVEL: {
|
||||
type: "string" as const,
|
||||
default: "debug",
|
||||
description: "Logging level",
|
||||
enum: ["error", "warn", "info", "debug", "trace"],
|
||||
},
|
||||
CACHE_DIRECTORY: {
|
||||
type: "string" as const,
|
||||
default: ".cache",
|
||||
description: "Directory for cache files",
|
||||
},
|
||||
CONFIG_DIRECTORY: {
|
||||
type: "string" as const,
|
||||
default: ".config",
|
||||
description: "Directory for configuration files",
|
||||
},
|
||||
DATA_DIRECTORY: {
|
||||
type: "string" as const,
|
||||
default: ".data",
|
||||
description: "Directory for data files",
|
||||
},
|
||||
},
|
||||
internal: {
|
||||
boilerplate: {
|
||||
configuration: {
|
||||
LOG_LEVEL: "debug",
|
||||
CACHE_DIRECTORY: ".cache",
|
||||
CONFIG_DIRECTORY: ".config",
|
||||
DATA_DIRECTORY: ".data",
|
||||
},
|
||||
},
|
||||
},
|
||||
};
|
||||
@@ -11,6 +11,7 @@ const envFile =
|
||||
|
||||
config({ path: resolve(process.cwd(), envFile) });
|
||||
|
||||
// Base configuration for Home Assistant
|
||||
export const HASS_CONFIG = {
|
||||
// Base configuration
|
||||
BASE_URL: process.env.HASS_HOST || "http://localhost:8123",
|
||||
|
||||
@@ -1,16 +1,7 @@
|
||||
import { config } from "dotenv";
|
||||
import { resolve } from "path";
|
||||
import { loadEnvironmentVariables } from "./loadEnv";
|
||||
|
||||
// Load environment variables based on NODE_ENV
|
||||
const envFile =
|
||||
process.env.NODE_ENV === "production"
|
||||
? ".env"
|
||||
: process.env.NODE_ENV === "test"
|
||||
? ".env.test"
|
||||
: ".env.development";
|
||||
|
||||
console.log(`Loading environment from ${envFile}`);
|
||||
config({ path: resolve(process.cwd(), envFile) });
|
||||
// Load environment variables from the appropriate files
|
||||
loadEnvironmentVariables();
|
||||
|
||||
// Home Assistant Configuration
|
||||
export const HASS_CONFIG = {
|
||||
|
||||
59
src/config/loadEnv.ts
Normal file
59
src/config/loadEnv.ts
Normal file
@@ -0,0 +1,59 @@
|
||||
import { config as dotenvConfig } from "dotenv";
|
||||
import { file } from "bun";
|
||||
import path from "path";
|
||||
|
||||
/**
|
||||
* Maps NODE_ENV values to their corresponding environment file names
|
||||
*/
|
||||
const ENV_FILE_MAPPING: Record<string, string> = {
|
||||
production: ".env.prod",
|
||||
development: ".env.dev",
|
||||
test: ".env.test",
|
||||
};
|
||||
|
||||
/**
|
||||
* Loads environment variables from the appropriate files based on NODE_ENV.
|
||||
* First loads environment-specific file, then overrides with generic .env if it exists.
|
||||
*/
|
||||
export async function loadEnvironmentVariables() {
|
||||
// Determine the current environment (default to 'development')
|
||||
const nodeEnv = (process.env.NODE_ENV || "development").toLowerCase();
|
||||
|
||||
// Get the environment-specific file name
|
||||
const envSpecificFile = ENV_FILE_MAPPING[nodeEnv];
|
||||
if (!envSpecificFile) {
|
||||
console.warn(`Unknown NODE_ENV value: ${nodeEnv}. Using .env.dev as fallback.`);
|
||||
}
|
||||
|
||||
const envFile = envSpecificFile || ".env.dev";
|
||||
const envPath = path.resolve(process.cwd(), envFile);
|
||||
|
||||
// Load the environment-specific file if it exists
|
||||
try {
|
||||
const envFileExists = await file(envPath).exists();
|
||||
if (envFileExists) {
|
||||
dotenvConfig({ path: envPath });
|
||||
console.log(`Loaded environment variables from ${envFile}`);
|
||||
} else {
|
||||
console.warn(`Environment-specific file ${envFile} not found.`);
|
||||
}
|
||||
} catch (error) {
|
||||
console.warn(`Error checking environment file ${envFile}:`, error);
|
||||
}
|
||||
|
||||
// Finally, check if there is a generic .env file present
|
||||
// If so, load it with the override option, so its values take precedence
|
||||
const genericEnvPath = path.resolve(process.cwd(), ".env");
|
||||
try {
|
||||
const genericEnvExists = await file(genericEnvPath).exists();
|
||||
if (genericEnvExists) {
|
||||
dotenvConfig({ path: genericEnvPath, override: true });
|
||||
console.log("Loaded and overrode with generic .env file");
|
||||
}
|
||||
} catch (error) {
|
||||
console.warn(`Error checking generic .env file:`, error);
|
||||
}
|
||||
}
|
||||
|
||||
// Export the environment file mapping for reference
|
||||
export const ENV_FILES = ENV_FILE_MAPPING;
|
||||
74
src/hass/types.ts
Normal file
74
src/hass/types.ts
Normal file
@@ -0,0 +1,74 @@
|
||||
import type { WebSocket } from 'ws';
|
||||
|
||||
export interface HassInstanceImpl {
|
||||
baseUrl: string;
|
||||
token: string;
|
||||
connect(): Promise<void>;
|
||||
disconnect(): Promise<void>;
|
||||
getStates(): Promise<any[]>;
|
||||
callService(domain: string, service: string, data?: any): Promise<void>;
|
||||
fetchStates(): Promise<any[]>;
|
||||
fetchState(entityId: string): Promise<any>;
|
||||
subscribeEvents(callback: (event: any) => void, eventType?: string): Promise<number>;
|
||||
unsubscribeEvents(subscriptionId: number): Promise<void>;
|
||||
}
|
||||
|
||||
export interface HassWebSocketClient {
|
||||
url: string;
|
||||
token: string;
|
||||
socket: WebSocket | null;
|
||||
connect(): Promise<void>;
|
||||
disconnect(): Promise<void>;
|
||||
send(message: any): Promise<void>;
|
||||
subscribe(callback: (data: any) => void): () => void;
|
||||
}
|
||||
|
||||
export interface HassState {
|
||||
entity_id: string;
|
||||
state: string;
|
||||
attributes: Record<string, any>;
|
||||
last_changed: string;
|
||||
last_updated: string;
|
||||
context: {
|
||||
id: string;
|
||||
parent_id: string | null;
|
||||
user_id: string | null;
|
||||
};
|
||||
}
|
||||
|
||||
export interface HassServiceCall {
|
||||
domain: string;
|
||||
service: string;
|
||||
target?: {
|
||||
entity_id?: string | string[];
|
||||
device_id?: string | string[];
|
||||
area_id?: string | string[];
|
||||
};
|
||||
service_data?: Record<string, any>;
|
||||
}
|
||||
|
||||
export interface HassEvent {
|
||||
event_type: string;
|
||||
data: any;
|
||||
origin: string;
|
||||
time_fired: string;
|
||||
context: {
|
||||
id: string;
|
||||
parent_id: string | null;
|
||||
user_id: string | null;
|
||||
};
|
||||
}
|
||||
|
||||
export type MockFunction<T extends (...args: any[]) => any> = {
|
||||
(...args: Parameters<T>): ReturnType<T>;
|
||||
mock: {
|
||||
calls: Parameters<T>[];
|
||||
results: { type: 'return' | 'throw'; value: any }[];
|
||||
instances: any[];
|
||||
mockImplementation(fn: T): MockFunction<T>;
|
||||
mockReturnValue(value: ReturnType<T>): MockFunction<T>;
|
||||
mockResolvedValue(value: Awaited<ReturnType<T>>): MockFunction<T>;
|
||||
mockRejectedValue(value: any): MockFunction<T>;
|
||||
mockReset(): void;
|
||||
};
|
||||
};
|
||||
301
src/index.ts
301
src/index.ts
@@ -1,169 +1,184 @@
|
||||
import "./polyfills.js";
|
||||
import { config } from "dotenv";
|
||||
import { resolve } from "path";
|
||||
import { Elysia } from "elysia";
|
||||
import { cors } from "@elysiajs/cors";
|
||||
import { swagger } from "@elysiajs/swagger";
|
||||
import {
|
||||
rateLimiter,
|
||||
securityHeaders,
|
||||
validateRequest,
|
||||
sanitizeInput,
|
||||
errorHandler,
|
||||
} from "./security/index.js";
|
||||
import {
|
||||
get_hass,
|
||||
call_service,
|
||||
list_devices,
|
||||
get_states,
|
||||
get_state,
|
||||
} from "./hass/index.js";
|
||||
import { z } from "zod";
|
||||
import {
|
||||
commonCommands,
|
||||
coverCommands,
|
||||
climateCommands,
|
||||
type Command,
|
||||
} from "./commands.js";
|
||||
import { speechService } from "./speech/index.js";
|
||||
import { APP_CONFIG } from "./config/app.config.js";
|
||||
/**
|
||||
* Home Assistant Model Context Protocol (MCP) Server
|
||||
* A standardized protocol for AI tools to interact with Home Assistant
|
||||
*/
|
||||
|
||||
// Load environment variables based on NODE_ENV
|
||||
const envFile =
|
||||
process.env.NODE_ENV === "production"
|
||||
? ".env"
|
||||
: process.env.NODE_ENV === "test"
|
||||
? ".env.test"
|
||||
: ".env.development";
|
||||
import express from 'express';
|
||||
import cors from 'cors';
|
||||
import swaggerUi from 'swagger-ui-express';
|
||||
import { MCPServer } from './mcp/MCPServer.js';
|
||||
import { loggingMiddleware, timeoutMiddleware } from './mcp/middleware/index.js';
|
||||
import { StdioTransport } from './mcp/transports/stdio.transport.js';
|
||||
import { HttpTransport } from './mcp/transports/http.transport.js';
|
||||
import { APP_CONFIG } from './config.js';
|
||||
import { logger } from "./utils/logger.js";
|
||||
import { openApiConfig } from './openapi.js';
|
||||
import { apiLimiter, authLimiter } from './middleware/rate-limit.middleware.js';
|
||||
import { SecurityMiddleware } from './security/enhanced-middleware.js';
|
||||
|
||||
console.log(`Loading environment from ${envFile}`);
|
||||
config({ path: resolve(process.cwd(), envFile) });
|
||||
// Home Assistant tools
|
||||
import { LightsControlTool } from './tools/homeassistant/lights.tool.js';
|
||||
import { ClimateControlTool } from './tools/homeassistant/climate.tool.js';
|
||||
|
||||
// Configuration
|
||||
const HASS_TOKEN = process.env.HASS_TOKEN;
|
||||
const PORT = parseInt(process.env.PORT || "4000", 10);
|
||||
// Home Assistant optional tools - these can be added as needed
|
||||
// import { ControlTool } from './tools/control.tool.js';
|
||||
// import { SceneTool } from './tools/scene.tool.js';
|
||||
// import { AutomationTool } from './tools/automation.tool.js';
|
||||
// import { NotifyTool } from './tools/notify.tool.js';
|
||||
// import { ListDevicesTool } from './tools/list-devices.tool.js';
|
||||
// import { HistoryTool } from './tools/history.tool.js';
|
||||
|
||||
console.log("Initializing Home Assistant connection...");
|
||||
|
||||
// Define Tool interface
|
||||
interface Tool {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters: z.ZodType<any>;
|
||||
execute: (params: any) => Promise<any>;
|
||||
/**
|
||||
* Check if running in stdio mode via command line args
|
||||
*/
|
||||
function isStdioMode(): boolean {
|
||||
return process.argv.includes('--stdio');
|
||||
}
|
||||
|
||||
// Array to store tools
|
||||
const tools: Tool[] = [];
|
||||
/**
|
||||
* Main function to start the MCP server
|
||||
*/
|
||||
async function main() {
|
||||
logger.info('Starting Home Assistant MCP Server...');
|
||||
|
||||
// Define the list devices tool
|
||||
const listDevicesTool: Tool = {
|
||||
name: "list_devices",
|
||||
description: "List all available Home Assistant devices",
|
||||
parameters: z.object({}),
|
||||
execute: async () => {
|
||||
try {
|
||||
const devices = await list_devices();
|
||||
return {
|
||||
success: true,
|
||||
devices,
|
||||
};
|
||||
} catch (error) {
|
||||
return {
|
||||
success: false,
|
||||
message:
|
||||
error instanceof Error ? error.message : "Unknown error occurred",
|
||||
};
|
||||
}
|
||||
},
|
||||
};
|
||||
// Check if we're in stdio mode from command line
|
||||
const useStdio = isStdioMode() || APP_CONFIG.useStdioTransport;
|
||||
|
||||
// Add tools to the array
|
||||
tools.push(listDevicesTool);
|
||||
// Configure server
|
||||
const EXECUTION_TIMEOUT = APP_CONFIG.executionTimeout;
|
||||
const STREAMING_ENABLED = APP_CONFIG.streamingEnabled;
|
||||
|
||||
// Add the Home Assistant control tool
|
||||
const controlTool: Tool = {
|
||||
name: "control",
|
||||
description: "Control Home Assistant devices and services",
|
||||
parameters: z.object({
|
||||
command: z.enum([
|
||||
...commonCommands,
|
||||
...coverCommands,
|
||||
...climateCommands,
|
||||
] as [string, ...string[]]),
|
||||
entity_id: z.string().describe("The ID of the entity to control"),
|
||||
}),
|
||||
execute: async (params: { command: Command; entity_id: string }) => {
|
||||
try {
|
||||
const [domain] = params.entity_id.split(".");
|
||||
await call_service(domain, params.command, {
|
||||
entity_id: params.entity_id,
|
||||
// Get the server instance (singleton)
|
||||
const server = MCPServer.getInstance();
|
||||
|
||||
// Register Home Assistant tools
|
||||
server.registerTool(new LightsControlTool());
|
||||
server.registerTool(new ClimateControlTool());
|
||||
|
||||
// Add optional tools here as needed
|
||||
// server.registerTool(new ControlTool());
|
||||
// server.registerTool(new SceneTool());
|
||||
// server.registerTool(new NotifyTool());
|
||||
// server.registerTool(new ListDevicesTool());
|
||||
// server.registerTool(new HistoryTool());
|
||||
|
||||
// Add middlewares
|
||||
server.use(loggingMiddleware);
|
||||
server.use(timeoutMiddleware(EXECUTION_TIMEOUT));
|
||||
|
||||
// Initialize transports
|
||||
if (useStdio) {
|
||||
logger.info('Using Standard I/O transport');
|
||||
|
||||
// Create and configure the stdio transport with debug enabled for stdio mode
|
||||
const stdioTransport = new StdioTransport({
|
||||
debug: true, // Always enable debug in stdio mode for better visibility
|
||||
silent: false // Never be silent in stdio mode
|
||||
});
|
||||
return {
|
||||
success: true,
|
||||
message: `Command ${params.command} executed successfully on ${params.entity_id}`,
|
||||
};
|
||||
|
||||
// Explicitly set the server reference to ensure access to tools
|
||||
stdioTransport.setServer(server);
|
||||
|
||||
// Register the transport
|
||||
server.registerTransport(stdioTransport);
|
||||
|
||||
// Special handling for stdio mode - don't start other transports
|
||||
if (isStdioMode()) {
|
||||
logger.info('Running in pure stdio mode (from CLI)');
|
||||
// Start the server
|
||||
await server.start();
|
||||
logger.info('MCP Server started successfully');
|
||||
|
||||
// Handle shutdown
|
||||
const shutdown = async () => {
|
||||
logger.info('Shutting down MCP Server...');
|
||||
try {
|
||||
await server.shutdown();
|
||||
logger.info('MCP Server shutdown complete');
|
||||
process.exit(0);
|
||||
} catch (error) {
|
||||
return {
|
||||
success: false,
|
||||
message:
|
||||
error instanceof Error ? error.message : "Unknown error occurred",
|
||||
};
|
||||
logger.error('Error during shutdown:', error);
|
||||
process.exit(1);
|
||||
}
|
||||
},
|
||||
};
|
||||
|
||||
// Add the control tool to the array
|
||||
tools.push(controlTool);
|
||||
// Register shutdown handlers
|
||||
process.on('SIGINT', shutdown);
|
||||
process.on('SIGTERM', shutdown);
|
||||
|
||||
// Initialize Elysia app with middleware
|
||||
const app = new Elysia()
|
||||
.use(cors())
|
||||
.use(swagger())
|
||||
.use(rateLimiter)
|
||||
.use(securityHeaders)
|
||||
.use(validateRequest)
|
||||
.use(sanitizeInput)
|
||||
.use(errorHandler);
|
||||
// Exit the function early as we're in stdio-only mode
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
// Health check endpoint
|
||||
app.get("/health", () => ({
|
||||
status: "ok",
|
||||
timestamp: new Date().toISOString(),
|
||||
version: "0.1.0",
|
||||
speech_enabled: APP_CONFIG.SPEECH.ENABLED,
|
||||
wake_word_enabled: APP_CONFIG.SPEECH.WAKE_WORD_ENABLED,
|
||||
speech_to_text_enabled: APP_CONFIG.SPEECH.SPEECH_TO_TEXT_ENABLED,
|
||||
// HTTP transport (only if not in pure stdio mode)
|
||||
if (APP_CONFIG.useHttpTransport) {
|
||||
logger.info('Using HTTP transport on port ' + APP_CONFIG.port);
|
||||
const app = express();
|
||||
|
||||
// Apply enhanced security middleware
|
||||
app.use(SecurityMiddleware.applySecurityHeaders);
|
||||
|
||||
// CORS configuration
|
||||
app.use(cors({
|
||||
origin: APP_CONFIG.corsOrigin,
|
||||
methods: ['GET', 'POST', 'PUT', 'DELETE', 'OPTIONS'],
|
||||
allowedHeaders: ['Content-Type', 'Authorization'],
|
||||
maxAge: 86400 // 24 hours
|
||||
}));
|
||||
|
||||
// Initialize speech service if enabled
|
||||
if (APP_CONFIG.SPEECH.ENABLED) {
|
||||
console.log("Initializing speech service...");
|
||||
speechService.initialize().catch((error) => {
|
||||
console.error("Failed to initialize speech service:", error);
|
||||
});
|
||||
}
|
||||
// Apply rate limiting to all routes
|
||||
app.use('/api', apiLimiter);
|
||||
app.use('/auth', authLimiter);
|
||||
|
||||
// Create API endpoints for each tool
|
||||
tools.forEach((tool) => {
|
||||
app.post(`/api/tools/${tool.name}`, async ({ body }: { body: Record<string, unknown> }) => {
|
||||
const result = await tool.execute(body);
|
||||
return result;
|
||||
// Swagger UI setup
|
||||
app.use('/api-docs', swaggerUi.serve, swaggerUi.setup(openApiConfig, {
|
||||
explorer: true,
|
||||
customCss: '.swagger-ui .topbar { display: none }',
|
||||
customSiteTitle: 'Home Assistant MCP API Documentation'
|
||||
}));
|
||||
|
||||
// Health check endpoint
|
||||
app.get('/health', (req, res) => {
|
||||
res.json({
|
||||
status: 'ok',
|
||||
version: process.env.npm_package_version || '1.0.0'
|
||||
});
|
||||
});
|
||||
|
||||
const httpTransport = new HttpTransport({
|
||||
port: APP_CONFIG.port,
|
||||
corsOrigin: APP_CONFIG.corsOrigin,
|
||||
apiPrefix: "/api/mcp",
|
||||
debug: APP_CONFIG.debugHttp
|
||||
});
|
||||
server.registerTransport(httpTransport);
|
||||
}
|
||||
|
||||
// Start the server
|
||||
app.listen(PORT, () => {
|
||||
console.log(`Server is running on port ${PORT}`);
|
||||
});
|
||||
await server.start();
|
||||
logger.info('MCP Server started successfully');
|
||||
|
||||
// Handle server shutdown
|
||||
process.on("SIGTERM", async () => {
|
||||
console.log("Received SIGTERM. Shutting down gracefully...");
|
||||
if (APP_CONFIG.SPEECH.ENABLED) {
|
||||
await speechService.shutdown().catch((error) => {
|
||||
console.error("Error shutting down speech service:", error);
|
||||
});
|
||||
}
|
||||
// Handle shutdown
|
||||
const shutdown = async () => {
|
||||
logger.info('Shutting down MCP Server...');
|
||||
try {
|
||||
await server.shutdown();
|
||||
logger.info('MCP Server shutdown complete');
|
||||
process.exit(0);
|
||||
} catch (error) {
|
||||
logger.error('Error during shutdown:', error);
|
||||
process.exit(1);
|
||||
}
|
||||
};
|
||||
|
||||
// Register shutdown handlers
|
||||
process.on('SIGINT', shutdown);
|
||||
process.on('SIGTERM', shutdown);
|
||||
}
|
||||
|
||||
// Run the main function
|
||||
main().catch(error => {
|
||||
logger.error('Error starting MCP Server:', error);
|
||||
process.exit(1);
|
||||
});
|
||||
|
||||
105
src/mcp/BaseTool.ts
Normal file
105
src/mcp/BaseTool.ts
Normal file
@@ -0,0 +1,105 @@
|
||||
/**
|
||||
* Base Tool Implementation for MCP
|
||||
*
|
||||
* This base class provides the foundation for all tools in the MCP implementation,
|
||||
* with typed parameters, validation, and error handling.
|
||||
*/
|
||||
|
||||
import { z } from 'zod';
|
||||
import { ToolDefinition, ToolMetadata, MCPResponseStream } from './types.js';
|
||||
|
||||
/**
|
||||
* Configuration options for creating a tool
|
||||
*/
|
||||
export interface ToolOptions<P = unknown> {
|
||||
name: string;
|
||||
description: string;
|
||||
version: string;
|
||||
parameters?: z.ZodType<P>;
|
||||
metadata?: ToolMetadata;
|
||||
}
|
||||
|
||||
/**
|
||||
* Base class for all MCP tools
|
||||
*
|
||||
* Provides:
|
||||
* - Parameter validation with Zod
|
||||
* - Error handling
|
||||
* - Streaming support
|
||||
* - Type safety
|
||||
*/
|
||||
export abstract class BaseTool<P = unknown, R = unknown> implements ToolDefinition {
|
||||
public readonly name: string;
|
||||
public readonly description: string;
|
||||
public readonly parameters?: z.ZodType<P>;
|
||||
public readonly metadata: ToolMetadata;
|
||||
|
||||
/**
|
||||
* Create a new tool
|
||||
*/
|
||||
constructor(options: ToolOptions<P>) {
|
||||
this.name = options.name;
|
||||
this.description = options.description;
|
||||
this.parameters = options.parameters;
|
||||
this.metadata = {
|
||||
version: options.version,
|
||||
category: options.metadata?.category || 'general',
|
||||
tags: options.metadata?.tags || [],
|
||||
examples: options.metadata?.examples || [],
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Execute the tool with the given parameters
|
||||
*
|
||||
* @param params The validated parameters for the tool
|
||||
* @param stream Optional stream for sending partial results
|
||||
* @returns The result of the tool execution
|
||||
*/
|
||||
abstract execute(params: P, stream?: MCPResponseStream): Promise<R>;
|
||||
|
||||
/**
|
||||
* Get the parameter schema as JSON schema
|
||||
*/
|
||||
public getParameterSchema(): Record<string, unknown> | undefined {
|
||||
if (!this.parameters) return undefined;
|
||||
return this.parameters.isOptional()
|
||||
? { type: 'object', properties: {} }
|
||||
: this.parameters.shape;
|
||||
}
|
||||
|
||||
/**
|
||||
* Get tool definition for registration
|
||||
*/
|
||||
public getDefinition(): ToolDefinition {
|
||||
return {
|
||||
name: this.name,
|
||||
description: this.description,
|
||||
parameters: this.parameters,
|
||||
metadata: this.metadata
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Validate parameters against the schema
|
||||
*
|
||||
* @param params Parameters to validate
|
||||
* @returns Validated parameters
|
||||
* @throws Error if validation fails
|
||||
*/
|
||||
protected validateParams(params: unknown): P {
|
||||
if (!this.parameters) {
|
||||
return {} as P;
|
||||
}
|
||||
|
||||
try {
|
||||
return this.parameters.parse(params);
|
||||
} catch (error) {
|
||||
if (error instanceof z.ZodError) {
|
||||
const issues = error.issues.map(issue => `${issue.path.join('.')}: ${issue.message}`).join(', ');
|
||||
throw new Error(`Parameter validation failed: ${issues}`);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
}
|
||||
453
src/mcp/MCPServer.ts
Normal file
453
src/mcp/MCPServer.ts
Normal file
@@ -0,0 +1,453 @@
|
||||
/**
|
||||
* MCPServer.ts
|
||||
*
|
||||
* Core implementation of the Model Context Protocol server.
|
||||
* This class manages tool registration, execution, and resource handling
|
||||
* while providing integration with various transport layers.
|
||||
*/
|
||||
|
||||
import { EventEmitter } from "events";
|
||||
import { z } from "zod";
|
||||
import { v4 as uuidv4 } from "uuid";
|
||||
import { logger } from "../utils/logger.js";
|
||||
|
||||
// Error code enum to break circular dependency
|
||||
export enum MCPErrorCode {
|
||||
// Standard JSON-RPC 2.0 error codes
|
||||
PARSE_ERROR = -32700,
|
||||
INVALID_REQUEST = -32600,
|
||||
METHOD_NOT_FOUND = -32601,
|
||||
INVALID_PARAMS = -32602,
|
||||
INTERNAL_ERROR = -32603,
|
||||
|
||||
// Custom MCP error codes
|
||||
TOOL_EXECUTION_ERROR = -32000,
|
||||
VALIDATION_ERROR = -32001,
|
||||
RESOURCE_NOT_FOUND = -32002,
|
||||
RESOURCE_BUSY = -32003,
|
||||
TIMEOUT = -32004,
|
||||
CANCELED = -32005,
|
||||
AUTHENTICATION_ERROR = -32006,
|
||||
AUTHORIZATION_ERROR = -32007,
|
||||
TRANSPORT_ERROR = -32008,
|
||||
STREAMING_ERROR = -32009
|
||||
}
|
||||
|
||||
// Server events enum to break circular dependency
|
||||
export enum MCPServerEvents {
|
||||
STARTING = "starting",
|
||||
STARTED = "started",
|
||||
SHUTTING_DOWN = "shuttingDown",
|
||||
SHUTDOWN = "shutdown",
|
||||
REQUEST_RECEIVED = "requestReceived",
|
||||
RESPONSE_SENT = "responseSent",
|
||||
RESPONSE_ERROR = "responseError",
|
||||
TOOL_REGISTERED = "toolRegistered",
|
||||
TRANSPORT_REGISTERED = "transportRegistered",
|
||||
CONFIG_UPDATED = "configUpdated"
|
||||
}
|
||||
|
||||
// Forward declarations to break circular dependency
|
||||
import type {
|
||||
ToolDefinition,
|
||||
MCPMiddleware,
|
||||
MCPRequest,
|
||||
MCPResponse,
|
||||
MCPContext,
|
||||
TransportLayer,
|
||||
MCPConfig,
|
||||
ResourceManager
|
||||
} from "./types.js";
|
||||
|
||||
/**
|
||||
* Main Model Context Protocol server class
|
||||
*/
|
||||
export class MCPServer extends EventEmitter {
|
||||
private static instance: MCPServer;
|
||||
private tools: Map<string, ToolDefinition> = new Map();
|
||||
private middlewares: MCPMiddleware[] = [];
|
||||
private transports: TransportLayer[] = [];
|
||||
private resourceManager: ResourceManager;
|
||||
private config: MCPConfig;
|
||||
private resources: Map<string, Map<string, any>> = new Map();
|
||||
|
||||
/**
|
||||
* Private constructor for singleton pattern
|
||||
*/
|
||||
private constructor(config: Partial<MCPConfig> = {}) {
|
||||
super();
|
||||
this.config = {
|
||||
maxRetries: 3,
|
||||
retryDelay: 1000,
|
||||
executionTimeout: 30000,
|
||||
streamingEnabled: true,
|
||||
maxPayloadSize: 10 * 1024 * 1024, // 10MB
|
||||
...config
|
||||
};
|
||||
|
||||
this.resourceManager = {
|
||||
acquire: this.acquireResource.bind(this),
|
||||
release: this.releaseResource.bind(this),
|
||||
list: this.listResources.bind(this)
|
||||
};
|
||||
|
||||
// Initialize with default middlewares
|
||||
this.use(this.validationMiddleware.bind(this));
|
||||
this.use(this.errorHandlingMiddleware.bind(this));
|
||||
|
||||
logger.info("MCP Server initialized");
|
||||
}
|
||||
|
||||
/**
|
||||
* Get singleton instance
|
||||
*/
|
||||
public static getInstance(config?: Partial<MCPConfig>): MCPServer {
|
||||
if (!MCPServer.instance) {
|
||||
MCPServer.instance = new MCPServer(config);
|
||||
} else if (config) {
|
||||
MCPServer.instance.configure(config);
|
||||
}
|
||||
return MCPServer.instance;
|
||||
}
|
||||
|
||||
/**
|
||||
* Update server configuration
|
||||
*/
|
||||
public configure(config: Partial<MCPConfig>): void {
|
||||
this.config = {
|
||||
...this.config,
|
||||
...config
|
||||
};
|
||||
logger.debug("MCP Server configuration updated", { config });
|
||||
this.emit(MCPServerEvents.CONFIG_UPDATED, this.config);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register a new tool with the server
|
||||
*/
|
||||
public registerTool(tool: ToolDefinition): void {
|
||||
if (this.tools.has(tool.name)) {
|
||||
logger.warn(`Tool '${tool.name}' is already registered. Overwriting.`);
|
||||
}
|
||||
|
||||
this.tools.set(tool.name, tool);
|
||||
logger.debug(`Tool '${tool.name}' registered`);
|
||||
this.emit(MCPServerEvents.TOOL_REGISTERED, tool);
|
||||
}
|
||||
|
||||
/**
|
||||
* Register multiple tools at once
|
||||
*/
|
||||
public registerTools(tools: ToolDefinition[]): void {
|
||||
tools.forEach(tool => this.registerTool(tool));
|
||||
}
|
||||
|
||||
/**
|
||||
* Get a tool by name
|
||||
*/
|
||||
public getTool(name: string): ToolDefinition | undefined {
|
||||
return this.tools.get(name);
|
||||
}
|
||||
|
||||
/**
|
||||
* Get all registered tools
|
||||
*/
|
||||
public getAllTools(): ToolDefinition[] {
|
||||
return Array.from(this.tools.values());
|
||||
}
|
||||
|
||||
/**
|
||||
* Register a transport layer
|
||||
*/
|
||||
public registerTransport(transport: TransportLayer): void {
|
||||
this.transports.push(transport);
|
||||
transport.initialize(this.handleRequest.bind(this));
|
||||
logger.debug(`Transport '${transport.name}' registered`);
|
||||
this.emit(MCPServerEvents.TRANSPORT_REGISTERED, transport);
|
||||
}
|
||||
|
||||
/**
|
||||
* Add a middleware to the pipeline
|
||||
*/
|
||||
public use(middleware: MCPMiddleware): void {
|
||||
this.middlewares.push(middleware);
|
||||
logger.debug("Middleware added");
|
||||
}
|
||||
|
||||
/**
|
||||
* Handle an incoming request through the middleware pipeline
|
||||
*/
|
||||
public async handleRequest(request: MCPRequest): Promise<MCPResponse> {
|
||||
const context: MCPContext = {
|
||||
requestId: request.id ?? uuidv4(),
|
||||
startTime: Date.now(),
|
||||
resourceManager: this.resourceManager,
|
||||
tools: this.tools,
|
||||
config: this.config,
|
||||
logger: logger.child({ requestId: request.id }),
|
||||
server: this,
|
||||
state: new Map()
|
||||
};
|
||||
|
||||
logger.debug(`Handling request: ${context.requestId}`, { method: request.method });
|
||||
this.emit(MCPServerEvents.REQUEST_RECEIVED, request, context);
|
||||
|
||||
let index = 0;
|
||||
const next = async (): Promise<MCPResponse> => {
|
||||
if (index < this.middlewares.length) {
|
||||
const middleware = this.middlewares[index++];
|
||||
return middleware(request, context, next);
|
||||
} else {
|
||||
return this.executeRequest(request, context);
|
||||
}
|
||||
};
|
||||
|
||||
try {
|
||||
const response = await next();
|
||||
this.emit(MCPServerEvents.RESPONSE_SENT, response, context);
|
||||
return response;
|
||||
} catch (error) {
|
||||
const errorResponse: MCPResponse = {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.INTERNAL_ERROR,
|
||||
message: error instanceof Error ? error.message : String(error)
|
||||
}
|
||||
};
|
||||
this.emit(MCPServerEvents.RESPONSE_ERROR, errorResponse, context);
|
||||
return errorResponse;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Execute a tool request after middleware processing
|
||||
*/
|
||||
private async executeRequest(request: MCPRequest, context: MCPContext): Promise<MCPResponse> {
|
||||
const { method, params = {} } = request;
|
||||
|
||||
// Special case for internal context retrieval (used by transports for initialization)
|
||||
if (method === "_internal_getContext") {
|
||||
return {
|
||||
id: request.id,
|
||||
result: {
|
||||
context: context,
|
||||
tools: Array.from(this.tools.values()).map(tool => ({
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
metadata: tool.metadata
|
||||
}))
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
const tool = this.tools.get(method);
|
||||
if (!tool) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.METHOD_NOT_FOUND,
|
||||
message: `Method not found: ${method}`
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
try {
|
||||
const result = await tool.execute(params, context);
|
||||
return {
|
||||
id: request.id,
|
||||
result
|
||||
};
|
||||
} catch (error) {
|
||||
logger.error(`Error executing tool ${method}:`, error);
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.TOOL_EXECUTION_ERROR,
|
||||
message: error instanceof Error ? error.message : String(error)
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Validation middleware
|
||||
*/
|
||||
private async validationMiddleware(
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> {
|
||||
const { method, params = {} } = request;
|
||||
|
||||
const tool = this.tools.get(method);
|
||||
if (!tool) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.METHOD_NOT_FOUND,
|
||||
message: `Method not found: ${method}`
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
if (tool.parameters && params) {
|
||||
try {
|
||||
// Validate parameters with the schema
|
||||
const validParams = tool.parameters.parse(params);
|
||||
// Update with validated params (which may include defaults)
|
||||
request.params = validParams;
|
||||
} catch (validationError) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.INVALID_PARAMS,
|
||||
message: "Invalid parameters",
|
||||
data: validationError instanceof Error ? validationError.message : String(validationError)
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
return next();
|
||||
}
|
||||
|
||||
/**
|
||||
* Error handling middleware
|
||||
*/
|
||||
private async errorHandlingMiddleware(
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> {
|
||||
try {
|
||||
return await next();
|
||||
} catch (error) {
|
||||
logger.error(`Uncaught error in request pipeline:`, error);
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.INTERNAL_ERROR,
|
||||
message: error instanceof Error ? error.message : "An unknown error occurred",
|
||||
data: error instanceof Error ? { name: error.name, stack: error.stack } : undefined
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Resource acquisition
|
||||
*/
|
||||
private async acquireResource(resourceType: string, resourceId: string, context: MCPContext): Promise<any> {
|
||||
logger.debug(`Acquiring resource: ${resourceType}/${resourceId}`);
|
||||
|
||||
// Initialize resource type map if not exists
|
||||
if (!this.resources.has(resourceType)) {
|
||||
this.resources.set(resourceType, new Map());
|
||||
}
|
||||
|
||||
const typeResources = this.resources.get(resourceType);
|
||||
|
||||
// Create resource if it doesn't exist
|
||||
if (!typeResources.has(resourceId)) {
|
||||
// Create a placeholder for the resource
|
||||
const resourceData = {
|
||||
id: resourceId,
|
||||
type: resourceType,
|
||||
createdAt: Date.now(),
|
||||
data: {}
|
||||
};
|
||||
|
||||
// Store the resource
|
||||
typeResources.set(resourceId, resourceData);
|
||||
|
||||
// Log resource creation
|
||||
await Promise.resolve(); // Add await to satisfy linter
|
||||
logger.debug(`Created new resource: ${resourceType}/${resourceId}`);
|
||||
|
||||
return resourceData;
|
||||
}
|
||||
|
||||
// Return existing resource
|
||||
return typeResources.get(resourceId);
|
||||
}
|
||||
|
||||
/**
|
||||
* Resource release
|
||||
*/
|
||||
private async releaseResource(resourceType: string, resourceId: string, context: MCPContext): Promise<void> {
|
||||
logger.debug(`Releasing resource: ${resourceType}/${resourceId}`);
|
||||
|
||||
// Check if type exists
|
||||
if (!this.resources.has(resourceType)) {
|
||||
return;
|
||||
}
|
||||
|
||||
const typeResources = this.resources.get(resourceType);
|
||||
|
||||
// Remove resource if it exists
|
||||
if (typeResources.has(resourceId)) {
|
||||
await Promise.resolve(); // Add await to satisfy linter
|
||||
typeResources.delete(resourceId);
|
||||
logger.debug(`Released resource: ${resourceType}/${resourceId}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* List available resources
|
||||
*/
|
||||
private async listResources(context: MCPContext, resourceType?: string): Promise<string[]> {
|
||||
if (resourceType) {
|
||||
logger.debug(`Listing resources of type ${resourceType}`);
|
||||
|
||||
if (!this.resources.has(resourceType)) {
|
||||
return [];
|
||||
}
|
||||
|
||||
await Promise.resolve(); // Add await to satisfy linter
|
||||
return Array.from(this.resources.get(resourceType).keys());
|
||||
} else {
|
||||
logger.debug('Listing all resource types');
|
||||
await Promise.resolve(); // Add await to satisfy linter
|
||||
return Array.from(this.resources.keys());
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Start the server
|
||||
*/
|
||||
public async start(): Promise<void> {
|
||||
logger.info("Starting MCP Server");
|
||||
this.emit(MCPServerEvents.STARTING);
|
||||
|
||||
// Start all transports
|
||||
for (const transport of this.transports) {
|
||||
await transport.start();
|
||||
}
|
||||
|
||||
this.emit(MCPServerEvents.STARTED);
|
||||
logger.info("MCP Server started");
|
||||
}
|
||||
|
||||
/**
|
||||
* Gracefully shut down the server
|
||||
*/
|
||||
public async shutdown(): Promise<void> {
|
||||
logger.info("Shutting down MCP Server");
|
||||
this.emit(MCPServerEvents.SHUTTING_DOWN);
|
||||
|
||||
// Stop all transports
|
||||
for (const transport of this.transports) {
|
||||
await transport.stop();
|
||||
}
|
||||
|
||||
// Clear resources
|
||||
this.tools.clear();
|
||||
this.middlewares = [];
|
||||
this.transports = [];
|
||||
this.resources.clear();
|
||||
|
||||
this.emit(MCPServerEvents.SHUTDOWN);
|
||||
this.removeAllListeners();
|
||||
logger.info("MCP Server shut down");
|
||||
}
|
||||
}
|
||||
153
src/mcp/index.ts
Normal file
153
src/mcp/index.ts
Normal file
@@ -0,0 +1,153 @@
|
||||
/**
|
||||
* MCP - Model Context Protocol Implementation
|
||||
*
|
||||
* This is the main entry point for the MCP implementation.
|
||||
* It exports all the components needed to use the MCP.
|
||||
*/
|
||||
|
||||
// Core MCP components
|
||||
export * from './MCPServer.js';
|
||||
export * from './types.js';
|
||||
export * from './BaseTool.js';
|
||||
|
||||
// Middleware
|
||||
export * from './middleware/index.js';
|
||||
|
||||
// Transports
|
||||
export * from './transports/stdio.transport.js';
|
||||
export * from './transports/http.transport.js';
|
||||
|
||||
// Utilities for AI assistants
|
||||
export * from './utils/claude.js';
|
||||
export * from './utils/cursor.js';
|
||||
export * from './utils/error.js';
|
||||
|
||||
// Helper function to create Claude-compatible tool definitions
|
||||
export function createClaudeToolDefinitions(tools: any[]): any[] {
|
||||
return tools.map(tool => {
|
||||
// Convert Zod schema to JSON Schema
|
||||
const parameters = tool.parameters ? {
|
||||
type: 'object',
|
||||
properties: {},
|
||||
required: []
|
||||
} : {
|
||||
type: 'object',
|
||||
properties: {},
|
||||
required: []
|
||||
};
|
||||
|
||||
return {
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
parameters
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
// Helper function to create Cursor-compatible tool definitions
|
||||
export function createCursorToolDefinitions(tools: any[]): any[] {
|
||||
return tools.map(tool => {
|
||||
// Convert to Cursor format
|
||||
return {
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
parameters: {}
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Model Context Protocol (MCP) Module
|
||||
*
|
||||
* This module provides the core MCP server implementation along with
|
||||
* tools, transports, and utilities for integrating with Claude and Cursor.
|
||||
*/
|
||||
|
||||
// Export server implementation
|
||||
export { MCPServer } from "./MCPServer.js";
|
||||
|
||||
// Export type definitions
|
||||
export * from "./types.js";
|
||||
|
||||
// Export transport layers
|
||||
export { StdioTransport } from "./transports/stdio.transport.js";
|
||||
|
||||
// Re-export tools base class
|
||||
export { BaseTool } from "../tools/base-tool.js";
|
||||
|
||||
// Re-export middleware
|
||||
export * from "./middleware/index.js";
|
||||
|
||||
// Import types for proper type definitions
|
||||
import { MCPServer } from "./MCPServer.js";
|
||||
import { StdioTransport } from "./transports/stdio.transport.js";
|
||||
import { ToolDefinition } from "./types.js";
|
||||
|
||||
/**
|
||||
* Utility function to create Claude-compatible function definitions
|
||||
*/
|
||||
export function createClaudeFunctions(tools: ToolDefinition[]): any[] {
|
||||
return tools.map(tool => {
|
||||
// If the tool has a toSchemaObject method, use it
|
||||
if ('toSchemaObject' in tool && typeof tool.toSchemaObject === 'function') {
|
||||
return tool.toSchemaObject();
|
||||
}
|
||||
|
||||
// Otherwise, manually convert the tool to a Claude function
|
||||
return {
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
parameters: {
|
||||
type: "object",
|
||||
properties: (tool as any).parameters?.properties || {},
|
||||
required: (tool as any).parameters?.required || []
|
||||
}
|
||||
};
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Utility function to create Cursor-compatible tool definitions
|
||||
*/
|
||||
export function createCursorTools(tools: ToolDefinition[]): any[] {
|
||||
return tools.map(tool => ({
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
parameters: Object.entries((tool as any).parameters?.properties || {}).reduce((acc, [key, value]) => {
|
||||
const param = value as any;
|
||||
acc[key] = {
|
||||
type: param.type || 'string',
|
||||
description: param.description || '',
|
||||
required: ((tool as any).parameters?.required || []).includes(key)
|
||||
};
|
||||
return acc;
|
||||
}, {} as Record<string, any>)
|
||||
}));
|
||||
}
|
||||
|
||||
/**
|
||||
* Create a standalone MCP server with stdio transport
|
||||
*/
|
||||
export function createStdioServer(options: {
|
||||
silent?: boolean;
|
||||
debug?: boolean;
|
||||
tools?: ToolDefinition[];
|
||||
} = {}): { server: MCPServer; transport: StdioTransport } {
|
||||
// Create server instance
|
||||
const server = MCPServer.getInstance();
|
||||
|
||||
// Create and register stdio transport
|
||||
const transport = new StdioTransport({
|
||||
silent: options.silent,
|
||||
debug: options.debug
|
||||
});
|
||||
|
||||
server.registerTransport(transport);
|
||||
|
||||
// Register tools if provided
|
||||
if (options.tools && Array.isArray(options.tools)) {
|
||||
server.registerTools(options.tools);
|
||||
}
|
||||
|
||||
return { server, transport };
|
||||
}
|
||||
172
src/mcp/middleware/index.ts
Normal file
172
src/mcp/middleware/index.ts
Normal file
@@ -0,0 +1,172 @@
|
||||
/**
|
||||
* MCP Middleware System
|
||||
*
|
||||
* This module provides middleware functionality for the MCP server,
|
||||
* allowing for request/response processing pipelines.
|
||||
*/
|
||||
|
||||
import { MCPMiddleware, MCPRequest, MCPResponse, MCPContext, MCPErrorCode } from "../types.js";
|
||||
import { logger } from "../../utils/logger.js";
|
||||
|
||||
/**
|
||||
* Middleware for validating requests against JSON Schema
|
||||
*/
|
||||
export const validationMiddleware: MCPMiddleware = async (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> => {
|
||||
const { method } = request;
|
||||
|
||||
const tool = context.tools.get(method);
|
||||
if (!tool) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.METHOD_NOT_FOUND,
|
||||
message: `Method not found: ${method}`
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
if (tool.parameters && request.params) {
|
||||
try {
|
||||
// Zod validation happens here
|
||||
const validatedParams = tool.parameters.parse(request.params);
|
||||
request.params = validatedParams;
|
||||
} catch (error) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.INVALID_PARAMS,
|
||||
message: "Invalid parameters",
|
||||
data: error instanceof Error ? error.message : String(error)
|
||||
}
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
return next();
|
||||
};
|
||||
|
||||
/**
|
||||
* Middleware for handling authentication
|
||||
*/
|
||||
export const authMiddleware = (authKey: string): MCPMiddleware => {
|
||||
return async (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> => {
|
||||
// Check for authentication in params
|
||||
const authToken = (request.params)?.auth_token;
|
||||
|
||||
if (!authToken || authToken !== authKey) {
|
||||
return {
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.AUTHENTICATION_ERROR,
|
||||
message: "Authentication failed"
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
// Remove auth token from params to keep them clean
|
||||
if (request.params && typeof request.params === 'object') {
|
||||
const { auth_token, ...cleanParams } = request.params;
|
||||
request.params = cleanParams;
|
||||
}
|
||||
|
||||
return next();
|
||||
};
|
||||
};
|
||||
|
||||
/**
|
||||
* Middleware for logging requests and responses
|
||||
*/
|
||||
export const loggingMiddleware: MCPMiddleware = async (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> => {
|
||||
const startTime = Date.now();
|
||||
logger.debug(`MCP Request: ${request.method}`, {
|
||||
id: request.id,
|
||||
method: request.method
|
||||
});
|
||||
|
||||
try {
|
||||
const response = await next();
|
||||
|
||||
const duration = Date.now() - startTime;
|
||||
logger.debug(`MCP Response: ${request.method}`, {
|
||||
id: request.id,
|
||||
method: request.method,
|
||||
success: !response.error,
|
||||
duration
|
||||
});
|
||||
|
||||
return response;
|
||||
} catch (error) {
|
||||
const duration = Date.now() - startTime;
|
||||
logger.error(`MCP Error: ${request.method}`, {
|
||||
id: request.id,
|
||||
method: request.method,
|
||||
error,
|
||||
duration
|
||||
});
|
||||
|
||||
throw error;
|
||||
}
|
||||
};
|
||||
|
||||
/**
|
||||
* Middleware for handling timeouts
|
||||
*/
|
||||
export const timeoutMiddleware = (timeoutMs: number): MCPMiddleware => {
|
||||
return async (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> => {
|
||||
return Promise.race([
|
||||
next(),
|
||||
new Promise<MCPResponse>((resolve) => {
|
||||
setTimeout(() => {
|
||||
resolve({
|
||||
id: request.id,
|
||||
error: {
|
||||
code: MCPErrorCode.TIMEOUT,
|
||||
message: `Request timed out after ${timeoutMs}ms`
|
||||
}
|
||||
});
|
||||
}, timeoutMs);
|
||||
})
|
||||
]);
|
||||
};
|
||||
};
|
||||
|
||||
/**
|
||||
* Utility to combine multiple middlewares
|
||||
*/
|
||||
export function combineMiddlewares(middlewares: MCPMiddleware[]): MCPMiddleware {
|
||||
return async (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
): Promise<MCPResponse> => {
|
||||
// Create a function that runs through all middlewares
|
||||
let index = 0;
|
||||
|
||||
const runMiddleware = async (): Promise<MCPResponse> => {
|
||||
if (index < middlewares.length) {
|
||||
const middleware = middlewares[index++];
|
||||
return middleware(request, context, runMiddleware);
|
||||
} else {
|
||||
return next();
|
||||
}
|
||||
};
|
||||
|
||||
return runMiddleware();
|
||||
};
|
||||
}
|
||||
42
src/mcp/transport.ts
Normal file
42
src/mcp/transport.ts
Normal file
@@ -0,0 +1,42 @@
|
||||
/**
|
||||
* Base Transport for MCP
|
||||
*
|
||||
* This module provides a base class for all transport implementations.
|
||||
*/
|
||||
|
||||
import { TransportLayer, MCPRequest, MCPResponse, MCPStreamPart, MCPNotification } from "./types.js";
|
||||
|
||||
/**
|
||||
* Abstract base class for all transports
|
||||
*/
|
||||
export abstract class BaseTransport implements TransportLayer {
|
||||
public name: string = "base";
|
||||
protected handler: ((request: MCPRequest) => Promise<MCPResponse>) | null = null;
|
||||
|
||||
/**
|
||||
* Initialize the transport with a request handler
|
||||
*/
|
||||
public initialize(handler: (request: MCPRequest) => Promise<MCPResponse>): void {
|
||||
this.handler = handler;
|
||||
}
|
||||
|
||||
/**
|
||||
* Start the transport
|
||||
*/
|
||||
public abstract start(): Promise<void>;
|
||||
|
||||
/**
|
||||
* Stop the transport
|
||||
*/
|
||||
public abstract stop(): Promise<void>;
|
||||
|
||||
/**
|
||||
* Send a notification to a client
|
||||
*/
|
||||
public sendNotification?(notification: MCPNotification): void;
|
||||
|
||||
/**
|
||||
* Send a streaming response part
|
||||
*/
|
||||
public sendStreamPart?(streamPart: MCPStreamPart): void;
|
||||
}
|
||||
426
src/mcp/transports/http.transport.ts
Normal file
426
src/mcp/transports/http.transport.ts
Normal file
@@ -0,0 +1,426 @@
|
||||
/**
|
||||
* HTTP Transport for MCP
|
||||
*
|
||||
* This module implements a JSON-RPC 2.0 transport layer over HTTP/HTTPS
|
||||
* for the Model Context Protocol. It supports both traditional request/response
|
||||
* patterns as well as streaming responses via Server-Sent Events (SSE).
|
||||
*/
|
||||
|
||||
import { Server as HttpServer } from "http";
|
||||
import express, { Express, Request, Response, NextFunction } from "express";
|
||||
// Using a direct import now that we have the types
|
||||
import cors from "cors";
|
||||
import { TransportLayer, MCPRequest, MCPResponse, MCPStreamPart, MCPNotification, MCPErrorCode } from "../types.js";
|
||||
import { logger } from "../../utils/logger.js";
|
||||
import { EventEmitter } from "events";
|
||||
|
||||
type ServerSentEventsClient = {
|
||||
id: string;
|
||||
response: Response;
|
||||
};
|
||||
|
||||
/**
|
||||
* Implementation of TransportLayer using HTTP/Express
|
||||
*/
|
||||
export class HttpTransport implements TransportLayer {
|
||||
public name = "http";
|
||||
private handler: ((request: MCPRequest) => Promise<MCPResponse>) | null = null;
|
||||
private app: Express;
|
||||
private server: HttpServer | null = null;
|
||||
private sseClients: Map<string, ServerSentEventsClient>;
|
||||
private events: EventEmitter;
|
||||
private initialized = false;
|
||||
private port: number;
|
||||
private corsOrigin: string | string[];
|
||||
private apiPrefix: string;
|
||||
private debug: boolean;
|
||||
|
||||
/**
|
||||
* Constructor for HttpTransport
|
||||
*/
|
||||
constructor(options: {
|
||||
port?: number;
|
||||
corsOrigin?: string | string[];
|
||||
apiPrefix?: string;
|
||||
debug?: boolean;
|
||||
} = {}) {
|
||||
this.port = options.port ?? (process.env.PORT ? parseInt(process.env.PORT, 10) : 3000);
|
||||
this.corsOrigin = options.corsOrigin ?? (process.env.CORS_ORIGIN || '*');
|
||||
this.apiPrefix = options.apiPrefix ?? '/api';
|
||||
this.debug = options.debug ?? (process.env.DEBUG_HTTP === "true");
|
||||
this.app = express();
|
||||
this.sseClients = new Map();
|
||||
this.events = new EventEmitter();
|
||||
|
||||
// Configure max event listeners
|
||||
this.events.setMaxListeners(100);
|
||||
}
|
||||
|
||||
/**
|
||||
* Initialize the transport with a request handler
|
||||
*/
|
||||
public initialize(handler: (request: MCPRequest) => Promise<MCPResponse>): void {
|
||||
if (this.initialized) {
|
||||
throw new Error("HttpTransport already initialized");
|
||||
}
|
||||
|
||||
this.handler = handler;
|
||||
this.initialized = true;
|
||||
|
||||
// Setup middleware
|
||||
this.setupMiddleware();
|
||||
|
||||
// Setup routes
|
||||
this.setupRoutes();
|
||||
|
||||
logger.info("HTTP transport initialized");
|
||||
}
|
||||
|
||||
/**
|
||||
* Setup Express middleware
|
||||
*/
|
||||
private setupMiddleware(): void {
|
||||
// JSON body parser
|
||||
this.app.use(express.json({ limit: '1mb' }));
|
||||
|
||||
// CORS configuration
|
||||
// Using the imported cors middleware
|
||||
try {
|
||||
this.app.use(cors({
|
||||
origin: this.corsOrigin,
|
||||
methods: ['GET', 'POST', 'OPTIONS'],
|
||||
allowedHeaders: ['Content-Type', 'Authorization'],
|
||||
credentials: true
|
||||
}));
|
||||
} catch (err) {
|
||||
logger.warn(`CORS middleware not available: ${String(err)}`);
|
||||
}
|
||||
|
||||
// Request logging
|
||||
if (this.debug) {
|
||||
this.app.use((req, res, next) => {
|
||||
logger.debug(`HTTP ${req.method} ${req.url}`);
|
||||
next();
|
||||
});
|
||||
}
|
||||
|
||||
// Error handling middleware
|
||||
this.app.use((err: Error, req: Request, res: Response, next: NextFunction) => {
|
||||
logger.error(`Express error: ${err.message}`);
|
||||
res.status(500).json({
|
||||
jsonrpc: "2.0",
|
||||
id: null,
|
||||
error: {
|
||||
code: MCPErrorCode.INTERNAL_ERROR,
|
||||
message: "Internal server error",
|
||||
data: this.debug ? { stack: err.stack } : undefined
|
||||
}
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Setup Express routes
|
||||
*/
|
||||
private setupRoutes(): void {
|
||||
// Health check endpoint
|
||||
this.app.get('/health', (req: Request, res: Response) => {
|
||||
res.status(200).json({
|
||||
status: 'ok',
|
||||
transport: 'http',
|
||||
timestamp: new Date().toISOString()
|
||||
});
|
||||
});
|
||||
|
||||
// Server info endpoint
|
||||
this.app.get(`${this.apiPrefix}/info`, (req: Request, res: Response) => {
|
||||
res.status(200).json({
|
||||
jsonrpc: "2.0",
|
||||
result: {
|
||||
name: "Model Context Protocol Server",
|
||||
version: "1.0.0",
|
||||
transport: "http",
|
||||
protocol: "json-rpc-2.0",
|
||||
features: ["streaming"],
|
||||
timestamp: new Date().toISOString()
|
||||
}
|
||||
});
|
||||
});
|
||||
|
||||
// SSE stream endpoint
|
||||
this.app.get(`${this.apiPrefix}/stream`, (req: Request, res: Response) => {
|
||||
const clientId = (req.query.clientId as string) || `client-${Date.now()}-${Math.random().toString(36).substring(2, 9)}`;
|
||||
|
||||
// Set headers for SSE
|
||||
res.setHeader('Content-Type', 'text/event-stream');
|
||||
res.setHeader('Cache-Control', 'no-cache');
|
||||
res.setHeader('Connection', 'keep-alive');
|
||||
|
||||
// Store the client
|
||||
this.sseClients.set(clientId, { id: clientId, response: res });
|
||||
|
||||
// Send initial connection established event
|
||||
res.write(`event: connected\ndata: ${JSON.stringify({ clientId })}\n\n`);
|
||||
|
||||
// Client disconnection handler
|
||||
req.on('close', () => {
|
||||
if (this.debug) {
|
||||
logger.debug(`SSE client disconnected: ${clientId}`);
|
||||
}
|
||||
this.sseClients.delete(clientId);
|
||||
});
|
||||
|
||||
if (this.debug) {
|
||||
logger.debug(`SSE client connected: ${clientId}`);
|
||||
}
|
||||
});
|
||||
|
||||
// JSON-RPC endpoint
|
||||
this.app.post(`${this.apiPrefix}/jsonrpc`, (req: Request, res: Response) => {
|
||||
void this.handleJsonRpcRequest(req, res);
|
||||
});
|
||||
|
||||
// Default 404 handler
|
||||
this.app.use((req: Request, res: Response) => {
|
||||
res.status(404).json({
|
||||
jsonrpc: "2.0",
|
||||
id: null,
|
||||
error: {
|
||||
code: MCPErrorCode.METHOD_NOT_FOUND,
|
||||
message: "Not found"
|
||||
}
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Handle a JSON-RPC request from HTTP
|
||||
*/
|
||||
private async handleJsonRpcRequest(req: Request, res: Response): Promise<void> {
|
||||
if (!this.handler) {
|
||||
res.status(500).json({
|
||||
jsonrpc: "2.0",
|
||||
id: req.body.id || null,
|
||||
error: {
|
||||
code: MCPErrorCode.INTERNAL_ERROR,
|
||||
message: "Transport not properly initialized"
|
||||
}
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
try {
|
||||
// Validate it's JSON-RPC 2.0
|
||||
if (!req.body.jsonrpc || req.body.jsonrpc !== "2.0") {
|
||||
res.status(400).json({
|
||||
jsonrpc: "2.0",
|
||||
id: req.body.id || null,
|
||||
error: {
|
||||
code: MCPErrorCode.INVALID_REQUEST,
|
||||
message: "Invalid JSON-RPC 2.0 request: missing or invalid jsonrpc version"
|
||||
}
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
// Check for batch requests
|
||||
if (Array.isArray(req.body)) {
|
||||
res.status(501).json({
|
||||
jsonrpc: "2.0",
|
||||
id: null,
|
||||
error: {
|
||||
code: MCPErrorCode.METHOD_NOT_FOUND,
|
||||
message: "Batch requests are not supported"
|
||||
}
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
// Handle request
|
||||
const request: MCPRequest = {
|
||||
jsonrpc: req.body.jsonrpc,
|
||||
id: req.body.id ?? null,
|
||||
method: req.body.method,
|
||||
params: req.body.params
|
||||
};
|
||||
|
||||
// Get streaming preference from query params
|
||||
const useStreaming = req.query.stream === 'true';
|
||||
|
||||
// Extract client ID if provided (for streaming)
|
||||
const clientId = (req.query.clientId as string) || (req.body.clientId as string);
|
||||
|
||||
// Check if this is a streaming request and client is connected
|
||||
if (useStreaming && clientId && this.sseClients.has(clientId)) {
|
||||
// Add streaming metadata to the request
|
||||
request.streaming = {
|
||||
enabled: true,
|
||||
clientId
|
||||
};
|
||||
}
|
||||
|
||||
// Process the request
|
||||
const response = await this.handler(request);
|
||||
|
||||
// Return the response
|
||||
res.status(200).json({
|
||||
jsonrpc: "2.0",
|
||||
...response
|
||||
});
|
||||
} catch (error) {
|
||||
logger.error(`Error handling JSON-RPC request: ${String(error)}`);
|
||||
|
||||
res.status(500).json({
|
||||
jsonrpc: "2.0",
|
||||
id: req.body?.id || null,
|
||||
error: {
|
||||
code: MCPErrorCode.INTERNAL_ERROR,
|
||||
message: error instanceof Error ? error.message : "Internal error",
|
||||
data: this.debug && error instanceof Error ? { stack: error.stack } : undefined
|
||||
}
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Start the HTTP server
|
||||
*/
|
||||
public async start(): Promise<void> {
|
||||
if (!this.initialized) {
|
||||
throw new Error("HttpTransport not initialized");
|
||||
}
|
||||
|
||||
return new Promise<void>((resolve, reject) => {
|
||||
try {
|
||||
this.server = this.app.listen(this.port, () => {
|
||||
logger.info(`HTTP transport started on port ${this.port}`);
|
||||
resolve();
|
||||
});
|
||||
|
||||
// Error handler
|
||||
this.server.on('error', (err) => {
|
||||
logger.error(`HTTP server error: ${String(err)}`);
|
||||
reject(err);
|
||||
});
|
||||
} catch (err) {
|
||||
logger.error(`Failed to start HTTP transport: ${String(err)}`);
|
||||
reject(err);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Stop the HTTP server
|
||||
*/
|
||||
public async stop(): Promise<void> {
|
||||
return new Promise<void>((resolve, reject) => {
|
||||
// Close server if running
|
||||
if (this.server) {
|
||||
this.server.close((err) => {
|
||||
if (err) {
|
||||
logger.error(`Error shutting down HTTP server: ${String(err)}`);
|
||||
reject(err);
|
||||
} else {
|
||||
logger.info("HTTP transport stopped");
|
||||
this.server = null;
|
||||
resolve();
|
||||
}
|
||||
});
|
||||
} else {
|
||||
resolve();
|
||||
}
|
||||
|
||||
// Close all SSE connections
|
||||
for (const client of this.sseClients.values()) {
|
||||
try {
|
||||
client.response.write(`event: shutdown\ndata: {}\n\n`);
|
||||
client.response.end();
|
||||
} catch (err) {
|
||||
logger.error(`Error closing SSE connection: ${String(err)}`);
|
||||
}
|
||||
}
|
||||
|
||||
// Clear all clients
|
||||
this.sseClients.clear();
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Send an SSE event to a specific client
|
||||
*/
|
||||
private sendSSEEvent(clientId: string, event: string, data: unknown): boolean {
|
||||
const client = this.sseClients.get(clientId);
|
||||
if (!client) {
|
||||
return false;
|
||||
}
|
||||
|
||||
try {
|
||||
const payload = JSON.stringify(data);
|
||||
client.response.write(`event: ${event}\ndata: ${payload}\n\n`);
|
||||
return true;
|
||||
} catch (err) {
|
||||
logger.error(`Error sending SSE event: ${String(err)}`);
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Send a notification to a client
|
||||
*/
|
||||
public sendNotification(notification: MCPNotification): void {
|
||||
// SSE notifications not supported without a client ID
|
||||
return;
|
||||
}
|
||||
|
||||
/**
|
||||
* Send a streaming response part
|
||||
*/
|
||||
public sendStreamPart(streamPart: MCPStreamPart): void {
|
||||
// Find the client ID in streaming metadata
|
||||
const clientId = streamPart.clientId;
|
||||
if (!clientId || !this.sseClients.has(clientId)) {
|
||||
logger.warn(`Cannot send stream part: client ${clientId || 'unknown'} not connected`);
|
||||
return;
|
||||
}
|
||||
|
||||
// Send the stream part as an SSE event
|
||||
const eventPayload = {
|
||||
jsonrpc: "2.0",
|
||||
id: streamPart.id,
|
||||
stream: {
|
||||
partId: streamPart.partId,
|
||||
final: streamPart.final,
|
||||
data: streamPart.data
|
||||
}
|
||||
};
|
||||
|
||||
this.sendSSEEvent(clientId, 'stream', eventPayload);
|
||||
|
||||
// Debug logging
|
||||
if (this.debug) {
|
||||
logger.debug(`Sent stream part to client ${clientId}: partId=${streamPart.partId}, final=${streamPart.final}`);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Broadcast a notification to all connected clients
|
||||
*/
|
||||
public broadcastNotification(event: string, data: unknown): void {
|
||||
for (const client of this.sseClients.values()) {
|
||||
try {
|
||||
const payload = JSON.stringify(data);
|
||||
client.response.write(`event: ${event}\ndata: ${payload}\n\n`);
|
||||
} catch (err) {
|
||||
logger.error(`Error broadcasting to client ${client.id}: ${String(err)}`);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Send a log message (not applicable for HTTP transport)
|
||||
*/
|
||||
public sendLogMessage(level: string, message: string, data?: unknown): void {
|
||||
// Log messages in HTTP context go to the logger, not to clients
|
||||
logger[level as keyof typeof logger]?.(message, data);
|
||||
}
|
||||
}
|
||||
329
src/mcp/transports/stdio.transport.ts
Normal file
329
src/mcp/transports/stdio.transport.ts
Normal file
@@ -0,0 +1,329 @@
|
||||
/**
|
||||
* Stdio Transport for MCP
|
||||
*
|
||||
* This module provides a transport that uses standard input/output
|
||||
* for JSON-RPC 2.0 communication. This is particularly useful for
|
||||
* integration with AI assistants like Claude, GPT, and Cursor.
|
||||
*/
|
||||
|
||||
import { BaseTransport } from "../transport.js";
|
||||
import { logger } from "../../utils/logger.js";
|
||||
import { MCPServer } from "../MCPServer.js";
|
||||
import type { MCPRequest, MCPResponse, ToolExecutionResult } from "../types.js";
|
||||
import { JSONRPCError } from "../utils/error.js";
|
||||
|
||||
/**
|
||||
* StdioTransport options
|
||||
*/
|
||||
export interface StdioTransportOptions {
|
||||
/** Whether to enable silent mode (suppress non-essential output) */
|
||||
silent?: boolean;
|
||||
/** Whether to enable debug mode */
|
||||
debug?: boolean;
|
||||
/** Reference to an MCPServer instance */
|
||||
server?: MCPServer;
|
||||
}
|
||||
|
||||
/**
|
||||
* Transport implementation for standard input/output
|
||||
* Communicates using JSON-RPC 2.0 protocol
|
||||
*/
|
||||
export class StdioTransport extends BaseTransport {
|
||||
private isStarted = false;
|
||||
private silent: boolean;
|
||||
private debug: boolean;
|
||||
private server: MCPServer | null = null;
|
||||
|
||||
constructor(options: StdioTransportOptions = {}) {
|
||||
super();
|
||||
this.silent = options.silent ?? false;
|
||||
this.debug = options.debug ?? false;
|
||||
|
||||
if (options.server) {
|
||||
this.server = options.server;
|
||||
}
|
||||
|
||||
// Configure stdin to not buffer input
|
||||
process.stdin.setEncoding('utf8');
|
||||
}
|
||||
|
||||
/**
|
||||
* Set the server reference to access tools and other server properties
|
||||
*/
|
||||
public setServer(server: MCPServer): void {
|
||||
this.server = server;
|
||||
}
|
||||
|
||||
/**
|
||||
* Start the transport and setup stdin/stdout handlers
|
||||
*/
|
||||
public async start(): Promise<void> {
|
||||
if (this.isStarted) return;
|
||||
|
||||
if (!this.silent) {
|
||||
logger.info('Starting stdio transport');
|
||||
}
|
||||
|
||||
// Setup input handling
|
||||
this.setupInputHandling();
|
||||
|
||||
this.isStarted = true;
|
||||
|
||||
if (!this.silent) {
|
||||
logger.info('Stdio transport started');
|
||||
}
|
||||
|
||||
// Send system info notification
|
||||
this.sendSystemInfo();
|
||||
|
||||
// Send available tools notification
|
||||
this.sendAvailableTools();
|
||||
}
|
||||
|
||||
/**
|
||||
* Send system information as a notification
|
||||
* This helps clients understand the capabilities of the server
|
||||
*/
|
||||
private sendSystemInfo(): void {
|
||||
const notification = {
|
||||
jsonrpc: '2.0',
|
||||
method: 'system.info',
|
||||
params: {
|
||||
name: 'Home Assistant Model Context Protocol Server',
|
||||
version: '1.0.0',
|
||||
transport: 'stdio',
|
||||
protocol: 'json-rpc-2.0',
|
||||
features: ['streaming'],
|
||||
timestamp: new Date().toISOString()
|
||||
}
|
||||
};
|
||||
|
||||
// Send directly to stdout
|
||||
process.stdout.write(JSON.stringify(notification) + '\n');
|
||||
}
|
||||
|
||||
/**
|
||||
* Send available tools as a notification
|
||||
* This helps clients know what tools are available to use
|
||||
*/
|
||||
private sendAvailableTools(): void {
|
||||
if (!this.server) {
|
||||
logger.warn('Cannot send available tools: server reference not set');
|
||||
return;
|
||||
}
|
||||
|
||||
const tools = this.server.getAllTools().map(tool => {
|
||||
// For parameters, create a simple JSON schema or empty object
|
||||
const parameters = tool.parameters
|
||||
? { type: 'object', properties: {} } // Simplified schema
|
||||
: { type: 'object', properties: {} };
|
||||
|
||||
return {
|
||||
name: tool.name,
|
||||
description: tool.description,
|
||||
parameters,
|
||||
metadata: tool.metadata
|
||||
};
|
||||
});
|
||||
|
||||
const notification = {
|
||||
jsonrpc: '2.0',
|
||||
method: 'tools.available',
|
||||
params: { tools }
|
||||
};
|
||||
|
||||
// Send directly to stdout
|
||||
process.stdout.write(JSON.stringify(notification) + '\n');
|
||||
}
|
||||
|
||||
/**
|
||||
* Set up the input handling for JSON-RPC requests
|
||||
*/
|
||||
private setupInputHandling(): void {
|
||||
let buffer = '';
|
||||
|
||||
process.stdin.on('data', (chunk: string) => {
|
||||
buffer += chunk;
|
||||
|
||||
try {
|
||||
// Look for complete JSON objects by matching opening and closing braces
|
||||
let startIndex = 0;
|
||||
let openBraces = 0;
|
||||
let inString = false;
|
||||
let escapeNext = false;
|
||||
|
||||
for (let i = 0; i < buffer.length; i++) {
|
||||
const char = buffer[i];
|
||||
|
||||
if (escapeNext) {
|
||||
escapeNext = false;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (char === '\\' && inString) {
|
||||
escapeNext = true;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (char === '"' && !escapeNext) {
|
||||
inString = !inString;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (!inString) {
|
||||
if (char === '{') {
|
||||
if (openBraces === 0) {
|
||||
startIndex = i;
|
||||
}
|
||||
openBraces++;
|
||||
} else if (char === '}') {
|
||||
openBraces--;
|
||||
|
||||
if (openBraces === 0) {
|
||||
// We have a complete JSON object
|
||||
const jsonStr = buffer.substring(startIndex, i + 1);
|
||||
this.handleJsonRequest(jsonStr);
|
||||
|
||||
// Remove the processed part from the buffer
|
||||
buffer = buffer.substring(i + 1);
|
||||
|
||||
// Reset the parser to start from the beginning of the new buffer
|
||||
i = -1;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
} catch (error) {
|
||||
if (this.debug) {
|
||||
logger.error('Error processing JSON-RPC input', error);
|
||||
}
|
||||
|
||||
this.sendErrorResponse(null, new JSONRPCError.ParseError('Invalid JSON'));
|
||||
}
|
||||
});
|
||||
|
||||
process.stdin.on('end', () => {
|
||||
if (!this.silent) {
|
||||
logger.info('Stdio transport: stdin ended');
|
||||
}
|
||||
process.exit(0);
|
||||
});
|
||||
|
||||
process.stdin.on('error', (error) => {
|
||||
logger.error('Stdio transport: stdin error', error);
|
||||
process.exit(1);
|
||||
});
|
||||
}
|
||||
|
||||
/**
|
||||
* Handle a JSON-RPC request
|
||||
*/
|
||||
private async handleJsonRequest(jsonStr: string): Promise<void> {
|
||||
try {
|
||||
const request = JSON.parse(jsonStr);
|
||||
|
||||
if (this.debug) {
|
||||
logger.debug(`Received request: ${jsonStr}`);
|
||||
}
|
||||
|
||||
if (!request.jsonrpc || request.jsonrpc !== '2.0') {
|
||||
return this.sendErrorResponse(
|
||||
request.id,
|
||||
new JSONRPCError.InvalidRequest('Invalid JSON-RPC 2.0 request')
|
||||
);
|
||||
}
|
||||
|
||||
const mcpRequest: MCPRequest = {
|
||||
jsonrpc: request.jsonrpc,
|
||||
id: request.id,
|
||||
method: request.method,
|
||||
params: request.params || {}
|
||||
};
|
||||
|
||||
if (!this.server) {
|
||||
return this.sendErrorResponse(
|
||||
request.id,
|
||||
new JSONRPCError.InternalError('Server not available')
|
||||
);
|
||||
}
|
||||
|
||||
// Delegate to the server to handle the request
|
||||
if (this.handler) {
|
||||
const response = await this.handler(mcpRequest);
|
||||
this.sendResponse(response);
|
||||
}
|
||||
|
||||
} catch (error) {
|
||||
if (error instanceof SyntaxError) {
|
||||
this.sendErrorResponse(null, new JSONRPCError.ParseError('Invalid JSON'));
|
||||
} else {
|
||||
this.sendErrorResponse(null, new JSONRPCError.InternalError('Internal error'));
|
||||
}
|
||||
|
||||
if (this.debug) {
|
||||
logger.error('Error handling JSON-RPC request', error);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Send a JSON-RPC error response
|
||||
*/
|
||||
private sendErrorResponse(id: string | number | null, error: JSONRPCError.JSONRPCError): void {
|
||||
const response = {
|
||||
jsonrpc: '2.0',
|
||||
id: id,
|
||||
error: {
|
||||
code: error.code,
|
||||
message: error.message,
|
||||
data: error.data
|
||||
}
|
||||
};
|
||||
|
||||
process.stdout.write(JSON.stringify(response) + '\n');
|
||||
}
|
||||
|
||||
/**
|
||||
* Send an MCPResponse to the client
|
||||
*/
|
||||
public sendResponse(response: MCPResponse): void {
|
||||
const jsonRpcResponse = {
|
||||
jsonrpc: '2.0',
|
||||
id: response.id,
|
||||
...(response.error
|
||||
? { error: response.error }
|
||||
: { result: response.result })
|
||||
};
|
||||
|
||||
process.stdout.write(JSON.stringify(jsonRpcResponse) + '\n');
|
||||
}
|
||||
|
||||
/**
|
||||
* Stream a partial response for long-running operations
|
||||
*/
|
||||
public streamResponsePart(requestId: string | number, result: ToolExecutionResult): void {
|
||||
const streamResponse = {
|
||||
jsonrpc: '2.0',
|
||||
method: 'stream.data',
|
||||
params: {
|
||||
id: requestId,
|
||||
data: result
|
||||
}
|
||||
};
|
||||
|
||||
process.stdout.write(JSON.stringify(streamResponse) + '\n');
|
||||
}
|
||||
|
||||
/**
|
||||
* Stop the transport
|
||||
*/
|
||||
public async stop(): Promise<void> {
|
||||
if (!this.isStarted) return;
|
||||
|
||||
if (!this.silent) {
|
||||
logger.info('Stopping stdio transport');
|
||||
}
|
||||
|
||||
this.isStarted = false;
|
||||
}
|
||||
}
|
||||
220
src/mcp/types.ts
Normal file
220
src/mcp/types.ts
Normal file
@@ -0,0 +1,220 @@
|
||||
/**
|
||||
* MCP Type Definitions
|
||||
*
|
||||
* This file contains all the type definitions used by the Model Context Protocol
|
||||
* implementation, including tools, transports, middleware, and resources.
|
||||
*/
|
||||
|
||||
import { z } from "zod";
|
||||
import { Logger } from "winston";
|
||||
import { MCPServer, MCPErrorCode, MCPServerEvents } from "./MCPServer.js";
|
||||
|
||||
/**
|
||||
* MCP Server configuration
|
||||
*/
|
||||
export interface MCPConfig {
|
||||
maxRetries: number;
|
||||
retryDelay: number;
|
||||
executionTimeout: number;
|
||||
streamingEnabled: boolean;
|
||||
maxPayloadSize: number;
|
||||
}
|
||||
|
||||
// Re-export enums from MCPServer
|
||||
export { MCPErrorCode, MCPServerEvents };
|
||||
|
||||
/**
|
||||
* Tool definition interface
|
||||
*/
|
||||
export interface ToolDefinition {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters?: z.ZodType<any>;
|
||||
returnType?: z.ZodType<any>;
|
||||
execute: (params: any, context: MCPContext) => Promise<any>;
|
||||
metadata?: ToolMetadata;
|
||||
}
|
||||
|
||||
/**
|
||||
* Tool metadata for categorization and discovery
|
||||
*/
|
||||
export interface ToolMetadata {
|
||||
category: string;
|
||||
version: string;
|
||||
tags?: string[];
|
||||
platforms?: string[];
|
||||
requiresAuth?: boolean;
|
||||
isStreaming?: boolean;
|
||||
examples?: ToolExample[];
|
||||
}
|
||||
|
||||
/**
|
||||
* Example usage for a tool
|
||||
*/
|
||||
export interface ToolExample {
|
||||
description: string;
|
||||
params: any;
|
||||
expectedResult?: any;
|
||||
}
|
||||
|
||||
/**
|
||||
* JSON-RPC Request
|
||||
*/
|
||||
export interface MCPRequest {
|
||||
jsonrpc: string;
|
||||
id: string | number | null;
|
||||
method: string;
|
||||
params?: Record<string, unknown>;
|
||||
streaming?: {
|
||||
enabled: boolean;
|
||||
clientId: string;
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* JSON-RPC 2.0 Response
|
||||
*/
|
||||
export interface MCPResponse {
|
||||
jsonrpc?: string;
|
||||
id?: string | number;
|
||||
result?: any;
|
||||
error?: MCPError;
|
||||
}
|
||||
|
||||
/**
|
||||
* JSON-RPC 2.0 Error
|
||||
*/
|
||||
export interface MCPError {
|
||||
code: number;
|
||||
message: string;
|
||||
data?: any;
|
||||
}
|
||||
|
||||
/**
|
||||
* JSON-RPC 2.0 Notification
|
||||
*/
|
||||
export interface MCPNotification {
|
||||
jsonrpc?: string;
|
||||
method: string;
|
||||
params?: any;
|
||||
}
|
||||
|
||||
/**
|
||||
* JSON-RPC Stream Part
|
||||
*/
|
||||
export interface MCPStreamPart {
|
||||
id: string | number;
|
||||
partId: string | number;
|
||||
final: boolean;
|
||||
data: unknown;
|
||||
clientId?: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Response Stream Interface for streaming operation results
|
||||
*/
|
||||
export interface MCPResponseStream {
|
||||
/**
|
||||
* Write partial result data to the stream
|
||||
*
|
||||
* @param data The partial result data
|
||||
* @returns True if the write was successful, false otherwise
|
||||
*/
|
||||
write(data: any): boolean;
|
||||
|
||||
/**
|
||||
* End the stream, indicating no more data will be sent
|
||||
*
|
||||
* @param data Optional final data to send
|
||||
*/
|
||||
end(data?: any): void;
|
||||
|
||||
/**
|
||||
* Check if streaming is enabled
|
||||
*/
|
||||
readonly isEnabled: boolean;
|
||||
|
||||
/**
|
||||
* Get the client ID for this stream
|
||||
*/
|
||||
readonly clientId?: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Context for tool execution
|
||||
*/
|
||||
export interface MCPContext {
|
||||
requestId: string | number;
|
||||
startTime: number;
|
||||
resourceManager: ResourceManager;
|
||||
tools: Map<string, ToolDefinition>;
|
||||
config: MCPConfig;
|
||||
logger: Logger;
|
||||
server: MCPServer;
|
||||
state?: Map<string, any>;
|
||||
}
|
||||
|
||||
/**
|
||||
* Resource manager interface
|
||||
*/
|
||||
export interface ResourceManager {
|
||||
acquire: (resourceType: string, resourceId: string, context: MCPContext) => Promise<any>;
|
||||
release: (resourceType: string, resourceId: string, context: MCPContext) => Promise<void>;
|
||||
list: (context: MCPContext, resourceType?: string) => Promise<string[]>;
|
||||
}
|
||||
|
||||
/**
|
||||
* Middleware function type
|
||||
*/
|
||||
export type MCPMiddleware = (
|
||||
request: MCPRequest,
|
||||
context: MCPContext,
|
||||
next: () => Promise<MCPResponse>
|
||||
) => Promise<MCPResponse>;
|
||||
|
||||
/**
|
||||
* Transport layer interface
|
||||
*/
|
||||
export interface TransportLayer {
|
||||
name: string;
|
||||
initialize: (handler: (request: MCPRequest) => Promise<MCPResponse>) => void;
|
||||
start: () => Promise<void>;
|
||||
stop: () => Promise<void>;
|
||||
sendNotification?: (notification: MCPNotification) => void;
|
||||
sendStreamPart?: (streamPart: MCPStreamPart) => void;
|
||||
}
|
||||
|
||||
/**
|
||||
* Claude-specific function call formats
|
||||
*/
|
||||
export interface ClaudeFunctionDefinition {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters: {
|
||||
type: string;
|
||||
properties: Record<string, {
|
||||
type: string;
|
||||
description: string;
|
||||
enum?: string[];
|
||||
}>;
|
||||
required: string[];
|
||||
};
|
||||
}
|
||||
|
||||
/**
|
||||
* Cursor-specific integration types
|
||||
*/
|
||||
export interface CursorToolDefinition {
|
||||
name: string;
|
||||
description: string;
|
||||
parameters: Record<string, {
|
||||
type: string;
|
||||
description: string;
|
||||
required: boolean;
|
||||
}>;
|
||||
}
|
||||
|
||||
/**
|
||||
* Tool execution result type used in streaming responses
|
||||
*/
|
||||
export type ToolExecutionResult = any;
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user