Add workspace utility scripts

- llm: Local LLM wrapper for llama-swap
- homelab-status: Quick K8s/cluster health check
- calc: Python/JS REPL for quick calculations
- transcribe: Whisper audio transcription wrapper

Added to fish PATH.
This commit is contained in:
William Valentin
2026-01-26 22:53:52 -08:00
parent 3aea7b4050
commit 2c3d6afcdd
5 changed files with 441 additions and 0 deletions

76
scripts/calc Executable file
View File

@@ -0,0 +1,76 @@
#!/usr/bin/env bash
# Quick code execution for calculations and data processing
# Usage: calc "expression"
# calc -p "python code"
# calc -j "javascript code"
set -e
MODE="python"
CODE=""
show_help() {
echo "Usage: calc [options] \"code\""
echo ""
echo "Options:"
echo " -p, --python Python mode (default)"
echo " -j, --js JavaScript mode (Node.js)"
echo " -h, --help Show this help"
echo ""
echo "Examples:"
echo " calc \"2 + 2\" # Quick math"
echo " calc \"sum([1,2,3,4,5])\" # Python functions"
echo " calc -p \"import math; math.pi\" # Python with imports"
echo " calc -j \"[1,2,3].map(x => x*2)\" # JavaScript"
echo ""
echo "Python mode has these pre-imported:"
echo " math, json, re, datetime, random, statistics"
}
while [[ $# -gt 0 ]]; do
case $1 in
-p|--python)
MODE="python"
shift
;;
-j|--js)
MODE="js"
shift
;;
-h|--help)
show_help
exit 0
;;
*)
if [[ -z "$CODE" ]]; then
CODE="$1"
else
CODE="$CODE $1"
fi
shift
;;
esac
done
if [[ -z "$CODE" ]]; then
show_help
exit 1
fi
case $MODE in
python)
python3 -c "
import math, json, re, sys
from datetime import datetime, date, timedelta
from random import random, randint, choice
from statistics import mean, median
from collections import Counter, defaultdict
result = $CODE
if result is not None: print(result)
"
;;
js)
node -e "console.log($CODE)"
;;
esac

135
scripts/homelab-status Executable file
View File

@@ -0,0 +1,135 @@
#!/usr/bin/env bash
# Quick homelab status check
# Usage: homelab-status [--full]
set -e
FULL=false
[[ "$1" == "--full" || "$1" == "-f" ]] && FULL=true
# Colors
RED='\033[0;31m'
GREEN='\033[0;32m'
YELLOW='\033[1;33m'
BLUE='\033[0;34m'
NC='\033[0m' # No Color
ok() { echo -e "${GREEN}✓${NC} $1"; }
warn() { echo -e "${YELLOW}⚠${NC} $1"; }
fail() { echo -e "${RED}✗${NC} $1"; }
info() { echo -e "${BLUE}${NC} $1"; }
echo "═══════════════════════════════════════"
echo " 🏠 HOMELAB STATUS"
echo "═══════════════════════════════════════"
echo ""
# K8s Cluster
echo "☸️ KUBERNETES"
echo "───────────────────────────────────────"
if ! kubectl cluster-info &>/dev/null; then
fail "Cluster unreachable"
else
# Nodes
NODES_TOTAL=$(kubectl get nodes --no-headers 2>/dev/null | wc -l)
NODES_READY=$(kubectl get nodes --no-headers 2>/dev/null | grep -c " Ready" || echo 0)
if [[ "$NODES_READY" == "$NODES_TOTAL" ]]; then
ok "Nodes: $NODES_READY/$NODES_TOTAL Ready"
else
warn "Nodes: $NODES_READY/$NODES_TOTAL Ready"
fi
# Pods
PODS_RUNNING=$(kubectl get pods -A --no-headers 2>/dev/null | grep -c "Running" || echo 0)
PODS_NOT_RUNNING=$(kubectl get pods -A --no-headers 2>/dev/null | grep -cvE "Running|Completed" || echo 0)
if [[ "$PODS_NOT_RUNNING" == "0" ]]; then
ok "Pods: $PODS_RUNNING Running"
else
warn "Pods: $PODS_RUNNING Running, $PODS_NOT_RUNNING not ready"
fi
# Show failed pods if any
if [[ "$PODS_FAILED" != "0" ]] || $FULL; then
kubectl get pods -A --no-headers 2>/dev/null | grep -E "(Error|CrashLoop|Failed|Pending)" | head -5 | while read -r line; do
echo " $line"
done
fi
fi
echo ""
# Alertmanager
echo "🚨 ALERTS"
echo "───────────────────────────────────────"
ALERTMANAGER_URL="http://alertmanager.monitoring.192.168.153.240.nip.io"
ALERTS=$(curl -sf "$ALERTMANAGER_URL/api/v2/alerts?active=true&silenced=false" 2>/dev/null || echo "[]")
ALERT_COUNT=$(echo "$ALERTS" | jq 'length' 2>/dev/null || echo "?")
if [[ "$ALERT_COUNT" == "0" ]]; then
ok "No active alerts"
elif [[ "$ALERT_COUNT" == "?" ]]; then
warn "Alertmanager unreachable"
else
warn "$ALERT_COUNT active alert(s)"
echo "$ALERTS" | jq -r '.[].labels.alertname' 2>/dev/null | head -5 | while read -r alert; do
echo " - $alert"
done
fi
echo ""
# Key Services
echo "🔌 SERVICES"
echo "───────────────────────────────────────"
check_service() {
local name="$1"
local url="$2"
if curl -sf --max-time 3 "$url" >/dev/null 2>&1; then
ok "$name"
else
fail "$name"
fi
}
check_service "Grafana" "http://grafana.monitoring.192.168.153.240.nip.io/api/health"
check_service "ArgoCD" "https://argocd.taildb3494.ts.net" 2>/dev/null || check_service "ArgoCD" "http://argocd-server.argocd.svc:80" 2>/dev/null || warn "ArgoCD (via Tailscale only)"
check_service "Longhorn" "http://ui.longhorn-system.192.168.153.240.nip.io"
# Ollama (homelab)
if curl -sf --max-time 3 "http://100.85.116.57:11434/api/tags" >/dev/null 2>&1; then
ok "Ollama (homelab)"
else
warn "Ollama (homelab) - unreachable"
fi
echo ""
# Local LLM
echo "🤖 LOCAL LLM"
echo "───────────────────────────────────────"
if curl -sf "http://127.0.0.1:8080/health" >/dev/null 2>&1; then
ok "llama-swap running"
MODELS=$(curl -s "http://127.0.0.1:8080/v1/models" | jq -r '.data[].id' 2>/dev/null | tr '\n' ', ' | sed 's/,$//')
info "Models: $MODELS"
else
warn "llama-swap not running"
info "Start: systemctl --user start llama-swap"
fi
echo ""
# Storage (if full)
if $FULL; then
echo "💾 STORAGE"
echo "───────────────────────────────────────"
kubectl get pvc -A --no-headers 2>/dev/null | head -10 | while read -r ns name status vol cap mode class age; do
if [[ "$status" == "Bound" ]]; then
ok "$ns/$name ($cap)"
else
warn "$ns/$name - $status"
fi
done
echo ""
fi
echo "═══════════════════════════════════════"
echo " $(date '+%Y-%m-%d %H:%M:%S %Z')"
echo "═══════════════════════════════════════"

99
scripts/llm Executable file
View File

@@ -0,0 +1,99 @@
#!/usr/bin/env bash
# Local LLM wrapper for llama-swap
# Usage: llm [model] "prompt"
# llm -m model "prompt"
# echo "prompt" | llm [model]
set -e
ENDPOINT="${LLAMA_SWAP_URL:-http://127.0.0.1:8080}"
DEFAULT_MODEL="${LLAMA_SWAP_MODEL:-gemma}"
MAX_TOKENS="${LLAMA_SWAP_MAX_TOKENS:-2048}"
# Parse args
MODEL="$DEFAULT_MODEL"
PROMPT=""
while [[ $# -gt 0 ]]; do
case $1 in
-m|--model)
MODEL="$2"
shift 2
;;
-t|--tokens)
MAX_TOKENS="$2"
shift 2
;;
-h|--help)
echo "Usage: llm [-m model] [-t max_tokens] \"prompt\""
echo ""
echo "Models: gemma, qwen3, coder, glm, reasoning, gpt-oss"
echo ""
echo "Examples:"
echo " llm \"What is 2+2?\""
echo " llm -m coder \"Write a Python hello world\""
echo " echo \"Explain this\" | llm qwen3"
echo ""
echo "Environment:"
echo " LLAMA_SWAP_URL Endpoint (default: http://127.0.0.1:8080)"
echo " LLAMA_SWAP_MODEL Default model (default: gemma)"
echo " LLAMA_SWAP_MAX_TOKENS Max tokens (default: 2048)"
exit 0
;;
-*)
echo "Unknown option: $1" >&2
exit 1
;;
*)
if [[ -z "$PROMPT" ]]; then
# Check if it's a model name
if [[ "$1" =~ ^(gemma|qwen3|coder|glm|reasoning|gpt-oss)$ ]]; then
MODEL="$1"
else
PROMPT="$1"
fi
else
PROMPT="$PROMPT $1"
fi
shift
;;
esac
done
# Read from stdin if no prompt
if [[ -z "$PROMPT" ]]; then
if [[ ! -t 0 ]]; then
PROMPT=$(cat)
else
echo "Error: No prompt provided" >&2
exit 1
fi
fi
# Check if llama-swap is running
if ! curl -sf "$ENDPOINT/health" >/dev/null 2>&1; then
echo "Error: llama-swap not running at $ENDPOINT" >&2
echo "Start with: systemctl --user start llama-swap" >&2
exit 1
fi
# Build JSON payload
JSON=$(jq -n \
--arg model "$MODEL" \
--arg prompt "$PROMPT" \
--argjson max_tokens "$MAX_TOKENS" \
'{model: $model, messages: [{role: "user", content: $prompt}], max_tokens: $max_tokens}')
# Make request and extract response
RESPONSE=$(curl -s "$ENDPOINT/v1/chat/completions" \
-H "Content-Type: application/json" \
-d "$JSON")
# Check for error
if echo "$RESPONSE" | jq -e '.error' >/dev/null 2>&1; then
echo "Error: $(echo "$RESPONSE" | jq -r '.error.message')" >&2
exit 1
fi
# Extract content
echo "$RESPONSE" | jq -r '.choices[0].message.content'

94
scripts/transcribe Executable file
View File

@@ -0,0 +1,94 @@
#!/usr/bin/env bash
# Transcribe audio files using Whisper
# Usage: transcribe <audio_file> [options]
# transcribe recording.mp3
# transcribe -m medium meeting.wav
set -e
MODEL="${WHISPER_MODEL:-base}"
LANGUAGE=""
OUTPUT_FORMAT="txt"
AUDIO_FILE=""
show_help() {
echo "Usage: transcribe [options] <audio_file>"
echo ""
echo "Options:"
echo " -m, --model MODEL Whisper model (tiny, base, small, medium, large)"
echo " Default: base (fast), use medium/large for accuracy"
echo " -l, --language LANG Force language (e.g., en, es, fr)"
echo " -f, --format FORMAT Output format (txt, json, srt, vtt)"
echo " -h, --help Show this help"
echo ""
echo "Examples:"
echo " transcribe meeting.mp3 # Quick transcription"
echo " transcribe -m medium interview.wav # Better accuracy"
echo " transcribe -l en -f srt podcast.mp3 # English subtitles"
echo ""
echo "Models (speed vs accuracy):"
echo " tiny - Fastest, lowest accuracy (~1GB VRAM)"
echo " base - Fast, good accuracy (~1GB VRAM) [default]"
echo " small - Balanced (~2GB VRAM)"
echo " medium - Better accuracy (~5GB VRAM)"
echo " large - Best accuracy (~10GB VRAM)"
echo ""
echo "Environment:"
echo " WHISPER_MODEL Default model (default: base)"
}
while [[ $# -gt 0 ]]; do
case $1 in
-m|--model)
MODEL="$2"
shift 2
;;
-l|--language)
LANGUAGE="$2"
shift 2
;;
-f|--format)
OUTPUT_FORMAT="$2"
shift 2
;;
-h|--help)
show_help
exit 0
;;
-*)
echo "Unknown option: $1" >&2
exit 1
;;
*)
AUDIO_FILE="$1"
shift
;;
esac
done
if [[ -z "$AUDIO_FILE" ]]; then
echo "Error: No audio file provided" >&2
show_help
exit 1
fi
if [[ ! -f "$AUDIO_FILE" ]]; then
echo "Error: File not found: $AUDIO_FILE" >&2
exit 1
fi
# Build whisper command
CMD="whisper \"$AUDIO_FILE\" --model $MODEL --output_format $OUTPUT_FORMAT"
if [[ -n "$LANGUAGE" ]]; then
CMD="$CMD --language $LANGUAGE"
fi
# Run transcription
echo "Transcribing: $AUDIO_FILE (model: $MODEL)" >&2
eval $CMD
# Show output location
BASE=$(basename "$AUDIO_FILE" | sed 's/\.[^.]*$//')
echo "" >&2
echo "Output: ${BASE}.${OUTPUT_FORMAT}" >&2