tftsr-devops_investigation/node_modules/@promptbook/utils/package.json

57 lines
1.4 KiB
JSON
Raw Normal View History

feat: initial implementation of TFTSR IT Triage & RCA application Implements Phases 1-8 of the TFTSR implementation plan. Rust backend (Tauri 2.x, src-tauri/): - Multi-provider AI: OpenAI-compatible, Anthropic, Gemini, Mistral, Ollama - PII detection engine: 11 regex patterns with overlap resolution - SQLCipher AES-256 encrypted database with 10 versioned migrations - 28 Tauri IPC commands for triage, analysis, document, and system ops - Ollama: hardware probe, model recommendations, pull/delete with events - RCA and blameless post-mortem Markdown document generators - PDF export via printpdf - Audit log: SHA-256 hash of every external data send - Integration stubs for Confluence, ServiceNow, Azure DevOps (v0.2) Frontend (React 18 + TypeScript + Vite, src/): - 9 pages: full triage workflow NewIssue→LogUpload→Triage→Resolution→RCA→Postmortem→History+Settings - 7 components: ChatWindow, TriageProgress, PiiDiffViewer, DocEditor, HardwareReport, ModelSelector, UI primitives - 3 Zustand stores: session, settings (persisted), history - Type-safe tauriCommands.ts matching Rust backend types exactly - 8 IT domain system prompts (Linux, Windows, Network, K8s, DB, Virt, HW, Obs) DevOps: - .woodpecker/test.yml: rustfmt, clippy, cargo test, tsc, vitest on every push - .woodpecker/release.yml: linux/amd64 + linux/arm64 builds, Gogs release upload Verified: - cargo check: zero errors - tsc --noEmit: zero errors - vitest run: 13/13 unit tests passing Co-Authored-By: Claude Sonnet 4.6 (1M context) <noreply@anthropic.com>
2026-03-15 03:36:25 +00:00
{
"name": "@promptbook/utils",
"version": "0.69.5",
"description": "Build responsible, controlled and transparent applications on top of LLM models!",
"private": false,
"sideEffects": false,
"repository": {
"type": "git",
"url": "https://github.com/webgptorg/promptbook"
},
"contributors": [
"Pavol Hejný <me@pavolhejny.com> (https://www.pavolhejny.com/)"
],
"keywords": [
"ai",
"llm",
"prompt",
"template",
"language-model",
"chatgpt",
"autogpt",
"machine-learning",
"natural-language-processing",
"nlp",
"openai",
"gpt-3",
"gpt-4",
"gpt-4o",
"gpt-4o-mini",
"o1",
"o1-mini",
"o1-preview",
"anthropic"
],
"license": "CC-BY-4.0",
"bugs": {
"url": "https://github.com/webgptorg/promptbook/issues"
},
"homepage": "https://www.npmjs.com/package/@promptbook/core",
"funding": [
{
"type": "individual",
"url": "https://buymeacoffee.com/hejny"
},
{
"type": "github",
"url": "https://github.com/webgptorg/promptbook/blob/main/README.md#%EF%B8%8F-contributing"
}
],
"main": "./umd/index.umd.js",
"module": "./esm/index.es.js",
"typings": "./esm/typings/src/_packages/utils.index.d.ts",
"dependencies": {
"spacetrim": "0.11.59"
}
}