Implements Phases 1-8 of the TFTSR implementation plan. Rust backend (Tauri 2.x, src-tauri/): - Multi-provider AI: OpenAI-compatible, Anthropic, Gemini, Mistral, Ollama - PII detection engine: 11 regex patterns with overlap resolution - SQLCipher AES-256 encrypted database with 10 versioned migrations - 28 Tauri IPC commands for triage, analysis, document, and system ops - Ollama: hardware probe, model recommendations, pull/delete with events - RCA and blameless post-mortem Markdown document generators - PDF export via printpdf - Audit log: SHA-256 hash of every external data send - Integration stubs for Confluence, ServiceNow, Azure DevOps (v0.2) Frontend (React 18 + TypeScript + Vite, src/): - 9 pages: full triage workflow NewIssue→LogUpload→Triage→Resolution→RCA→Postmortem→History+Settings - 7 components: ChatWindow, TriageProgress, PiiDiffViewer, DocEditor, HardwareReport, ModelSelector, UI primitives - 3 Zustand stores: session, settings (persisted), history - Type-safe tauriCommands.ts matching Rust backend types exactly - 8 IT domain system prompts (Linux, Windows, Network, K8s, DB, Virt, HW, Obs) DevOps: - .woodpecker/test.yml: rustfmt, clippy, cargo test, tsc, vitest on every push - .woodpecker/release.yml: linux/amd64 + linux/arm64 builds, Gogs release upload Verified: - cargo check: zero errors - tsc --noEmit: zero errors - vitest run: 13/13 unit tests passing Co-Authored-By: Claude Sonnet 4.6 (1M context) <noreply@anthropic.com>
71 lines
3.3 KiB
JavaScript
71 lines
3.3 KiB
JavaScript
'use strict';
|
|
|
|
Object.defineProperty(exports, "__esModule", {
|
|
value: true
|
|
});
|
|
exports.default = cargo;
|
|
|
|
var _queue = require('./internal/queue.js');
|
|
|
|
var _queue2 = _interopRequireDefault(_queue);
|
|
|
|
function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { default: obj }; }
|
|
|
|
/**
|
|
* Creates a `cargoQueue` object with the specified payload. Tasks added to the
|
|
* cargoQueue will be processed together (up to the `payload` limit) in `concurrency` parallel workers.
|
|
* If the all `workers` are in progress, the task is queued until one becomes available. Once
|
|
* a `worker` has completed some tasks, each callback of those tasks is
|
|
* called. Check out [these](https://camo.githubusercontent.com/6bbd36f4cf5b35a0f11a96dcd2e97711ffc2fb37/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313637363837312f36383130382f62626330636662302d356632392d313165322d393734662d3333393763363464633835382e676966) [animations](https://camo.githubusercontent.com/f4810e00e1c5f5f8addbe3e9f49064fd5d102699/68747470733a2f2f662e636c6f75642e6769746875622e636f6d2f6173736574732f313637363837312f36383130312f38346339323036362d356632392d313165322d383134662d3964336430323431336266642e676966)
|
|
* for how `cargo` and `queue` work.
|
|
*
|
|
* While [`queue`]{@link module:ControlFlow.queue} passes only one task to one of a group of workers
|
|
* at a time, and [`cargo`]{@link module:ControlFlow.cargo} passes an array of tasks to a single worker,
|
|
* the cargoQueue passes an array of tasks to multiple parallel workers.
|
|
*
|
|
* @name cargoQueue
|
|
* @static
|
|
* @memberOf module:ControlFlow
|
|
* @method
|
|
* @see [async.queue]{@link module:ControlFlow.queue}
|
|
* @see [async.cargo]{@link module:ControlFLow.cargo}
|
|
* @category Control Flow
|
|
* @param {AsyncFunction} worker - An asynchronous function for processing an array
|
|
* of queued tasks. Invoked with `(tasks, callback)`.
|
|
* @param {number} [concurrency=1] - An `integer` for determining how many
|
|
* `worker` functions should be run in parallel. If omitted, the concurrency
|
|
* defaults to `1`. If the concurrency is `0`, an error is thrown.
|
|
* @param {number} [payload=Infinity] - An optional `integer` for determining
|
|
* how many tasks should be processed per round; if omitted, the default is
|
|
* unlimited.
|
|
* @returns {module:ControlFlow.QueueObject} A cargoQueue object to manage the tasks. Callbacks can
|
|
* attached as certain properties to listen for specific events during the
|
|
* lifecycle of the cargoQueue and inner queue.
|
|
* @example
|
|
*
|
|
* // create a cargoQueue object with payload 2 and concurrency 2
|
|
* var cargoQueue = async.cargoQueue(function(tasks, callback) {
|
|
* for (var i=0; i<tasks.length; i++) {
|
|
* console.log('hello ' + tasks[i].name);
|
|
* }
|
|
* callback();
|
|
* }, 2, 2);
|
|
*
|
|
* // add some items
|
|
* cargoQueue.push({name: 'foo'}, function(err) {
|
|
* console.log('finished processing foo');
|
|
* });
|
|
* cargoQueue.push({name: 'bar'}, function(err) {
|
|
* console.log('finished processing bar');
|
|
* });
|
|
* cargoQueue.push({name: 'baz'}, function(err) {
|
|
* console.log('finished processing baz');
|
|
* });
|
|
* cargoQueue.push({name: 'boo'}, function(err) {
|
|
* console.log('finished processing boo');
|
|
* });
|
|
*/
|
|
function cargo(worker, concurrency, payload) {
|
|
return (0, _queue2.default)(worker, concurrency, payload);
|
|
}
|
|
module.exports = exports.default; |