chore(security): enhance environment configuration, CI workflows, and wallet daemon with security improvements

- Restructure .env.example with security-focused documentation, service-specific environment file references, and AWS Secrets Manager integration
- Update CLI tests workflow to single Python 3.13 version, add pytest-mock dependency, and consolidate test execution with coverage
- Add comprehensive security validation to package publishing workflow with manual approval gates, secret scanning, and release
This commit is contained in:
oib
2026-03-03 10:33:46 +01:00
parent 00d00cb964
commit f353e00172
220 changed files with 42506 additions and 921 deletions

View File

@@ -30,6 +30,28 @@ Provides GPU-powered LLM inference services through Ollama, allowing miners to e
## Quick Start
### Prerequisites
- Ollama installed and running locally (`ollama serve`)
- At least one model pulled (example: `ollama pull mistral:latest`)
- Python 3.11+ with `pip install -e .` if running from repo root
### Minimal Usage Example
```bash
# 1) Run miner (exposes inference endpoint for jobs)
python3 miner_plugin.py --host 0.0.0.0 --port 8001
# 2) In another terminal, submit a job via client
python3 client_plugin.py chat mistral:latest "Summarize the AITBC marketplace in 3 bullets"
# 3) View logs/results
tail -f miner.log
```
Optional environment variables:
- `OLLAMA_HOST` (default: http://127.0.0.1:11434)
- `OLLAMA_MODELS` (comma-separated list to register; defaults to discovered models)
- `OLLAMA_MAX_CONCURRENCY` (default: 2)
### 1. Start Ollama (if not running)
```bash
ollama serve