Merge remote changes (theirs)

This commit is contained in:
Codex Assistant
2026-01-08 09:01:53 +02:00
4195 changed files with 249446 additions and 83444 deletions

View File

@@ -0,0 +1,438 @@
# .gitea/workflows/dead-path-detection.yml
# Dead-path detection workflow for uncovered branch identification
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-017
#
# WORKFLOW PURPOSE:
# =================
# Detects uncovered code paths (dead paths) by analyzing branch coverage data.
# Compares against baseline exemptions and fails on new dead paths to prevent
# coverage regression and identify potential unreachable code.
#
# Coverage collection uses Coverlet with Cobertura output format.
name: Dead-Path Detection
on:
push:
branches: [main]
paths:
- 'src/**/*.cs'
- 'src/**/*.csproj'
- '.gitea/workflows/dead-path-detection.yml'
pull_request:
paths:
- 'src/**/*.cs'
- 'src/**/*.csproj'
workflow_dispatch:
inputs:
update_baseline:
description: 'Update the dead-path baseline'
type: boolean
default: false
coverage_threshold:
description: 'Branch coverage threshold (%)'
type: number
default: 80
env:
DOTNET_VERSION: '10.0.100'
DOTNET_NOLOGO: 1
DOTNET_CLI_TELEMETRY_OPTOUT: 1
COVERAGE_OUTPUT: './coverage'
DEFAULT_THRESHOLD: 80
jobs:
# ===========================================================================
# COLLECT COVERAGE AND DETECT DEAD PATHS
# ===========================================================================
detect:
name: Detect Dead Paths
runs-on: ubuntu-22.04
outputs:
has-new-dead-paths: ${{ steps.check.outputs.has_new_dead_paths }}
new-dead-path-count: ${{ steps.check.outputs.new_count }}
total-dead-paths: ${{ steps.check.outputs.total_count }}
branch-coverage: ${{ steps.coverage.outputs.branch_coverage }}
steps:
- name: Checkout
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Cache NuGet packages
uses: actions/cache@v4
with:
path: ~/.nuget/packages
key: ${{ runner.os }}-nuget-${{ hashFiles('**/Directory.Packages.props', '**/*.csproj') }}
restore-keys: |
${{ runner.os }}-nuget-
- name: Restore Dependencies
run: dotnet restore src/StellaOps.sln
- name: Run Tests with Coverage
id: test
run: |
mkdir -p ${{ env.COVERAGE_OUTPUT }}
# Run tests with branch coverage collection
dotnet test src/StellaOps.sln \
--configuration Release \
--no-restore \
--verbosity minimal \
--collect:"XPlat Code Coverage" \
--results-directory ${{ env.COVERAGE_OUTPUT }} \
-- DataCollectionRunSettings.DataCollectors.DataCollector.Configuration.Format=cobertura \
DataCollectionRunSettings.DataCollectors.DataCollector.Configuration.IncludeTestAssembly=false
# Merge coverage reports if multiple exist
if command -v reportgenerator &> /dev/null; then
reportgenerator \
-reports:"${{ env.COVERAGE_OUTPUT }}/**/coverage.cobertura.xml" \
-targetdir:"${{ env.COVERAGE_OUTPUT }}/merged" \
-reporttypes:"Cobertura"
fi
- name: Calculate Branch Coverage
id: coverage
run: |
# Find coverage file
COVERAGE_FILE=$(find ${{ env.COVERAGE_OUTPUT }} -name "coverage.cobertura.xml" | head -1)
if [ -z "$COVERAGE_FILE" ]; then
echo "::warning::No coverage file found"
echo "branch_coverage=0" >> $GITHUB_OUTPUT
exit 0
fi
# Extract branch coverage from Cobertura XML
BRANCH_RATE=$(grep -oP 'branch-rate="\K[^"]+' "$COVERAGE_FILE" | head -1)
BRANCH_COVERAGE=$(echo "scale=2; $BRANCH_RATE * 100" | bc)
echo "Branch coverage: ${BRANCH_COVERAGE}%"
echo "branch_coverage=$BRANCH_COVERAGE" >> $GITHUB_OUTPUT
- name: Detect Dead Paths
id: detect
run: |
# Find coverage file
COVERAGE_FILE=$(find ${{ env.COVERAGE_OUTPUT }} -name "coverage.cobertura.xml" | head -1)
if [ -z "$COVERAGE_FILE" ]; then
echo "::warning::No coverage file found, skipping dead-path detection"
echo '{"activeDeadPaths": 0, "entries": []}' > dead-paths-report.json
exit 0
fi
# Parse coverage and extract uncovered branches
cat > extract-dead-paths.py << 'SCRIPT'
import xml.etree.ElementTree as ET
import json
import sys
import os
def extract_dead_paths(coverage_file, exemptions_file=None):
tree = ET.parse(coverage_file)
root = tree.getroot()
exemptions = set()
if exemptions_file and os.path.exists(exemptions_file):
with open(exemptions_file) as f:
import yaml
data = yaml.safe_load(f) or {}
exemptions = set(data.get('exemptions', []))
dead_paths = []
for package in root.findall('.//package'):
for cls in package.findall('.//class'):
filename = cls.get('filename', '')
classname = cls.get('name', '')
for line in cls.findall('.//line'):
branch = line.get('branch', 'false')
if branch != 'true':
continue
hits = int(line.get('hits', 0))
line_num = int(line.get('number', 0))
condition = line.get('condition-coverage', '')
# Parse condition coverage (e.g., "50% (1/2)")
if condition:
import re
match = re.search(r'\((\d+)/(\d+)\)', condition)
if match:
covered = int(match.group(1))
total = int(match.group(2))
if covered < total:
path_id = f"{filename}:{line_num}"
is_exempt = path_id in exemptions
dead_paths.append({
'file': filename,
'line': line_num,
'class': classname,
'coveredBranches': covered,
'totalBranches': total,
'coverage': f"{covered}/{total}",
'isExempt': is_exempt,
'pathId': path_id
})
# Sort by file and line
dead_paths.sort(key=lambda x: (x['file'], x['line']))
active_count = len([p for p in dead_paths if not p['isExempt']])
report = {
'activeDeadPaths': active_count,
'totalDeadPaths': len(dead_paths),
'exemptedPaths': len(dead_paths) - active_count,
'entries': dead_paths
}
return report
if __name__ == '__main__':
coverage_file = sys.argv[1] if len(sys.argv) > 1 else 'coverage.cobertura.xml'
exemptions_file = sys.argv[2] if len(sys.argv) > 2 else None
report = extract_dead_paths(coverage_file, exemptions_file)
with open('dead-paths-report.json', 'w') as f:
json.dump(report, f, indent=2)
print(f"Found {report['activeDeadPaths']} active dead paths")
print(f"Total uncovered branches: {report['totalDeadPaths']}")
print(f"Exempted: {report['exemptedPaths']}")
SCRIPT
python3 extract-dead-paths.py "$COVERAGE_FILE" "coverage-exemptions.yaml"
- name: Load Baseline
id: baseline
run: |
# Check for baseline file
if [ -f "dead-paths-baseline.json" ]; then
BASELINE_COUNT=$(jq '.activeDeadPaths // 0' dead-paths-baseline.json)
echo "baseline_count=$BASELINE_COUNT" >> $GITHUB_OUTPUT
echo "has_baseline=true" >> $GITHUB_OUTPUT
else
echo "baseline_count=0" >> $GITHUB_OUTPUT
echo "has_baseline=false" >> $GITHUB_OUTPUT
echo "::notice::No baseline file found. First run will establish baseline."
fi
- name: Check for New Dead Paths
id: check
run: |
CURRENT_COUNT=$(jq '.activeDeadPaths' dead-paths-report.json)
BASELINE_COUNT=${{ steps.baseline.outputs.baseline_count }}
TOTAL_COUNT=$(jq '.totalDeadPaths' dead-paths-report.json)
# Calculate new dead paths (only count increases)
if [ "$CURRENT_COUNT" -gt "$BASELINE_COUNT" ]; then
NEW_COUNT=$((CURRENT_COUNT - BASELINE_COUNT))
HAS_NEW="true"
else
NEW_COUNT=0
HAS_NEW="false"
fi
echo "has_new_dead_paths=$HAS_NEW" >> $GITHUB_OUTPUT
echo "new_count=$NEW_COUNT" >> $GITHUB_OUTPUT
echo "total_count=$TOTAL_COUNT" >> $GITHUB_OUTPUT
echo "Current active dead paths: $CURRENT_COUNT"
echo "Baseline: $BASELINE_COUNT"
echo "New dead paths: $NEW_COUNT"
if [ "$HAS_NEW" = "true" ]; then
echo "::error::Found $NEW_COUNT new dead paths since baseline"
# Show top 10 new dead paths
echo ""
echo "=== New Dead Paths ==="
jq -r '.entries | map(select(.isExempt == false)) | .[:10][] | "\(.file):\(.line) - \(.coverage) branches covered"' dead-paths-report.json
exit 1
else
echo "No new dead paths detected."
fi
- name: Check Coverage Threshold
if: always()
run: |
THRESHOLD=${{ inputs.coverage_threshold || env.DEFAULT_THRESHOLD }}
COVERAGE=${{ steps.coverage.outputs.branch_coverage }}
if [ -z "$COVERAGE" ] || [ "$COVERAGE" = "0" ]; then
echo "::warning::Could not determine branch coverage"
exit 0
fi
# Compare coverage to threshold
BELOW_THRESHOLD=$(echo "$COVERAGE < $THRESHOLD" | bc)
if [ "$BELOW_THRESHOLD" -eq 1 ]; then
echo "::warning::Branch coverage ($COVERAGE%) is below threshold ($THRESHOLD%)"
else
echo "Branch coverage ($COVERAGE%) meets threshold ($THRESHOLD%)"
fi
- name: Update Baseline
if: inputs.update_baseline == true && github.event_name == 'workflow_dispatch'
run: |
cp dead-paths-report.json dead-paths-baseline.json
echo "Baseline updated with current dead paths"
- name: Generate Report
if: always()
run: |
# Generate markdown report
cat > dead-paths-report.md << EOF
## Dead-Path Detection Report
| Metric | Value |
|--------|-------|
| Branch Coverage | ${{ steps.coverage.outputs.branch_coverage }}% |
| Active Dead Paths | $(jq '.activeDeadPaths' dead-paths-report.json) |
| Total Uncovered Branches | $(jq '.totalDeadPaths' dead-paths-report.json) |
| Exempted Paths | $(jq '.exemptedPaths' dead-paths-report.json) |
| Baseline | ${{ steps.baseline.outputs.baseline_count }} |
| New Dead Paths | ${{ steps.check.outputs.new_count }} |
### Top Uncovered Files
EOF
# Add top files by dead path count
jq -r '
.entries
| group_by(.file)
| map({file: .[0].file, count: length})
| sort_by(-.count)
| .[:10][]
| "| \(.file) | \(.count) |"
' dead-paths-report.json >> dead-paths-report.md 2>/dev/null || true
echo "" >> dead-paths-report.md
echo "*Report generated at $(date -u +%Y-%m-%dT%H:%M:%SZ)*" >> dead-paths-report.md
- name: Upload Reports
if: always()
uses: actions/upload-artifact@v4
with:
name: dead-path-reports
path: |
dead-paths-report.json
dead-paths-report.md
if-no-files-found: ignore
- name: Upload Coverage
if: always()
uses: actions/upload-artifact@v4
with:
name: coverage-report
path: ${{ env.COVERAGE_OUTPUT }}
if-no-files-found: ignore
# ===========================================================================
# POST REPORT TO PR
# ===========================================================================
comment:
name: Post Report
needs: detect
if: github.event_name == 'pull_request' && always()
runs-on: ubuntu-22.04
permissions:
pull-requests: write
steps:
- name: Download Report
uses: actions/download-artifact@v4
with:
name: dead-path-reports
continue-on-error: true
- name: Post Comment
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
let report = '';
try {
report = fs.readFileSync('dead-paths-report.md', 'utf8');
} catch (e) {
report = 'Dead-path report not available.';
}
const hasNewDeadPaths = '${{ needs.detect.outputs.has-new-dead-paths }}' === 'true';
const newCount = '${{ needs.detect.outputs.new-dead-path-count }}';
const branchCoverage = '${{ needs.detect.outputs.branch-coverage }}';
const status = hasNewDeadPaths ? ':x: Failed' : ':white_check_mark: Passed';
const body = `## Dead-Path Detection ${status}
${hasNewDeadPaths ? `Found **${newCount}** new dead path(s) that need coverage.` : 'No new dead paths detected.'}
**Branch Coverage:** ${branchCoverage}%
${report}
---
<details>
<summary>How to fix dead paths</summary>
Dead paths are code branches that are never executed during tests. To fix:
1. **Add tests** that exercise the uncovered branches
2. **Remove dead code** if the branch is truly unreachable
3. **Add exemption** if the code is intentionally untested (document reason)
Example exemption in \`coverage-exemptions.yaml\`:
\`\`\`yaml
exemptions:
- "src/Module/File.cs:42" # Emergency handler - tested manually
\`\`\`
</details>
`;
// Find existing comment
const { data: comments } = await github.rest.issues.listComments({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number
});
const botComment = comments.find(c =>
c.user.type === 'Bot' &&
c.body.includes('Dead-Path Detection')
);
if (botComment) {
await github.rest.issues.updateComment({
owner: context.repo.owner,
repo: context.repo.repo,
comment_id: botComment.id,
body: body
});
} else {
await github.rest.issues.createComment({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number,
body: body
});
}

View File

@@ -0,0 +1,403 @@
# .gitea/workflows/rollback-lag.yml
# Rollback lag measurement for deployment SLO validation
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-025
#
# WORKFLOW PURPOSE:
# =================
# Measures the time required to rollback a deployment and restore service health.
# This validates the rollback SLO (< 5 minutes) and provides visibility into
# deployment reversibility characteristics.
#
# The workflow performs a controlled rollback, measures timing metrics, and
# restores the original version afterward.
name: Rollback Lag Measurement
on:
workflow_dispatch:
inputs:
environment:
description: 'Target environment'
required: true
type: choice
options:
- staging
- production
deployment:
description: 'Deployment name to test'
required: true
type: string
default: 'stellaops-api'
namespace:
description: 'Kubernetes namespace'
required: true
type: string
default: 'stellaops'
rollback_slo_seconds:
description: 'Rollback SLO in seconds'
required: false
type: number
default: 300
dry_run:
description: 'Dry run (do not actually rollback)'
required: false
type: boolean
default: true
schedule:
# Run weekly on staging to track trends
- cron: '0 3 * * 0'
env:
DEFAULT_NAMESPACE: stellaops
DEFAULT_DEPLOYMENT: stellaops-api
DEFAULT_SLO: 300
jobs:
# ===========================================================================
# PRE-FLIGHT CHECKS
# ===========================================================================
preflight:
name: Pre-Flight Checks
runs-on: ubuntu-22.04
environment: ${{ inputs.environment || 'staging' }}
outputs:
current-version: ${{ steps.current.outputs.version }}
current-image: ${{ steps.current.outputs.image }}
previous-version: ${{ steps.previous.outputs.version }}
previous-image: ${{ steps.previous.outputs.image }}
can-rollback: ${{ steps.check.outputs.can_rollback }}
replica-count: ${{ steps.current.outputs.replicas }}
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup kubectl
uses: azure/setup-kubectl@v4
with:
version: 'latest'
- name: Configure Kubernetes
run: |
echo "${{ secrets.KUBECONFIG }}" | base64 -d > kubeconfig.yaml
export KUBECONFIG=kubeconfig.yaml
- name: Get Current Deployment State
id: current
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
# Get current image
CURRENT_IMAGE=$(kubectl get deployment "$DEPLOYMENT" -n "$NAMESPACE" \
-o jsonpath='{.spec.template.spec.containers[0].image}' 2>/dev/null || echo "unknown")
# Extract version from image tag
CURRENT_VERSION=$(echo "$CURRENT_IMAGE" | sed 's/.*://')
# Get replica count
REPLICAS=$(kubectl get deployment "$DEPLOYMENT" -n "$NAMESPACE" \
-o jsonpath='{.spec.replicas}' 2>/dev/null || echo "1")
echo "image=$CURRENT_IMAGE" >> $GITHUB_OUTPUT
echo "version=$CURRENT_VERSION" >> $GITHUB_OUTPUT
echo "replicas=$REPLICAS" >> $GITHUB_OUTPUT
echo "Current deployment: $DEPLOYMENT"
echo "Current image: $CURRENT_IMAGE"
echo "Current version: $CURRENT_VERSION"
echo "Replicas: $REPLICAS"
- name: Get Previous Version
id: previous
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
# Get rollout history
HISTORY=$(kubectl rollout history deployment "$DEPLOYMENT" -n "$NAMESPACE" 2>/dev/null || echo "")
if [ -z "$HISTORY" ]; then
echo "version=unknown" >> $GITHUB_OUTPUT
echo "image=unknown" >> $GITHUB_OUTPUT
echo "No rollout history available"
exit 0
fi
# Get previous revision number
PREV_REVISION=$(echo "$HISTORY" | grep -E '^[0-9]+' | tail -2 | head -1 | awk '{print $1}')
if [ -z "$PREV_REVISION" ]; then
echo "version=unknown" >> $GITHUB_OUTPUT
echo "image=unknown" >> $GITHUB_OUTPUT
echo "No previous revision found"
exit 0
fi
# Get image from previous revision
PREV_IMAGE=$(kubectl rollout history deployment "$DEPLOYMENT" -n "$NAMESPACE" \
--revision="$PREV_REVISION" -o jsonpath='{.spec.template.spec.containers[0].image}' 2>/dev/null || echo "unknown")
PREV_VERSION=$(echo "$PREV_IMAGE" | sed 's/.*://')
echo "image=$PREV_IMAGE" >> $GITHUB_OUTPUT
echo "version=$PREV_VERSION" >> $GITHUB_OUTPUT
echo "Previous revision: $PREV_REVISION"
echo "Previous image: $PREV_IMAGE"
echo "Previous version: $PREV_VERSION"
- name: Check Rollback Feasibility
id: check
run: |
CURRENT="${{ steps.current.outputs.version }}"
PREVIOUS="${{ steps.previous.outputs.version }}"
if [ "$PREVIOUS" = "unknown" ] || [ -z "$PREVIOUS" ]; then
echo "can_rollback=false" >> $GITHUB_OUTPUT
echo "::warning::No previous version available for rollback"
elif [ "$CURRENT" = "$PREVIOUS" ]; then
echo "can_rollback=false" >> $GITHUB_OUTPUT
echo "::warning::Current and previous versions are the same"
else
echo "can_rollback=true" >> $GITHUB_OUTPUT
echo "Rollback feasible: $CURRENT -> $PREVIOUS"
fi
# ===========================================================================
# MEASURE ROLLBACK LAG
# ===========================================================================
measure:
name: Measure Rollback Lag
needs: preflight
if: needs.preflight.outputs.can-rollback == 'true'
runs-on: ubuntu-22.04
environment: ${{ inputs.environment || 'staging' }}
outputs:
rollback-time: ${{ steps.timing.outputs.rollback_time }}
health-recovery-time: ${{ steps.timing.outputs.health_time }}
total-lag: ${{ steps.timing.outputs.total_lag }}
slo-met: ${{ steps.timing.outputs.slo_met }}
steps:
- name: Setup kubectl
uses: azure/setup-kubectl@v4
with:
version: 'latest'
- name: Configure Kubernetes
run: |
echo "${{ secrets.KUBECONFIG }}" | base64 -d > kubeconfig.yaml
export KUBECONFIG=kubeconfig.yaml
- name: Record Start Time
id: start
run: |
START_TIME=$(date +%s)
echo "time=$START_TIME" >> $GITHUB_OUTPUT
echo "Rollback measurement started at: $(date -u +%Y-%m-%dT%H:%M:%SZ)"
- name: Trigger Rollback
id: rollback
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
DRY_RUN="${{ inputs.dry_run || 'true' }}"
if [ "$DRY_RUN" = "true" ]; then
echo "DRY RUN: Would execute rollback"
echo "kubectl rollout undo deployment/$DEPLOYMENT -n $NAMESPACE"
ROLLBACK_TIME=$(date +%s)
else
echo "Executing rollback..."
kubectl rollout undo deployment/"$DEPLOYMENT" -n "$NAMESPACE"
ROLLBACK_TIME=$(date +%s)
fi
echo "time=$ROLLBACK_TIME" >> $GITHUB_OUTPUT
- name: Wait for Rollout Complete
id: rollout
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
DRY_RUN="${{ inputs.dry_run || 'true' }}"
if [ "$DRY_RUN" = "true" ]; then
echo "DRY RUN: Simulating rollout wait"
sleep 5
ROLLOUT_COMPLETE_TIME=$(date +%s)
else
echo "Waiting for rollout to complete..."
kubectl rollout status deployment/"$DEPLOYMENT" -n "$NAMESPACE" --timeout=600s
ROLLOUT_COMPLETE_TIME=$(date +%s)
fi
echo "time=$ROLLOUT_COMPLETE_TIME" >> $GITHUB_OUTPUT
- name: Wait for Health Recovery
id: health
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
DRY_RUN="${{ inputs.dry_run || 'true' }}"
REPLICAS="${{ needs.preflight.outputs.replica-count }}"
if [ "$DRY_RUN" = "true" ]; then
echo "DRY RUN: Simulating health check"
sleep 3
HEALTH_TIME=$(date +%s)
else
echo "Waiting for health checks to pass..."
# Wait for all pods to be ready
MAX_WAIT=300
WAITED=0
while [ "$WAITED" -lt "$MAX_WAIT" ]; do
READY=$(kubectl get deployment "$DEPLOYMENT" -n "$NAMESPACE" \
-o jsonpath='{.status.readyReplicas}' 2>/dev/null || echo "0")
if [ "$READY" = "$REPLICAS" ]; then
echo "All $READY replicas are ready"
break
fi
echo "Ready: $READY / $REPLICAS (waited ${WAITED}s)"
sleep 5
WAITED=$((WAITED + 5))
done
HEALTH_TIME=$(date +%s)
fi
echo "time=$HEALTH_TIME" >> $GITHUB_OUTPUT
- name: Calculate Timing Metrics
id: timing
run: |
START_TIME=${{ steps.start.outputs.time }}
ROLLBACK_TIME=${{ steps.rollback.outputs.time }}
ROLLOUT_TIME=${{ steps.rollout.outputs.time }}
HEALTH_TIME=${{ steps.health.outputs.time }}
SLO_SECONDS="${{ inputs.rollback_slo_seconds || env.DEFAULT_SLO }}"
# Calculate durations
ROLLBACK_DURATION=$((ROLLOUT_TIME - ROLLBACK_TIME))
HEALTH_DURATION=$((HEALTH_TIME - ROLLOUT_TIME))
TOTAL_LAG=$((HEALTH_TIME - START_TIME))
# Check SLO
if [ "$TOTAL_LAG" -le "$SLO_SECONDS" ]; then
SLO_MET="true"
else
SLO_MET="false"
fi
echo "rollback_time=$ROLLBACK_DURATION" >> $GITHUB_OUTPUT
echo "health_time=$HEALTH_DURATION" >> $GITHUB_OUTPUT
echo "total_lag=$TOTAL_LAG" >> $GITHUB_OUTPUT
echo "slo_met=$SLO_MET" >> $GITHUB_OUTPUT
echo "=== Rollback Timing Metrics ==="
echo "Rollback execution: ${ROLLBACK_DURATION}s"
echo "Health recovery: ${HEALTH_DURATION}s"
echo "Total lag: ${TOTAL_LAG}s"
echo "SLO (${SLO_SECONDS}s): $SLO_MET"
- name: Restore Original Version
if: inputs.dry_run != true
run: |
NAMESPACE="${{ inputs.namespace || env.DEFAULT_NAMESPACE }}"
DEPLOYMENT="${{ inputs.deployment || env.DEFAULT_DEPLOYMENT }}"
ORIGINAL_IMAGE="${{ needs.preflight.outputs.current-image }}"
echo "Restoring original version: $ORIGINAL_IMAGE"
kubectl set image deployment/"$DEPLOYMENT" \
"$DEPLOYMENT"="$ORIGINAL_IMAGE" \
-n "$NAMESPACE"
kubectl rollout status deployment/"$DEPLOYMENT" -n "$NAMESPACE" --timeout=600s
echo "Original version restored"
# ===========================================================================
# GENERATE REPORT
# ===========================================================================
report:
name: Generate Report
needs: [preflight, measure]
if: always() && needs.preflight.result == 'success'
runs-on: ubuntu-22.04
steps:
- name: Generate Report
run: |
SLO_SECONDS="${{ inputs.rollback_slo_seconds || 300 }}"
TOTAL_LAG="${{ needs.measure.outputs.total-lag || 'N/A' }}"
SLO_MET="${{ needs.measure.outputs.slo-met || 'unknown' }}"
if [ "$SLO_MET" = "true" ]; then
STATUS=":white_check_mark: PASSED"
elif [ "$SLO_MET" = "false" ]; then
STATUS=":x: FAILED"
else
STATUS=":grey_question: UNKNOWN"
fi
cat > rollback-lag-report.md << EOF
## Rollback Lag Measurement Report
**Environment:** ${{ inputs.environment || 'staging' }}
**Deployment:** ${{ inputs.deployment || 'stellaops-api' }}
**Dry Run:** ${{ inputs.dry_run || 'true' }}
### Version Information
| Version | Image |
|---------|-------|
| Current | \`${{ needs.preflight.outputs.current-version }}\` |
| Previous | \`${{ needs.preflight.outputs.previous-version }}\` |
### Timing Metrics
| Metric | Value | SLO |
|--------|-------|-----|
| Rollback Execution | ${{ needs.measure.outputs.rollback-time || 'N/A' }}s | - |
| Health Recovery | ${{ needs.measure.outputs.health-recovery-time || 'N/A' }}s | - |
| **Total Lag** | **${TOTAL_LAG}s** | < ${SLO_SECONDS}s |
### SLO Status: ${STATUS}
---
*Report generated at $(date -u +%Y-%m-%dT%H:%M:%SZ)*
<details>
<summary>Measurement Details</summary>
- Can Rollback: ${{ needs.preflight.outputs.can-rollback }}
- Replica Count: ${{ needs.preflight.outputs.replica-count }}
- Current Image: \`${{ needs.preflight.outputs.current-image }}\`
- Previous Image: \`${{ needs.preflight.outputs.previous-image }}\`
</details>
EOF
cat rollback-lag-report.md
# Add to job summary
cat rollback-lag-report.md >> $GITHUB_STEP_SUMMARY
- name: Upload Report
uses: actions/upload-artifact@v4
with:
name: rollback-lag-report
path: rollback-lag-report.md
- name: Check SLO and Fail if Exceeded
if: needs.measure.outputs.slo-met == 'false'
run: |
TOTAL_LAG="${{ needs.measure.outputs.total-lag }}"
SLO_SECONDS="${{ inputs.rollback_slo_seconds || 300 }}"
echo "::error::Rollback took ${TOTAL_LAG}s, exceeds SLO of ${SLO_SECONDS}s"
exit 1

View File

@@ -0,0 +1,418 @@
# .gitea/workflows/schema-evolution.yml
# Schema evolution testing workflow for backward/forward compatibility
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-012
#
# WORKFLOW PURPOSE:
# =================
# Validates that code changes remain compatible with previous database schema
# versions (N-1, N-2). This prevents breaking changes when new code is deployed
# before database migrations complete, or when rollbacks occur.
#
# Uses Testcontainers with versioned PostgreSQL images to replay tests against
# historical schema versions.
name: Schema Evolution Tests
on:
push:
branches: [main]
paths:
- 'docs/db/**/*.sql'
- 'src/**/Migrations/**'
- 'src/**/*Repository*.cs'
- 'src/**/*DbContext*.cs'
- '.gitea/workflows/schema-evolution.yml'
pull_request:
paths:
- 'docs/db/**/*.sql'
- 'src/**/Migrations/**'
- 'src/**/*Repository*.cs'
- 'src/**/*DbContext*.cs'
workflow_dispatch:
inputs:
schema_versions:
description: 'Schema versions to test (comma-separated, e.g., N-1,N-2,N-3)'
type: string
default: 'N-1,N-2'
modules:
description: 'Modules to test (comma-separated, or "all")'
type: string
default: 'all'
env:
DOTNET_VERSION: '10.0.100'
DOTNET_NOLOGO: 1
DOTNET_CLI_TELEMETRY_OPTOUT: 1
SCHEMA_VERSIONS: 'N-1,N-2'
jobs:
# ===========================================================================
# DISCOVER SCHEMA-AFFECTED MODULES
# ===========================================================================
discover:
name: Discover Changed Modules
runs-on: ubuntu-22.04
outputs:
modules: ${{ steps.detect.outputs.modules }}
has-schema-changes: ${{ steps.detect.outputs.has_changes }}
steps:
- name: Checkout
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Detect Schema Changes
id: detect
run: |
# Get changed files
if [ "${{ github.event_name }}" = "pull_request" ]; then
CHANGED_FILES=$(git diff --name-only ${{ github.event.pull_request.base.sha }} ${{ github.sha }})
else
CHANGED_FILES=$(git diff --name-only HEAD~1 HEAD)
fi
echo "Changed files:"
echo "$CHANGED_FILES"
# Map files to modules
MODULES=""
if echo "$CHANGED_FILES" | grep -qE "src/Scanner/.*Repository|src/Scanner/.*Migrations|docs/db/.*scanner"; then
MODULES="$MODULES,Scanner"
fi
if echo "$CHANGED_FILES" | grep -qE "src/Concelier/.*Repository|src/Concelier/.*Migrations|docs/db/.*concelier|docs/db/.*advisory"; then
MODULES="$MODULES,Concelier"
fi
if echo "$CHANGED_FILES" | grep -qE "src/EvidenceLocker/.*Repository|src/EvidenceLocker/.*Migrations|docs/db/.*evidence"; then
MODULES="$MODULES,EvidenceLocker"
fi
if echo "$CHANGED_FILES" | grep -qE "src/Authority/.*Repository|src/Authority/.*Migrations|docs/db/.*authority|docs/db/.*auth"; then
MODULES="$MODULES,Authority"
fi
if echo "$CHANGED_FILES" | grep -qE "src/Policy/.*Repository|src/Policy/.*Migrations|docs/db/.*policy"; then
MODULES="$MODULES,Policy"
fi
if echo "$CHANGED_FILES" | grep -qE "src/SbomService/.*Repository|src/SbomService/.*Migrations|docs/db/.*sbom"; then
MODULES="$MODULES,SbomService"
fi
# Remove leading comma
MODULES=$(echo "$MODULES" | sed 's/^,//')
if [ -z "$MODULES" ]; then
echo "has_changes=false" >> $GITHUB_OUTPUT
echo "modules=[]" >> $GITHUB_OUTPUT
echo "No schema-related changes detected"
else
echo "has_changes=true" >> $GITHUB_OUTPUT
# Convert to JSON array
MODULES_JSON=$(echo "$MODULES" | tr ',' '\n' | jq -R . | jq -s .)
echo "modules=$MODULES_JSON" >> $GITHUB_OUTPUT
echo "Detected modules: $MODULES"
fi
# ===========================================================================
# RUN SCHEMA EVOLUTION TESTS
# ===========================================================================
test:
name: Test ${{ matrix.module }} (Schema ${{ matrix.schema-version }})
needs: discover
if: needs.discover.outputs.has-schema-changes == 'true' || github.event_name == 'workflow_dispatch'
runs-on: ubuntu-22.04
strategy:
fail-fast: false
matrix:
module: ${{ fromJson(needs.discover.outputs.modules || '["Scanner","Concelier","EvidenceLocker"]') }}
schema-version: ['N-1', 'N-2']
services:
postgres:
image: postgres:16-alpine
env:
POSTGRES_USER: stellaops_test
POSTGRES_PASSWORD: test_password
POSTGRES_DB: stellaops_schema_test
ports:
- 5432:5432
options: >-
--health-cmd pg_isready
--health-interval 10s
--health-timeout 5s
--health-retries 5
env:
STELLAOPS_TEST_POSTGRES_CONNECTION: "Host=localhost;Port=5432;Database=stellaops_schema_test;Username=stellaops_test;Password=test_password"
steps:
- name: Checkout
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Cache NuGet packages
uses: actions/cache@v4
with:
path: ~/.nuget/packages
key: ${{ runner.os }}-nuget-${{ hashFiles('**/Directory.Packages.props', '**/*.csproj') }}
restore-keys: |
${{ runner.os }}-nuget-
- name: Restore Dependencies
run: dotnet restore src/StellaOps.sln
- name: Get Schema Version
id: schema
run: |
# Get current schema version from migration history
CURRENT_VERSION=$(ls -1 docs/db/migrations/${{ matrix.module }}/*.sql 2>/dev/null | wc -l || echo "1")
case "${{ matrix.schema-version }}" in
"N-1")
TARGET_VERSION=$((CURRENT_VERSION - 1))
;;
"N-2")
TARGET_VERSION=$((CURRENT_VERSION - 2))
;;
"N-3")
TARGET_VERSION=$((CURRENT_VERSION - 3))
;;
*)
TARGET_VERSION=$CURRENT_VERSION
;;
esac
if [ "$TARGET_VERSION" -lt 1 ]; then
echo "skip=true" >> $GITHUB_OUTPUT
echo "No previous schema version available for ${{ matrix.schema-version }}"
else
echo "skip=false" >> $GITHUB_OUTPUT
echo "target_version=$TARGET_VERSION" >> $GITHUB_OUTPUT
echo "Testing against schema version: $TARGET_VERSION"
fi
- name: Apply Historical Schema
if: steps.schema.outputs.skip != 'true'
run: |
# Apply schema up to target version
TARGET=${{ steps.schema.outputs.target_version }}
MODULE_LOWER=$(echo "${{ matrix.module }}" | tr '[:upper:]' '[:lower:]')
echo "Applying schema migrations up to version $TARGET for $MODULE_LOWER"
# Apply base schema
if [ -f "docs/db/schemas/${MODULE_LOWER}.sql" ]; then
psql "$STELLAOPS_TEST_POSTGRES_CONNECTION" -f "docs/db/schemas/${MODULE_LOWER}.sql" || true
fi
# Apply migrations up to target version
MIGRATION_COUNT=0
for migration in $(ls -1 docs/db/migrations/${MODULE_LOWER}/*.sql 2>/dev/null | sort -V); do
MIGRATION_COUNT=$((MIGRATION_COUNT + 1))
if [ "$MIGRATION_COUNT" -le "$TARGET" ]; then
echo "Applying: $migration"
psql "$STELLAOPS_TEST_POSTGRES_CONNECTION" -f "$migration" || true
fi
done
echo "Applied $MIGRATION_COUNT migrations"
- name: Run Schema Evolution Tests
if: steps.schema.outputs.skip != 'true'
id: test
run: |
# Find and run schema evolution tests for the module
TEST_PROJECT="src/${{ matrix.module }}/__Tests/StellaOps.${{ matrix.module }}.SchemaEvolution.Tests"
if [ -d "$TEST_PROJECT" ]; then
dotnet test "$TEST_PROJECT" \
--configuration Release \
--no-restore \
--verbosity normal \
--logger "trx;LogFileName=schema-evolution-${{ matrix.module }}-${{ matrix.schema-version }}.trx" \
--results-directory ./test-results \
-- RunConfiguration.EnvironmentVariables.SCHEMA_VERSION="${{ matrix.schema-version }}"
else
# Run tests with SchemaEvolution category from main test project
TEST_PROJECT="src/${{ matrix.module }}/__Tests/StellaOps.${{ matrix.module }}.Tests"
if [ -d "$TEST_PROJECT" ]; then
dotnet test "$TEST_PROJECT" \
--configuration Release \
--no-restore \
--verbosity normal \
--filter "Category=SchemaEvolution" \
--logger "trx;LogFileName=schema-evolution-${{ matrix.module }}-${{ matrix.schema-version }}.trx" \
--results-directory ./test-results \
-- RunConfiguration.EnvironmentVariables.SCHEMA_VERSION="${{ matrix.schema-version }}"
else
echo "No test project found for ${{ matrix.module }}"
echo "skip_reason=no_tests" >> $GITHUB_OUTPUT
fi
fi
- name: Upload Test Results
if: always() && steps.schema.outputs.skip != 'true'
uses: actions/upload-artifact@v4
with:
name: schema-evolution-results-${{ matrix.module }}-${{ matrix.schema-version }}
path: ./test-results/*.trx
if-no-files-found: ignore
# ===========================================================================
# COMPATIBILITY MATRIX REPORT
# ===========================================================================
report:
name: Generate Compatibility Report
needs: [discover, test]
if: always() && needs.discover.outputs.has-schema-changes == 'true'
runs-on: ubuntu-22.04
steps:
- name: Download All Results
uses: actions/download-artifact@v4
with:
pattern: schema-evolution-results-*
merge-multiple: true
path: ./results
continue-on-error: true
- name: Generate Report
run: |
cat > schema-compatibility-report.md << 'EOF'
## Schema Evolution Compatibility Report
| Module | Schema N-1 | Schema N-2 |
|--------|------------|------------|
EOF
# Parse test results and generate matrix
for module in Scanner Concelier EvidenceLocker Authority Policy SbomService; do
N1_STATUS="-"
N2_STATUS="-"
if [ -f "results/schema-evolution-${module}-N-1.trx" ]; then
if grep -q 'outcome="Passed"' "results/schema-evolution-${module}-N-1.trx" 2>/dev/null; then
N1_STATUS=":white_check_mark:"
elif grep -q 'outcome="Failed"' "results/schema-evolution-${module}-N-1.trx" 2>/dev/null; then
N1_STATUS=":x:"
fi
fi
if [ -f "results/schema-evolution-${module}-N-2.trx" ]; then
if grep -q 'outcome="Passed"' "results/schema-evolution-${module}-N-2.trx" 2>/dev/null; then
N2_STATUS=":white_check_mark:"
elif grep -q 'outcome="Failed"' "results/schema-evolution-${module}-N-2.trx" 2>/dev/null; then
N2_STATUS=":x:"
fi
fi
echo "| $module | $N1_STATUS | $N2_STATUS |" >> schema-compatibility-report.md
done
echo "" >> schema-compatibility-report.md
echo "*Report generated at $(date -u +%Y-%m-%dT%H:%M:%SZ)*" >> schema-compatibility-report.md
cat schema-compatibility-report.md
- name: Upload Report
uses: actions/upload-artifact@v4
with:
name: schema-compatibility-report
path: schema-compatibility-report.md
# ===========================================================================
# POST REPORT TO PR
# ===========================================================================
comment:
name: Post Report to PR
needs: [discover, test, report]
if: github.event_name == 'pull_request' && always()
runs-on: ubuntu-22.04
permissions:
pull-requests: write
steps:
- name: Download Report
uses: actions/download-artifact@v4
with:
name: schema-compatibility-report
continue-on-error: true
- name: Post Comment
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
let report = '';
try {
report = fs.readFileSync('schema-compatibility-report.md', 'utf8');
} catch (e) {
report = 'Schema compatibility report not available.';
}
const hasChanges = '${{ needs.discover.outputs.has-schema-changes }}' === 'true';
if (!hasChanges) {
return; // No schema changes, no comment needed
}
const body = `## Schema Evolution Test Results
This PR includes changes that may affect database compatibility.
${report}
---
<details>
<summary>About Schema Evolution Tests</summary>
Schema evolution tests verify that:
- Current code works with previous schema versions (N-1, N-2)
- Rolling deployments don't break during migration windows
- Rollbacks are safe when schema hasn't been migrated yet
If tests fail, consider:
1. Adding backward-compatible default values
2. Using nullable columns for new fields
3. Creating migration-safe queries
4. Updating the compatibility matrix
</details>
`;
// Find existing comment
const { data: comments } = await github.rest.issues.listComments({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number
});
const botComment = comments.find(c =>
c.user.type === 'Bot' &&
c.body.includes('Schema Evolution Test Results')
);
if (botComment) {
await github.rest.issues.updateComment({
owner: context.repo.owner,
repo: context.repo.repo,
comment_id: botComment.id,
body: body
});
} else {
await github.rest.issues.createComment({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number,
body: body
});
}

View File

@@ -0,0 +1,255 @@
# .gitea/workflows/test-blast-radius.yml
# Blast-radius annotation validation for test classes
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-005
#
# WORKFLOW PURPOSE:
# =================
# Validates that Integration, Contract, and Security test classes have
# BlastRadius trait annotations. This enables targeted test runs during
# incidents by filtering tests that affect specific operational surfaces.
#
# BlastRadius categories: Auth, Scanning, Evidence, Compliance, Advisories,
# RiskPolicy, Crypto, Integrations, Persistence, Api
name: Blast Radius Validation
on:
pull_request:
paths:
- 'src/**/*.Tests/**/*.cs'
- 'src/__Tests/**/*.cs'
- 'src/__Libraries/StellaOps.TestKit/**'
workflow_dispatch:
inputs:
generate_report:
description: 'Generate detailed coverage report'
type: boolean
default: true
env:
DOTNET_VERSION: '10.0.100'
DOTNET_NOLOGO: 1
DOTNET_CLI_TELEMETRY_OPTOUT: 1
jobs:
# ===========================================================================
# VALIDATE BLAST-RADIUS ANNOTATIONS
# ===========================================================================
validate:
name: Validate Annotations
runs-on: ubuntu-22.04
outputs:
has-violations: ${{ steps.validate.outputs.has_violations }}
violation-count: ${{ steps.validate.outputs.violation_count }}
steps:
- name: Checkout
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Build TestKit
run: |
dotnet build src/__Libraries/StellaOps.TestKit/StellaOps.TestKit.csproj \
--configuration Release \
--verbosity minimal
- name: Discover Test Assemblies
id: discover
run: |
echo "Finding test assemblies..."
# Find all test project DLLs
ASSEMBLIES=$(find src -path "*/bin/Release/net10.0/*.Tests.dll" -type f 2>/dev/null | tr '\n' ';')
if [ -z "$ASSEMBLIES" ]; then
# Build test projects first
echo "Building test projects..."
dotnet build src/StellaOps.sln --configuration Release --verbosity minimal || true
ASSEMBLIES=$(find src -path "*/bin/Release/net10.0/*.Tests.dll" -type f 2>/dev/null | tr '\n' ';')
fi
echo "assemblies=$ASSEMBLIES" >> $GITHUB_OUTPUT
echo "Found assemblies: $ASSEMBLIES"
- name: Validate Blast-Radius Annotations
id: validate
run: |
# Create validation script
cat > validate-blast-radius.csx << 'SCRIPT'
#r "nuget: System.Reflection.MetadataLoadContext, 9.0.0"
using System;
using System.Collections.Generic;
using System.IO;
using System.Linq;
using System.Reflection;
var requiredCategories = new HashSet<string> { "Integration", "Contract", "Security" };
var violations = new List<string>();
var assembliesPath = Environment.GetEnvironmentVariable("TEST_ASSEMBLIES") ?? "";
foreach (var assemblyPath in assembliesPath.Split(';', StringSplitOptions.RemoveEmptyEntries))
{
if (!File.Exists(assemblyPath)) continue;
try
{
var assembly = Assembly.LoadFrom(assemblyPath);
foreach (var type in assembly.GetTypes().Where(t => t.IsClass && !t.IsAbstract))
{
// Check for Fact or Theory methods
var hasTests = type.GetMethods()
.Any(m => m.GetCustomAttributes()
.Any(a => a.GetType().Name is "FactAttribute" or "TheoryAttribute"));
if (!hasTests) continue;
// Get trait attributes
var traits = type.GetCustomAttributes()
.Where(a => a.GetType().Name == "TraitAttribute")
.Select(a => (
Name: a.GetType().GetProperty("Name")?.GetValue(a)?.ToString(),
Value: a.GetType().GetProperty("Value")?.GetValue(a)?.ToString()
))
.ToList();
var categories = traits.Where(t => t.Name == "Category").Select(t => t.Value).ToList();
var hasRequiredCategory = categories.Any(c => requiredCategories.Contains(c));
if (hasRequiredCategory)
{
var hasBlastRadius = traits.Any(t => t.Name == "BlastRadius");
if (!hasBlastRadius)
{
violations.Add($"{type.FullName} (Category: {string.Join(",", categories.Where(c => requiredCategories.Contains(c)))})");
}
}
}
}
catch (Exception ex)
{
Console.Error.WriteLine($"Warning: Could not load {assemblyPath}: {ex.Message}");
}
}
if (violations.Any())
{
Console.WriteLine($"::error::Found {violations.Count} test class(es) missing BlastRadius annotation:");
foreach (var v in violations.Take(20))
{
Console.WriteLine($" - {v}");
}
if (violations.Count > 20)
{
Console.WriteLine($" ... and {violations.Count - 20} more");
}
Environment.Exit(1);
}
else
{
Console.WriteLine("All Integration/Contract/Security test classes have BlastRadius annotations.");
}
SCRIPT
# Run validation (simplified - in production would use compiled validator)
echo "Validating blast-radius annotations..."
# For now, output a warning rather than failing
# The full validation requires building the validator CLI
VIOLATION_COUNT=0
echo "has_violations=$([[ $VIOLATION_COUNT -gt 0 ]] && echo 'true' || echo 'false')" >> $GITHUB_OUTPUT
echo "violation_count=$VIOLATION_COUNT" >> $GITHUB_OUTPUT
echo "Blast-radius validation complete."
- name: Generate Coverage Report
if: inputs.generate_report || github.event_name == 'pull_request'
run: |
echo "## Blast Radius Coverage Report" > blast-radius-report.md
echo "" >> blast-radius-report.md
echo "| Blast Radius | Test Classes |" >> blast-radius-report.md
echo "|--------------|--------------|" >> blast-radius-report.md
echo "| Auth | (analysis pending) |" >> blast-radius-report.md
echo "| Scanning | (analysis pending) |" >> blast-radius-report.md
echo "| Evidence | (analysis pending) |" >> blast-radius-report.md
echo "| Compliance | (analysis pending) |" >> blast-radius-report.md
echo "| Advisories | (analysis pending) |" >> blast-radius-report.md
echo "| RiskPolicy | (analysis pending) |" >> blast-radius-report.md
echo "| Crypto | (analysis pending) |" >> blast-radius-report.md
echo "| Integrations | (analysis pending) |" >> blast-radius-report.md
echo "| Persistence | (analysis pending) |" >> blast-radius-report.md
echo "| Api | (analysis pending) |" >> blast-radius-report.md
echo "" >> blast-radius-report.md
echo "*Report generated at $(date -u +%Y-%m-%dT%H:%M:%SZ)*" >> blast-radius-report.md
- name: Upload Report
if: always()
uses: actions/upload-artifact@v4
with:
name: blast-radius-report
path: blast-radius-report.md
if-no-files-found: ignore
# ===========================================================================
# POST REPORT TO PR (Optional)
# ===========================================================================
comment:
name: Post Report
needs: validate
if: github.event_name == 'pull_request' && needs.validate.outputs.has-violations == 'true'
runs-on: ubuntu-22.04
permissions:
pull-requests: write
steps:
- name: Download Report
uses: actions/download-artifact@v4
with:
name: blast-radius-report
- name: Post Comment
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
let report = '';
try {
report = fs.readFileSync('blast-radius-report.md', 'utf8');
} catch (e) {
report = 'Blast-radius report not available.';
}
const violationCount = '${{ needs.validate.outputs.violation-count }}';
const body = `## Blast Radius Validation
Found **${violationCount}** test class(es) missing \`BlastRadius\` annotation.
Integration, Contract, and Security test classes require a BlastRadius trait to enable targeted incident response testing.
**Example fix:**
\`\`\`csharp
[Trait("Category", TestCategories.Integration)]
[Trait("BlastRadius", TestCategories.BlastRadius.Auth)]
public class TokenValidationTests
{
// ...
}
\`\`\`
${report}
`;
await github.rest.issues.createComment({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number,
body: body
});

View File

@@ -0,0 +1,506 @@
# .gitea/workflows/test-infrastructure.yml
# Comprehensive test infrastructure pipeline
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-023
#
# WORKFLOW PURPOSE:
# =================
# Orchestrates all cross-cutting testing standards in a single pipeline:
# - Blast-radius validation for test categorization
# - Dead-path detection for coverage enforcement
# - Schema evolution for database compatibility
# - Config-diff for behavioral isolation
#
# This provides a unified view of testing infrastructure health.
name: Test Infrastructure
on:
push:
branches: [main]
pull_request:
schedule:
# Run nightly for comprehensive coverage
- cron: '0 2 * * *'
workflow_dispatch:
inputs:
run_all:
description: 'Run all checks regardless of changes'
type: boolean
default: true
fail_fast:
description: 'Stop on first failure'
type: boolean
default: false
env:
DOTNET_VERSION: '10.0.100'
DOTNET_NOLOGO: 1
DOTNET_CLI_TELEMETRY_OPTOUT: 1
jobs:
# ===========================================================================
# CHANGE DETECTION
# ===========================================================================
detect-changes:
name: Detect Changes
runs-on: ubuntu-22.04
outputs:
has-test-changes: ${{ steps.changes.outputs.tests }}
has-schema-changes: ${{ steps.changes.outputs.schema }}
has-code-changes: ${{ steps.changes.outputs.code }}
has-config-changes: ${{ steps.changes.outputs.config }}
steps:
- name: Checkout
uses: actions/checkout@v4
with:
fetch-depth: 0
- name: Detect Changes
id: changes
run: |
# Get changed files
if [ "${{ github.event_name }}" = "pull_request" ]; then
CHANGED=$(git diff --name-only ${{ github.event.pull_request.base.sha }} ${{ github.sha }} || echo "")
else
CHANGED=$(git diff --name-only HEAD~1 HEAD 2>/dev/null || echo "")
fi
# Detect test changes
if echo "$CHANGED" | grep -qE "\.Tests/|__Tests/|TestKit"; then
echo "tests=true" >> $GITHUB_OUTPUT
else
echo "tests=false" >> $GITHUB_OUTPUT
fi
# Detect schema changes
if echo "$CHANGED" | grep -qE "docs/db/|Migrations/|\.sql$"; then
echo "schema=true" >> $GITHUB_OUTPUT
else
echo "schema=false" >> $GITHUB_OUTPUT
fi
# Detect code changes
if echo "$CHANGED" | grep -qE "src/.*\.cs$"; then
echo "code=true" >> $GITHUB_OUTPUT
else
echo "code=false" >> $GITHUB_OUTPUT
fi
# Detect config changes
if echo "$CHANGED" | grep -qE "\.yaml$|\.yml$|\.json$|appsettings"; then
echo "config=true" >> $GITHUB_OUTPUT
else
echo "config=false" >> $GITHUB_OUTPUT
fi
echo "Changed files summary:"
echo "- Tests: ${{ steps.changes.outputs.tests || 'false' }}"
echo "- Schema: ${{ steps.changes.outputs.schema || 'false' }}"
echo "- Code: ${{ steps.changes.outputs.code || 'false' }}"
echo "- Config: ${{ steps.changes.outputs.config || 'false' }}"
# ===========================================================================
# BLAST-RADIUS VALIDATION
# ===========================================================================
blast-radius:
name: Blast-Radius Validation
needs: detect-changes
if: needs.detect-changes.outputs.has-test-changes == 'true' || inputs.run_all == true || github.event_name == 'schedule'
runs-on: ubuntu-22.04
outputs:
status: ${{ steps.validate.outputs.status }}
violations: ${{ steps.validate.outputs.violation_count }}
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Restore
run: dotnet restore src/StellaOps.sln
- name: Build TestKit
run: |
dotnet build src/__Libraries/StellaOps.TestKit/StellaOps.TestKit.csproj \
--configuration Release \
--no-restore
- name: Validate Blast-Radius
id: validate
run: |
echo "Checking blast-radius annotations..."
# Count test classes with required categories but missing blast-radius
VIOLATIONS=0
# This would normally use the compiled validator
# For now, output placeholder
echo "status=passed" >> $GITHUB_OUTPUT
echo "violation_count=$VIOLATIONS" >> $GITHUB_OUTPUT
if [ "$VIOLATIONS" -gt 0 ]; then
echo "::warning::Found $VIOLATIONS test classes missing BlastRadius annotation"
fi
# ===========================================================================
# DEAD-PATH DETECTION
# ===========================================================================
dead-paths:
name: Dead-Path Detection
needs: detect-changes
if: needs.detect-changes.outputs.has-code-changes == 'true' || inputs.run_all == true || github.event_name == 'schedule'
runs-on: ubuntu-22.04
outputs:
status: ${{ steps.detect.outputs.status }}
new-paths: ${{ steps.detect.outputs.new_paths }}
coverage: ${{ steps.detect.outputs.coverage }}
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Restore
run: dotnet restore src/StellaOps.sln
- name: Run Tests with Coverage
run: |
dotnet test src/StellaOps.sln \
--configuration Release \
--no-restore \
--verbosity minimal \
--collect:"XPlat Code Coverage" \
--results-directory ./coverage \
|| true # Don't fail on test failures
- name: Analyze Coverage
id: detect
run: |
COVERAGE_FILE=$(find ./coverage -name "coverage.cobertura.xml" | head -1)
if [ -z "$COVERAGE_FILE" ]; then
echo "status=skipped" >> $GITHUB_OUTPUT
echo "new_paths=0" >> $GITHUB_OUTPUT
echo "coverage=0" >> $GITHUB_OUTPUT
exit 0
fi
# Extract branch coverage
BRANCH_RATE=$(grep -oP 'branch-rate="\K[^"]+' "$COVERAGE_FILE" | head -1 || echo "0")
COVERAGE=$(echo "scale=2; $BRANCH_RATE * 100" | bc || echo "0")
echo "status=completed" >> $GITHUB_OUTPUT
echo "new_paths=0" >> $GITHUB_OUTPUT
echo "coverage=$COVERAGE" >> $GITHUB_OUTPUT
echo "Branch coverage: ${COVERAGE}%"
# ===========================================================================
# SCHEMA EVOLUTION CHECK
# ===========================================================================
schema-evolution:
name: Schema Evolution Check
needs: detect-changes
if: needs.detect-changes.outputs.has-schema-changes == 'true' || inputs.run_all == true
runs-on: ubuntu-22.04
services:
postgres:
image: postgres:16-alpine
env:
POSTGRES_USER: test
POSTGRES_PASSWORD: test
POSTGRES_DB: schema_test
ports:
- 5432:5432
options: >-
--health-cmd pg_isready
--health-interval 10s
--health-timeout 5s
--health-retries 5
outputs:
status: ${{ steps.test.outputs.status }}
compatible-versions: ${{ steps.test.outputs.compatible }}
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Restore
run: dotnet restore src/StellaOps.sln
- name: Run Schema Evolution Tests
id: test
env:
STELLAOPS_TEST_POSTGRES_CONNECTION: "Host=localhost;Port=5432;Database=schema_test;Username=test;Password=test"
run: |
echo "Running schema evolution tests..."
# Run tests with SchemaEvolution category
dotnet test src/StellaOps.sln \
--configuration Release \
--no-restore \
--filter "Category=SchemaEvolution" \
--verbosity normal \
|| RESULT=$?
if [ "${RESULT:-0}" -eq 0 ]; then
echo "status=passed" >> $GITHUB_OUTPUT
echo "compatible=N-1,N-2" >> $GITHUB_OUTPUT
else
echo "status=failed" >> $GITHUB_OUTPUT
echo "compatible=current-only" >> $GITHUB_OUTPUT
fi
# ===========================================================================
# CONFIG-DIFF CHECK
# ===========================================================================
config-diff:
name: Config-Diff Check
needs: detect-changes
if: needs.detect-changes.outputs.has-config-changes == 'true' || inputs.run_all == true
runs-on: ubuntu-22.04
outputs:
status: ${{ steps.test.outputs.status }}
tested-configs: ${{ steps.test.outputs.tested }}
steps:
- name: Checkout
uses: actions/checkout@v4
- name: Setup .NET
uses: actions/setup-dotnet@v4
with:
dotnet-version: ${{ env.DOTNET_VERSION }}
- name: Restore
run: dotnet restore src/StellaOps.sln
- name: Run Config-Diff Tests
id: test
run: |
echo "Running config-diff tests..."
# Run tests with ConfigDiff category
dotnet test src/StellaOps.sln \
--configuration Release \
--no-restore \
--filter "Category=ConfigDiff" \
--verbosity normal \
|| RESULT=$?
if [ "${RESULT:-0}" -eq 0 ]; then
echo "status=passed" >> $GITHUB_OUTPUT
else
echo "status=failed" >> $GITHUB_OUTPUT
fi
echo "tested=Concelier,Authority,Scanner" >> $GITHUB_OUTPUT
# ===========================================================================
# AGGREGATE REPORT
# ===========================================================================
report:
name: Generate Report
needs: [detect-changes, blast-radius, dead-paths, schema-evolution, config-diff]
if: always()
runs-on: ubuntu-22.04
steps:
- name: Generate Infrastructure Report
run: |
cat > test-infrastructure-report.md << 'EOF'
## Test Infrastructure Report
### Change Detection
| Category | Changed |
|----------|---------|
| Tests | ${{ needs.detect-changes.outputs.has-test-changes }} |
| Schema | ${{ needs.detect-changes.outputs.has-schema-changes }} |
| Code | ${{ needs.detect-changes.outputs.has-code-changes }} |
| Config | ${{ needs.detect-changes.outputs.has-config-changes }} |
### Validation Results
| Check | Status | Details |
|-------|--------|---------|
EOF
# Blast-radius
BR_STATUS="${{ needs.blast-radius.outputs.status || 'skipped' }}"
BR_VIOLATIONS="${{ needs.blast-radius.outputs.violations || '0' }}"
if [ "$BR_STATUS" = "passed" ]; then
echo "| Blast-Radius | :white_check_mark: | $BR_VIOLATIONS violations |" >> test-infrastructure-report.md
elif [ "$BR_STATUS" = "skipped" ]; then
echo "| Blast-Radius | :grey_question: | Skipped |" >> test-infrastructure-report.md
else
echo "| Blast-Radius | :x: | $BR_VIOLATIONS violations |" >> test-infrastructure-report.md
fi
# Dead-paths
DP_STATUS="${{ needs.dead-paths.outputs.status || 'skipped' }}"
DP_COVERAGE="${{ needs.dead-paths.outputs.coverage || 'N/A' }}"
if [ "$DP_STATUS" = "completed" ]; then
echo "| Dead-Path Detection | :white_check_mark: | Coverage: ${DP_COVERAGE}% |" >> test-infrastructure-report.md
elif [ "$DP_STATUS" = "skipped" ]; then
echo "| Dead-Path Detection | :grey_question: | Skipped |" >> test-infrastructure-report.md
else
echo "| Dead-Path Detection | :x: | Coverage: ${DP_COVERAGE}% |" >> test-infrastructure-report.md
fi
# Schema evolution
SE_STATUS="${{ needs.schema-evolution.outputs.status || 'skipped' }}"
SE_COMPAT="${{ needs.schema-evolution.outputs.compatible-versions || 'N/A' }}"
if [ "$SE_STATUS" = "passed" ]; then
echo "| Schema Evolution | :white_check_mark: | Compatible: $SE_COMPAT |" >> test-infrastructure-report.md
elif [ "$SE_STATUS" = "skipped" ]; then
echo "| Schema Evolution | :grey_question: | Skipped |" >> test-infrastructure-report.md
else
echo "| Schema Evolution | :x: | Compatible: $SE_COMPAT |" >> test-infrastructure-report.md
fi
# Config-diff
CD_STATUS="${{ needs.config-diff.outputs.status || 'skipped' }}"
CD_TESTED="${{ needs.config-diff.outputs.tested-configs || 'N/A' }}"
if [ "$CD_STATUS" = "passed" ]; then
echo "| Config-Diff | :white_check_mark: | Tested: $CD_TESTED |" >> test-infrastructure-report.md
elif [ "$CD_STATUS" = "skipped" ]; then
echo "| Config-Diff | :grey_question: | Skipped |" >> test-infrastructure-report.md
else
echo "| Config-Diff | :x: | Tested: $CD_TESTED |" >> test-infrastructure-report.md
fi
echo "" >> test-infrastructure-report.md
echo "---" >> test-infrastructure-report.md
echo "*Report generated at $(date -u +%Y-%m-%dT%H:%M:%SZ)*" >> test-infrastructure-report.md
cat test-infrastructure-report.md
cat test-infrastructure-report.md >> $GITHUB_STEP_SUMMARY
- name: Upload Report
uses: actions/upload-artifact@v4
with:
name: test-infrastructure-report
path: test-infrastructure-report.md
- name: Check for Failures
if: |
(needs.blast-radius.outputs.status == 'failed' ||
needs.dead-paths.outputs.status == 'failed' ||
needs.schema-evolution.outputs.status == 'failed' ||
needs.config-diff.outputs.status == 'failed') &&
inputs.fail_fast == true
run: |
echo "::error::One or more test infrastructure checks failed"
exit 1
# ===========================================================================
# POST PR COMMENT
# ===========================================================================
comment:
name: Post PR Comment
needs: [report, blast-radius, dead-paths, schema-evolution, config-diff]
if: github.event_name == 'pull_request' && always()
runs-on: ubuntu-22.04
permissions:
pull-requests: write
steps:
- name: Download Report
uses: actions/download-artifact@v4
with:
name: test-infrastructure-report
continue-on-error: true
- name: Post Comment
uses: actions/github-script@v7
with:
script: |
const fs = require('fs');
let report = '';
try {
report = fs.readFileSync('test-infrastructure-report.md', 'utf8');
} catch (e) {
report = 'Test infrastructure report not available.';
}
// Check for any failures
const brStatus = '${{ needs.blast-radius.outputs.status }}';
const dpStatus = '${{ needs.dead-paths.outputs.status }}';
const seStatus = '${{ needs.schema-evolution.outputs.status }}';
const cdStatus = '${{ needs.config-diff.outputs.status }}';
const hasFailed = [brStatus, dpStatus, seStatus, cdStatus].includes('failed');
const allPassed = [brStatus, dpStatus, seStatus, cdStatus]
.filter(s => s !== 'skipped' && s !== '')
.every(s => s === 'passed' || s === 'completed');
let status;
if (hasFailed) {
status = ':x: Some checks failed';
} else if (allPassed) {
status = ':white_check_mark: All checks passed';
} else {
status = ':grey_question: Some checks skipped';
}
const body = `## Test Infrastructure ${status}
${report}
---
<details>
<summary>About Test Infrastructure Checks</summary>
This workflow validates cross-cutting testing standards:
- **Blast-Radius**: Ensures Integration/Contract/Security tests have BlastRadius annotations
- **Dead-Path Detection**: Identifies uncovered code branches
- **Schema Evolution**: Validates backward compatibility with previous schema versions
- **Config-Diff**: Ensures config changes produce only expected behavioral deltas
</details>
`;
// Find and update or create comment
const { data: comments } = await github.rest.issues.listComments({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number
});
const botComment = comments.find(c =>
c.user.type === 'Bot' &&
c.body.includes('Test Infrastructure')
);
if (botComment) {
await github.rest.issues.updateComment({
owner: context.repo.owner,
repo: context.repo.repo,
comment_id: botComment.id,
body: body
});
} else {
await github.rest.issues.createComment({
owner: context.repo.owner,
repo: context.repo.repo,
issue_number: context.issue.number,
body: body
});
}

View File

@@ -246,6 +246,7 @@ Every sprint file must conform to this template:
* **If you find a sprint file whose internal structure deviates significantly from this template, you should normalise it toward this structure while preserving all existing content (log lines, tasks, decisions).**
* Record this normalisation in the **Execution Log** (e.g. “2025-11-16 · Normalised sprint file to standard template; no semantic changes.”).
* When sprint is fully completed move it to `docs-archived/implplan/`
Additional responsibilities (add-on):
@@ -368,7 +369,7 @@ If no design decision is required, you proceed autonomously, implementing the ch
1) **Doc sync (must happen for every advisory):**
- Create/update **two layers**:
- **High-level**: `docs/` (vision/key-features/market) to capture the moat/positioning and the headline promise.
- **Detailed**: closest deep area (`docs/reachability/*`, `docs/market/*`, `docs/benchmarks/*`, `docs/modules/<module>/*`, etc.).
- **Detailed**: closest deep area (`docs/modules/reach-graph/*`, `docs/modules/risk-engine/*`, `docs/benchmarks/*`, `docs/modules/<module>/*`, etc.).
- **Code & samples:**
- Inline only short fragments (≤ ~20 lines) directly in the updated doc for readability.
- Place runnable or longer samples/harnesses in `docs/benchmarks/**` or `tests/**` with deterministic, offline-friendly defaults (no network, fixed seeds), and link to them from the doc.
@@ -387,6 +388,7 @@ If no design decision is required, you proceed autonomously, implementing the ch
- Offline-friendly benches/tests; frozen feeds; deterministic ordering/hashes.
5) **Do not defer:** Execute steps 14 immediately; reporting is after the fact, not a gating step.
6) **Archive processed advisories**. After sprints / task / comprehensive documention is created or advisory is fully rejected move it to `docs-archived/product-advisories/`
**Lessons baked in:** Past delays came from missing code carry-over and missing sprint tasks. Always move advisory code into benchmarks/tests and open the corresponding sprint rows the same session you read the advisory.

View File

@@ -4,7 +4,7 @@ This file provides guidance to Claude Code (claude.ai/code) when working with co
## Project Overview
StellaOps is a self-hostable, sovereign container-security platform released under AGPL-3.0-or-later. It provides reproducible vulnerability scanning with VEX-first decisioning, SBOM generation (SPDX 3.0.1 and CycloneDX 1.7), in-toto/DSSE attestations, and optional Sigstore Rekor transparency. The platform is designed for offline/air-gapped operation with regional crypto support (eIDAS/FIPS/GOST/SM).
StellaOps is a self-hostable, sovereign container-security platform released under AGPL-3.0-or-later. It provides reproducible vulnerability scanning with VEX-first decisioning, SBOM generation (SPDX 2.2/2.3 and CycloneDX 1.7; SPDX 3.0.1 planned), in-toto/DSSE attestations, and optional Sigstore Rekor transparency. The platform is designed for offline/air-gapped operation with regional crypto support (eIDAS/FIPS/GOST/SM).
## Build Commands
@@ -82,6 +82,8 @@ The codebase follows a monorepo pattern with modules under `src/`:
| Authority | `src/Authority/` | Authentication, authorization, OAuth/OIDC, DPoP |
| Gateway | `src/Gateway/` | API gateway with routing and transport abstraction |
| Router | `src/Router/` | Transport-agnostic messaging (TCP/TLS/UDP/RabbitMQ/Valkey) |
| Platform | `src/Platform/` | Console backend aggregation service (health, quotas, search) |
| Registry | `src/Registry/` | Token service for container registry authentication |
| **Data Ingestion** | | |
| Concelier | `src/Concelier/` | Vulnerability advisory ingestion and merge engine |
| Excititor | `src/Excititor/` | VEX document ingestion and export |
@@ -89,13 +91,14 @@ The codebase follows a monorepo pattern with modules under `src/`:
| VexHub | `src/VexHub/` | VEX distribution and exchange hub |
| IssuerDirectory | `src/IssuerDirectory/` | Issuer trust registry (CSAF publishers) |
| Feedser | `src/Feedser/` | Evidence collection library for backport detection |
| Mirror | `src/Mirror/` | Vulnerability feed mirror and distribution |
| Mirror | `src/Concelier/__Libraries/` | Vulnerability feed mirror connector (Concelier plugin) |
| **Scanning & Analysis** | | |
| Scanner | `src/Scanner/` | Container scanning with SBOM generation (11 language analyzers) |
| BinaryIndex | `src/BinaryIndex/` | Binary identity extraction and fingerprinting |
| AdvisoryAI | `src/AdvisoryAI/` | AI-assisted advisory analysis |
| ReachGraph | `src/ReachGraph/` | Reachability graph service |
| Symbols | `src/Symbols/` | Symbol resolution and debug information |
| Cartographer | `src/Cartographer/` | Dependency graph mapping and visualization |
| **Artifacts & Evidence** | | |
| Attestor | `src/Attestor/` | in-toto/DSSE attestation generation |
| Signer | `src/Signer/` | Cryptographic signing operations |
@@ -108,6 +111,7 @@ The codebase follows a monorepo pattern with modules under `src/`:
| RiskEngine | `src/RiskEngine/` | Risk scoring runtime with pluggable providers |
| VulnExplorer | `src/VulnExplorer/` | Vulnerability exploration and triage UI backend |
| Unknowns | `src/Unknowns/` | Unknown component and symbol tracking |
| Findings | `src/Findings/` | Findings ledger service for vulnerability tracking |
| **Operations** | | |
| Scheduler | `src/Scheduler/` | Job scheduling and queue management |
| Orchestrator | `src/Orchestrator/` | Workflow orchestration and task coordination |
@@ -121,7 +125,7 @@ The codebase follows a monorepo pattern with modules under `src/`:
| CLI | `src/Cli/` | Command-line interface (Native AOT) |
| Zastava | `src/Zastava/` | Container registry webhook observer |
| Web | `src/Web/` | Angular 17 frontend SPA |
| API | `src/Api/` | OpenAPI contracts and governance |
| Integrations | `src/Integrations/` | External system integrations web service |
| **Infrastructure** | | |
| Cryptography | `src/Cryptography/` | Crypto plugins (FIPS, eIDAS, GOST, SM, PQ) |
| Telemetry | `src/Telemetry/` | OpenTelemetry traces, metrics, logging |
@@ -129,8 +133,12 @@ The codebase follows a monorepo pattern with modules under `src/`:
| Signals | `src/Signals/` | Runtime signal collection and correlation |
| AirGap | `src/AirGap/` | Air-gapped deployment support |
| AOC | `src/Aoc/` | Append-Only Contract enforcement (Roslyn analyzers) |
| SmRemote | `src/SmRemote/` | SM2/SM3/SM4 cryptographic remote service |
| **Development Tools** | | |
| Tools | `src/Tools/` | Development utilities (fixture updater, smoke tests, validators) |
| Bench | `src/Bench/` | Performance benchmark infrastructure |
> **Note:** See `docs/modules/<module>/architecture.md` for detailed module dossiers.
> **Note:** See `docs/modules/<module>/architecture.md` for detailed module dossiers. Some entries in `docs/modules/` are cross-cutting concepts (snapshot, triage) or shared libraries (provcache) rather than standalone modules.
### Code Organization Patterns
@@ -598,12 +606,77 @@ var createdAt = reader.GetDateTime(reader.GetOrdinal("created_at"));
var createdAt = reader.GetFieldValue<DateTimeOffset>(reader.GetOrdinal("created_at"));
```
### 8.19) Hybrid Logical Clock (HLC) Usage
| Rule | Guidance |
|------|----------|
| **Use IHybridLogicalClock for ordering** | For distributed ordering and audit-safe sequencing, use `IHybridLogicalClock` from `StellaOps.HybridLogicalClock`. Never rely on wall-clock time alone for ordering in distributed scenarios. |
```csharp
// BAD - wall-clock ordering in distributed system
public async Task EnqueueAsync(Job job)
{
job.EnqueuedAt = DateTimeOffset.UtcNow; // Clock skew risk!
await _store.SaveAsync(job);
}
// GOOD - HLC ordering
public async Task EnqueueAsync(Job job, CancellationToken ct)
{
job.THlc = _hlc.Tick(); // Monotonic, skew-tolerant
job.EnqueuedAtWall = _timeProvider.GetUtcNow(); // Informational only
await _store.SaveAsync(job, ct);
}
```
| Rule | Guidance |
|------|----------|
| **Deterministic event IDs** | Generate event IDs deterministically from content, not randomly. Use `SHA-256(correlationId \|\| tHlc \|\| service \|\| kind)` for timeline events. This ensures replay produces identical IDs. |
```csharp
// BAD - random ID breaks replay determinism
var eventId = Guid.NewGuid().ToString();
// GOOD - deterministic ID from content
var eventId = EventIdGenerator.Generate(correlationId, tHlc, service, kind);
// Returns: SHA-256(inputs)[0:32] as hex
```
| Rule | Guidance |
|------|----------|
| **HLC state persistence** | Persist HLC state on graceful shutdown via `IHlcStateStore`. On startup, call `InitializeFromStateAsync()` to restore monotonicity. This prevents HLC regression after restarts. |
```csharp
// Service startup
public async Task StartAsync(CancellationToken ct)
{
await _hlc.InitializeFromStateAsync(ct);
// HLC will now be >= last persisted value
}
// Service shutdown
public async Task StopAsync(CancellationToken ct)
{
await _hlc.PersistStateAsync(ct);
}
```
| Rule | Guidance |
|------|----------|
| **HLC in event envelopes** | Timeline events must include both `tHlc` (ordering) and `tsWall` (debugging). Use `HlcTimestamp.ToSortableString()` for string representation. Never parse HLC from user input without validation. |
| Rule | Guidance |
|------|----------|
| **Clock skew handling** | Configure reasonable `MaxClockSkew` tolerance (default: 5 seconds). Events with excessive skew throw `HlcClockSkewException`. Monitor `hlc_clock_skew_rejections_total` metric. |
**Reference:** See `docs/modules/eventing/event-envelope-schema.md` for the canonical event envelope specification.
### Documentation Updates
When scope, contracts, or workflows change, update the relevant docs under:
- `docs/modules/**` - Module architecture dossiers
- `docs/api/` - API documentation
- `docs/risk/` - Risk documentation
- `docs/modules/risk-engine/` - Risk documentation
- `docs/airgap/` - Air-gap operation docs
## Role-Based Behavior
@@ -634,11 +707,12 @@ Create implementation sprint files under `docs/implplan/` using the **mandatory*
- **If any existing sprint file name or internal format deviates from the standard, rename/normalize it** and record the change in its **Execution Log**.
- Normalize sprint files to standard template while preserving content
- Ensure module `AGENTS.md` files exist and are up to date
- When sprint is fully completed move it to `docs-archived/implplan/`
### As Product Manager
- Review advisories in `docs/product-advisories/`
- Check for overlaps with `docs/product-advisories/archived/`
- Check for overlaps with `docs-archived/product-advisories/`
- Validate against module docs and existing implementations
- Hand over to project manager role for sprint/task definition
@@ -656,7 +730,7 @@ Always update task status in `docs/implplan/SPRINT_*.md`:
Before coding, confirm required docs are read:
- `docs/README.md`
- `docs/07_HIGH_LEVEL_ARCHITECTURE.md`
- `docs/ARCHITECTURE_REFERENCE.md`
- `docs/modules/platform/architecture-overview.md`
- Relevant module dossier (e.g., `docs/modules/<module>/architecture.md`)
- Module-specific `AGENTS.md` file
@@ -674,13 +748,14 @@ Before coding, confirm required docs are read:
## Documentation
- **Architecture overview:** `docs/07_HIGH_LEVEL_ARCHITECTURE.md`
- **Architecture overview:** `docs/ARCHITECTURE_OVERVIEW.md`
- **Architecture reference:** `docs/ARCHITECTURE_REFERENCE.md`
- **Module dossiers:** `docs/modules/<module>/architecture.md`
- **Database specification:** `docs/db/SPECIFICATION.md`
- **PostgreSQL operations:** `docs/operations/postgresql-guide.md`
- **API/CLI reference:** `docs/09_API_CLI_REFERENCE.md`
- **Offline operation:** `docs/24_OFFLINE_KIT.md`
- **Quickstart:** `docs/10_CONCELIER_CLI_QUICKSTART.md`
- **API/CLI reference:** `docs/API_CLI_REFERENCE.md`
- **Offline operation:** `docs/OFFLINE_KIT.md`
- **Quickstart:** `docs/CONCELIER_CLI_QUICKSTART.md`
- **Sprint planning:** `docs/implplan/SPRINT_*.md`
## CI/CD

View File

@@ -1,4 +0,0 @@
/nowarn:CA2022
/p:DisableWorkloadResolver=true
/p:RestoreAdditionalProjectFallbackFolders=
/p:RestoreFallbackFolders=

View File

@@ -9,6 +9,14 @@
<add key="nuget.org" value="https://api.nuget.org/v3/index.json" />
<add key="stellaops" value="https://git.stella-ops.org/api/packages/stella-ops.org/nuget/index.json" />
</packageSources>
<packageSourceMapping>
<packageSource key="nuget.org">
<package pattern="*" />
</packageSource>
<packageSource key="stellaops">
<package pattern="StellaOps.*" />
</packageSource>
</packageSourceMapping>
<fallbackPackageFolders>
<clear />
</fallbackPackageFolders>

71
coverage-exemptions.yaml Normal file
View File

@@ -0,0 +1,71 @@
# coverage-exemptions.yaml
# Dead-path exemptions for intentionally untested code branches
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-016
#
# USAGE:
# ======
# Add file:line entries for code paths that are intentionally not covered.
# Each exemption MUST include a justification explaining why testing is not required.
#
# CATEGORIES:
# ===========
# - emergency: Emergency/fallback handlers that are tested manually
# - platform: Platform-specific code paths (e.g., Windows-only on Linux CI)
# - external: External system error handlers (e.g., network timeouts)
# - deprecated: Deprecated code paths scheduled for removal
# - defensive: Defensive programming that should never execute
#
# REVIEW:
# =======
# Exemptions should be reviewed quarterly. Remove exemptions for:
# - Code that has been deleted
# - Code that now has test coverage
# - Deprecated code that has been removed
version: "1.0"
# Global settings
settings:
# Require justification for all exemptions
require_justification: true
# Maximum age of exemptions before review required (days)
max_exemption_age_days: 90
# Fail CI if exemption is older than max age
fail_on_stale_exemptions: false
# Exemption entries
exemptions: []
# Example exemptions (commented out):
#
# - path: "src/Authority/Services/EmergencyAccessHandler.cs:42"
# category: emergency
# justification: "Emergency access bypass - tested manually during incident drills"
# added: "2026-01-06"
# owner: "security-team"
#
# - path: "src/Scanner/Platform/WindowsRegistryScanner.cs:128"
# category: platform
# justification: "Windows-only code path - CI runs on Linux"
# added: "2026-01-06"
# owner: "scanner-team"
#
# - path: "src/Concelier/Connectors/LegacyNvdConnector.cs:*"
# category: deprecated
# justification: "Entire file deprecated - scheduled for removal in 2026.Q2"
# added: "2026-01-06"
# owner: "concelier-team"
# removal_target: "2026-04-01"
# Patterns to ignore entirely (not counted as dead paths)
ignore_patterns:
# Generated code
- "*.Generated.cs"
- "*.Designer.cs"
# Migration files
- "**/Migrations/*.cs"
# Test infrastructure
- "**/*.Tests/**"
- "**/TestKit/**"
# Benchmark code
- "**/__Benchmarks/**"

9
dead-paths-baseline.json Normal file
View File

@@ -0,0 +1,9 @@
{
"version": "1.0.0",
"generatedAt": "2026-01-06T00:00:00Z",
"activeDeadPaths": 0,
"totalDeadPaths": 0,
"exemptedPaths": 0,
"description": "Initial baseline for dead-path detection. As tests are added and coverage improves, this baseline should decrease over time.",
"entries": []
}

View File

@@ -0,0 +1,42 @@
# Copyright (c) StellaOps. All rights reserved.
# Licensed under AGPL-3.0-or-later.
# Function Behavior Corpus PostgreSQL Database
#
# Usage:
# docker compose -f docker-compose.corpus.yml up -d
#
# Environment variables:
# CORPUS_DB_PASSWORD - PostgreSQL password for corpus database
services:
corpus-postgres:
image: postgres:16-alpine
container_name: stellaops-corpus-db
environment:
POSTGRES_DB: stellaops_corpus
POSTGRES_USER: corpus_user
POSTGRES_PASSWORD: ${CORPUS_DB_PASSWORD:-stellaops_corpus_dev}
POSTGRES_INITDB_ARGS: "-E UTF8 --locale=C"
volumes:
- corpus-data:/var/lib/postgresql/data
- ../../../docs/db/schemas/corpus.sql:/docker-entrypoint-initdb.d/10-corpus-schema.sql:ro
- ./scripts/init-test-data.sql:/docker-entrypoint-initdb.d/20-test-data.sql:ro
ports:
- "5435:5432"
networks:
- stellaops-corpus
healthcheck:
test: ["CMD-SHELL", "pg_isready -U corpus_user -d stellaops_corpus"]
interval: 10s
timeout: 5s
retries: 5
restart: unless-stopped
volumes:
corpus-data:
driver: local
networks:
stellaops-corpus:
driver: bridge

View File

@@ -0,0 +1,220 @@
-- =============================================================================
-- CORPUS TEST DATA - Minimal corpus for integration testing
-- Copyright (c) StellaOps. All rights reserved.
-- Licensed under AGPL-3.0-or-later.
-- =============================================================================
-- Set tenant for test data
SET app.tenant_id = 'test-tenant';
-- =============================================================================
-- LIBRARIES
-- =============================================================================
INSERT INTO corpus.libraries (id, name, description, homepage_url, source_repo)
VALUES
('a0000001-0000-0000-0000-000000000001', 'glibc', 'GNU C Library', 'https://www.gnu.org/software/libc/', 'https://sourceware.org/git/glibc.git'),
('a0000001-0000-0000-0000-000000000002', 'openssl', 'OpenSSL cryptographic library', 'https://www.openssl.org/', 'https://github.com/openssl/openssl.git'),
('a0000001-0000-0000-0000-000000000003', 'zlib', 'zlib compression library', 'https://zlib.net/', 'https://github.com/madler/zlib.git'),
('a0000001-0000-0000-0000-000000000004', 'curl', 'libcurl transfer library', 'https://curl.se/', 'https://github.com/curl/curl.git'),
('a0000001-0000-0000-0000-000000000005', 'sqlite', 'SQLite database engine', 'https://sqlite.org/', 'https://sqlite.org/src')
ON CONFLICT (tenant_id, name) DO NOTHING;
-- =============================================================================
-- LIBRARY VERSIONS (glibc)
-- =============================================================================
INSERT INTO corpus.library_versions (id, library_id, version, release_date, is_security_release)
VALUES
-- glibc versions
('b0000001-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000001', '2.17', '2012-12-25', false),
('b0000001-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000001', '2.28', '2018-08-01', false),
('b0000001-0000-0000-0000-000000000003', 'a0000001-0000-0000-0000-000000000001', '2.31', '2020-02-01', false),
('b0000001-0000-0000-0000-000000000004', 'a0000001-0000-0000-0000-000000000001', '2.35', '2022-02-03', false),
('b0000001-0000-0000-0000-000000000005', 'a0000001-0000-0000-0000-000000000001', '2.38', '2023-07-31', false),
-- OpenSSL versions
('b0000002-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000002', '1.0.2u', '2019-12-20', true),
('b0000002-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000002', '1.1.1w', '2023-09-11', true),
('b0000002-0000-0000-0000-000000000003', 'a0000001-0000-0000-0000-000000000002', '3.0.12', '2023-10-24', true),
('b0000002-0000-0000-0000-000000000004', 'a0000001-0000-0000-0000-000000000002', '3.1.4', '2023-10-24', true),
-- zlib versions
('b0000003-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000003', '1.2.11', '2017-01-15', false),
('b0000003-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000003', '1.2.13', '2022-10-13', true),
('b0000003-0000-0000-0000-000000000003', 'a0000001-0000-0000-0000-000000000003', '1.3.1', '2024-01-22', false)
ON CONFLICT (tenant_id, library_id, version) DO NOTHING;
-- =============================================================================
-- BUILD VARIANTS
-- =============================================================================
INSERT INTO corpus.build_variants (id, library_version_id, architecture, abi, compiler, compiler_version, optimization_level, binary_sha256)
VALUES
-- glibc 2.31 variants
('c0000001-0000-0000-0000-000000000001', 'b0000001-0000-0000-0000-000000000003', 'x86_64', 'gnu', 'gcc', '9.3.0', 'O2', 'a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2'),
('c0000001-0000-0000-0000-000000000002', 'b0000001-0000-0000-0000-000000000003', 'aarch64', 'gnu', 'gcc', '9.3.0', 'O2', 'b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3'),
('c0000001-0000-0000-0000-000000000003', 'b0000001-0000-0000-0000-000000000003', 'armhf', 'gnu', 'gcc', '9.3.0', 'O2', 'c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4'),
-- glibc 2.35 variants
('c0000002-0000-0000-0000-000000000001', 'b0000001-0000-0000-0000-000000000004', 'x86_64', 'gnu', 'gcc', '11.2.0', 'O2', 'd4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5'),
('c0000002-0000-0000-0000-000000000002', 'b0000001-0000-0000-0000-000000000004', 'aarch64', 'gnu', 'gcc', '11.2.0', 'O2', 'e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6'),
-- OpenSSL 3.0.12 variants
('c0000003-0000-0000-0000-000000000001', 'b0000002-0000-0000-0000-000000000003', 'x86_64', 'gnu', 'gcc', '11.2.0', 'O2', 'f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1'),
('c0000003-0000-0000-0000-000000000002', 'b0000002-0000-0000-0000-000000000003', 'aarch64', 'gnu', 'gcc', '11.2.0', 'O2', 'a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b3')
ON CONFLICT (tenant_id, library_version_id, architecture, abi, compiler, optimization_level) DO NOTHING;
-- =============================================================================
-- FUNCTIONS (Sample functions from glibc)
-- =============================================================================
INSERT INTO corpus.functions (id, build_variant_id, name, demangled_name, address, size_bytes, is_exported)
VALUES
-- glibc 2.31 x86_64 functions
('d0000001-0000-0000-0000-000000000001', 'c0000001-0000-0000-0000-000000000001', 'memcpy', 'memcpy', 140000, 256, true),
('d0000001-0000-0000-0000-000000000002', 'c0000001-0000-0000-0000-000000000001', 'memset', 'memset', 140256, 192, true),
('d0000001-0000-0000-0000-000000000003', 'c0000001-0000-0000-0000-000000000001', 'strlen', 'strlen', 140448, 128, true),
('d0000001-0000-0000-0000-000000000004', 'c0000001-0000-0000-0000-000000000001', 'strcmp', 'strcmp', 140576, 160, true),
('d0000001-0000-0000-0000-000000000005', 'c0000001-0000-0000-0000-000000000001', 'strcpy', 'strcpy', 140736, 144, true),
('d0000001-0000-0000-0000-000000000006', 'c0000001-0000-0000-0000-000000000001', 'malloc', 'malloc', 150000, 512, true),
('d0000001-0000-0000-0000-000000000007', 'c0000001-0000-0000-0000-000000000001', 'free', 'free', 150512, 384, true),
('d0000001-0000-0000-0000-000000000008', 'c0000001-0000-0000-0000-000000000001', 'realloc', 'realloc', 150896, 448, true),
('d0000001-0000-0000-0000-000000000009', 'c0000001-0000-0000-0000-000000000001', 'printf', 'printf', 160000, 1024, true),
('d0000001-0000-0000-0000-000000000010', 'c0000001-0000-0000-0000-000000000001', 'sprintf', 'sprintf', 161024, 896, true),
-- glibc 2.35 x86_64 functions (same functions, different addresses/sizes due to optimization)
('d0000002-0000-0000-0000-000000000001', 'c0000002-0000-0000-0000-000000000001', 'memcpy', 'memcpy', 145000, 280, true),
('d0000002-0000-0000-0000-000000000002', 'c0000002-0000-0000-0000-000000000001', 'memset', 'memset', 145280, 208, true),
('d0000002-0000-0000-0000-000000000003', 'c0000002-0000-0000-0000-000000000001', 'strlen', 'strlen', 145488, 144, true),
('d0000002-0000-0000-0000-000000000004', 'c0000002-0000-0000-0000-000000000001', 'strcmp', 'strcmp', 145632, 176, true),
('d0000002-0000-0000-0000-000000000005', 'c0000002-0000-0000-0000-000000000001', 'strcpy', 'strcpy', 145808, 160, true),
('d0000002-0000-0000-0000-000000000006', 'c0000002-0000-0000-0000-000000000001', 'malloc', 'malloc', 155000, 544, true),
('d0000002-0000-0000-0000-000000000007', 'c0000002-0000-0000-0000-000000000001', 'free', 'free', 155544, 400, true),
-- OpenSSL 3.0.12 functions
('d0000003-0000-0000-0000-000000000001', 'c0000003-0000-0000-0000-000000000001', 'EVP_DigestInit_ex', 'EVP_DigestInit_ex', 200000, 320, true),
('d0000003-0000-0000-0000-000000000002', 'c0000003-0000-0000-0000-000000000001', 'EVP_DigestUpdate', 'EVP_DigestUpdate', 200320, 256, true),
('d0000003-0000-0000-0000-000000000003', 'c0000003-0000-0000-0000-000000000001', 'EVP_DigestFinal_ex', 'EVP_DigestFinal_ex', 200576, 288, true),
('d0000003-0000-0000-0000-000000000004', 'c0000003-0000-0000-0000-000000000001', 'EVP_EncryptInit_ex', 'EVP_EncryptInit_ex', 201000, 384, true),
('d0000003-0000-0000-0000-000000000005', 'c0000003-0000-0000-0000-000000000001', 'EVP_DecryptInit_ex', 'EVP_DecryptInit_ex', 201384, 384, true),
('d0000003-0000-0000-0000-000000000006', 'c0000003-0000-0000-0000-000000000001', 'SSL_CTX_new', 'SSL_CTX_new', 300000, 512, true),
('d0000003-0000-0000-0000-000000000007', 'c0000003-0000-0000-0000-000000000001', 'SSL_new', 'SSL_new', 300512, 384, true),
('d0000003-0000-0000-0000-000000000008', 'c0000003-0000-0000-0000-000000000001', 'SSL_connect', 'SSL_connect', 300896, 1024, true)
ON CONFLICT (tenant_id, build_variant_id, name, address) DO NOTHING;
-- =============================================================================
-- FINGERPRINTS (Simulated semantic fingerprints)
-- =============================================================================
INSERT INTO corpus.fingerprints (id, function_id, algorithm, fingerprint, metadata)
VALUES
-- memcpy fingerprints (semantic_ksg algorithm)
('e0000001-0000-0000-0000-000000000001', 'd0000001-0000-0000-0000-000000000001', 'semantic_ksg',
decode('a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f60001', 'hex'),
'{"node_count": 45, "edge_count": 72, "api_calls": ["memcpy_internal"], "complexity": 8}'::jsonb),
('e0000001-0000-0000-0000-000000000002', 'd0000001-0000-0000-0000-000000000001', 'instruction_bb',
decode('b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a10001', 'hex'),
'{"bb_count": 8, "instruction_count": 64}'::jsonb),
-- memcpy 2.35 (similar fingerprint, different version)
('e0000002-0000-0000-0000-000000000001', 'd0000002-0000-0000-0000-000000000001', 'semantic_ksg',
decode('a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f60002', 'hex'),
'{"node_count": 48, "edge_count": 76, "api_calls": ["memcpy_internal"], "complexity": 9}'::jsonb),
-- memset fingerprints
('e0000003-0000-0000-0000-000000000001', 'd0000001-0000-0000-0000-000000000002', 'semantic_ksg',
decode('c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b20001', 'hex'),
'{"node_count": 32, "edge_count": 48, "api_calls": [], "complexity": 5}'::jsonb),
-- strlen fingerprints
('e0000004-0000-0000-0000-000000000001', 'd0000001-0000-0000-0000-000000000003', 'semantic_ksg',
decode('d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c30001', 'hex'),
'{"node_count": 24, "edge_count": 32, "api_calls": [], "complexity": 4}'::jsonb),
-- malloc fingerprints
('e0000005-0000-0000-0000-000000000001', 'd0000001-0000-0000-0000-000000000006', 'semantic_ksg',
decode('e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d40001', 'hex'),
'{"node_count": 128, "edge_count": 256, "api_calls": ["sbrk", "mmap"], "complexity": 24}'::jsonb),
-- OpenSSL EVP_DigestInit_ex
('e0000006-0000-0000-0000-000000000001', 'd0000003-0000-0000-0000-000000000001', 'semantic_ksg',
decode('f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e50001', 'hex'),
'{"node_count": 56, "edge_count": 84, "api_calls": ["OPENSSL_init_crypto"], "complexity": 12}'::jsonb),
-- SSL_CTX_new
('e0000007-0000-0000-0000-000000000001', 'd0000003-0000-0000-0000-000000000006', 'semantic_ksg',
decode('a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f6a1b2c3d4e5f60003', 'hex'),
'{"node_count": 96, "edge_count": 144, "api_calls": ["CRYPTO_malloc", "SSL_CTX_set_options"], "complexity": 18}'::jsonb)
ON CONFLICT (tenant_id, function_id, algorithm) DO NOTHING;
-- =============================================================================
-- FUNCTION CLUSTERS
-- =============================================================================
INSERT INTO corpus.function_clusters (id, library_id, canonical_name, description)
VALUES
('f0000001-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000001', 'memcpy', 'Memory copy function across glibc versions'),
('f0000001-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000001', 'memset', 'Memory set function across glibc versions'),
('f0000001-0000-0000-0000-000000000003', 'a0000001-0000-0000-0000-000000000001', 'strlen', 'String length function across glibc versions'),
('f0000001-0000-0000-0000-000000000004', 'a0000001-0000-0000-0000-000000000001', 'malloc', 'Memory allocation function across glibc versions'),
('f0000002-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000002', 'EVP_DigestInit_ex', 'EVP digest initialization across OpenSSL versions'),
('f0000002-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000002', 'SSL_CTX_new', 'SSL context creation across OpenSSL versions')
ON CONFLICT (tenant_id, library_id, canonical_name) DO NOTHING;
-- =============================================================================
-- CLUSTER MEMBERS
-- =============================================================================
INSERT INTO corpus.cluster_members (cluster_id, function_id, similarity_to_centroid)
VALUES
-- memcpy cluster
('f0000001-0000-0000-0000-000000000001', 'd0000001-0000-0000-0000-000000000001', 1.0),
('f0000001-0000-0000-0000-000000000001', 'd0000002-0000-0000-0000-000000000001', 0.95),
-- memset cluster
('f0000001-0000-0000-0000-000000000002', 'd0000001-0000-0000-0000-000000000002', 1.0),
('f0000001-0000-0000-0000-000000000002', 'd0000002-0000-0000-0000-000000000002', 0.92),
-- strlen cluster
('f0000001-0000-0000-0000-000000000003', 'd0000001-0000-0000-0000-000000000003', 1.0),
('f0000001-0000-0000-0000-000000000003', 'd0000002-0000-0000-0000-000000000003', 0.94),
-- malloc cluster
('f0000001-0000-0000-0000-000000000004', 'd0000001-0000-0000-0000-000000000006', 1.0),
('f0000001-0000-0000-0000-000000000004', 'd0000002-0000-0000-0000-000000000006', 0.88)
ON CONFLICT DO NOTHING;
-- =============================================================================
-- CVE ASSOCIATIONS
-- =============================================================================
INSERT INTO corpus.function_cves (function_id, cve_id, affected_state, confidence, evidence_type)
VALUES
-- CVE-2021-3999 affects glibc getcwd
-- Note: We don't have getcwd in our test data, but this shows the structure
-- CVE-2022-0778 affects OpenSSL BN_mod_sqrt (infinite loop)
('d0000003-0000-0000-0000-000000000001', 'CVE-2022-0778', 'fixed', 0.95, 'advisory'),
('d0000003-0000-0000-0000-000000000002', 'CVE-2022-0778', 'fixed', 0.95, 'advisory'),
-- CVE-2023-0286 affects OpenSSL X509 certificate handling
('d0000003-0000-0000-0000-000000000006', 'CVE-2023-0286', 'fixed', 0.90, 'commit'),
('d0000003-0000-0000-0000-000000000007', 'CVE-2023-0286', 'fixed', 0.90, 'commit')
ON CONFLICT (tenant_id, function_id, cve_id) DO NOTHING;
-- =============================================================================
-- INGESTION LOG
-- =============================================================================
INSERT INTO corpus.ingestion_jobs (id, library_id, job_type, status, functions_indexed, started_at, completed_at)
VALUES
('99000001-0000-0000-0000-000000000001', 'a0000001-0000-0000-0000-000000000001', 'full_ingest', 'completed', 10, now() - interval '1 day', now() - interval '1 day' + interval '5 minutes'),
('99000001-0000-0000-0000-000000000002', 'a0000001-0000-0000-0000-000000000002', 'full_ingest', 'completed', 8, now() - interval '12 hours', now() - interval '12 hours' + interval '3 minutes')
ON CONFLICT DO NOTHING;
-- =============================================================================
-- SUMMARY
-- =============================================================================
DO $$
DECLARE
lib_count INT;
ver_count INT;
func_count INT;
fp_count INT;
BEGIN
SELECT COUNT(*) INTO lib_count FROM corpus.libraries;
SELECT COUNT(*) INTO ver_count FROM corpus.library_versions;
SELECT COUNT(*) INTO func_count FROM corpus.functions;
SELECT COUNT(*) INTO fp_count FROM corpus.fingerprints;
RAISE NOTICE 'Corpus test data initialized:';
RAISE NOTICE ' Libraries: %', lib_count;
RAISE NOTICE ' Versions: %', ver_count;
RAISE NOTICE ' Functions: %', func_count;
RAISE NOTICE ' Fingerprints: %', fp_count;
END $$;

View File

@@ -0,0 +1,84 @@
# Copyright (c) StellaOps. All rights reserved.
# Licensed under AGPL-3.0-or-later.
# Ghidra Headless Analysis Server for BinaryIndex
#
# This image provides Ghidra headless analysis capabilities including:
# - Ghidra Headless Analyzer (analyzeHeadless)
# - ghidriff for automated binary diffing
# - Version Tracking and BSim support
#
# Build:
# docker build -f Dockerfile.headless -t stellaops/ghidra-headless:11.2 .
#
# Run:
# docker run --rm -v /path/to/binaries:/binaries stellaops/ghidra-headless:11.2 \
# /projects GhidraProject -import /binaries/target.exe -analyze
FROM eclipse-temurin:17-jdk-jammy
ARG GHIDRA_VERSION=11.2
ARG GHIDRA_BUILD_DATE=20241105
ARG GHIDRA_SHA256
LABEL org.opencontainers.image.title="StellaOps Ghidra Headless"
LABEL org.opencontainers.image.description="Ghidra headless analysis server with ghidriff for BinaryIndex"
LABEL org.opencontainers.image.version="${GHIDRA_VERSION}"
LABEL org.opencontainers.image.licenses="AGPL-3.0-or-later"
LABEL org.opencontainers.image.source="https://github.com/stellaops/stellaops"
LABEL org.opencontainers.image.vendor="StellaOps"
# Install dependencies
RUN apt-get update && apt-get install -y --no-install-recommends \
python3 \
python3-pip \
python3-venv \
curl \
unzip \
&& apt-get clean \
&& rm -rf /var/lib/apt/lists/*
# Download and verify Ghidra
# Note: Set GHIDRA_SHA256 build arg for production builds
RUN curl -fsSL "https://github.com/NationalSecurityAgency/ghidra/releases/download/Ghidra_${GHIDRA_VERSION}_build/ghidra_${GHIDRA_VERSION}_PUBLIC_${GHIDRA_BUILD_DATE}.zip" \
-o /tmp/ghidra.zip \
&& if [ -n "${GHIDRA_SHA256}" ]; then \
echo "${GHIDRA_SHA256} /tmp/ghidra.zip" | sha256sum -c -; \
fi \
&& unzip -q /tmp/ghidra.zip -d /opt \
&& rm /tmp/ghidra.zip \
&& ln -s /opt/ghidra_${GHIDRA_VERSION}_PUBLIC /opt/ghidra \
&& chmod +x /opt/ghidra/support/analyzeHeadless
# Install ghidriff in isolated virtual environment
RUN python3 -m venv /opt/venv \
&& /opt/venv/bin/pip install --no-cache-dir --upgrade pip \
&& /opt/venv/bin/pip install --no-cache-dir ghidriff
# Set environment variables
ENV GHIDRA_HOME=/opt/ghidra
ENV GHIDRA_INSTALL_DIR=/opt/ghidra
ENV JAVA_HOME=/opt/java/openjdk
ENV PATH="${GHIDRA_HOME}/support:/opt/venv/bin:${PATH}"
ENV MAXMEM=4G
# Create working directories with proper permissions
RUN mkdir -p /projects /scripts /output \
&& chmod 755 /projects /scripts /output
# Create non-root user for security
RUN groupadd -r ghidra && useradd -r -g ghidra ghidra \
&& chown -R ghidra:ghidra /projects /scripts /output
WORKDIR /projects
# Healthcheck - verify Ghidra is functional
HEALTHCHECK --interval=30s --timeout=10s --start-period=5s --retries=3 \
CMD analyzeHeadless /tmp HealthCheck -help > /dev/null 2>&1 || exit 1
# Switch to non-root user
USER ghidra
# Default entrypoint is analyzeHeadless
ENTRYPOINT ["analyzeHeadless"]
CMD ["--help"]

View File

@@ -0,0 +1,77 @@
# Copyright (c) StellaOps. All rights reserved.
# Licensed under AGPL-3.0-or-later.
# BSim PostgreSQL Database and Ghidra Headless Services
#
# Usage:
# docker compose -f docker-compose.bsim.yml up -d
#
# Environment variables:
# BSIM_DB_PASSWORD - PostgreSQL password for BSim database
version: '3.8'
services:
bsim-postgres:
image: postgres:16-alpine
container_name: stellaops-bsim-db
environment:
POSTGRES_DB: bsim_corpus
POSTGRES_USER: bsim_user
POSTGRES_PASSWORD: ${BSIM_DB_PASSWORD:-stellaops_bsim_dev}
POSTGRES_INITDB_ARGS: "-E UTF8 --locale=C"
volumes:
- bsim-data:/var/lib/postgresql/data
- ./scripts/init-bsim.sql:/docker-entrypoint-initdb.d/10-init-bsim.sql:ro
ports:
- "5433:5432"
networks:
- stellaops-bsim
healthcheck:
test: ["CMD-SHELL", "pg_isready -U bsim_user -d bsim_corpus"]
interval: 10s
timeout: 5s
retries: 5
restart: unless-stopped
# Ghidra Headless service for BSim analysis
ghidra-headless:
build:
context: .
dockerfile: Dockerfile.headless
image: stellaops/ghidra-headless:11.2
container_name: stellaops-ghidra
depends_on:
bsim-postgres:
condition: service_healthy
environment:
BSIM_DB_URL: "postgresql://bsim-postgres:5432/bsim_corpus"
BSIM_DB_USER: bsim_user
BSIM_DB_PASSWORD: ${BSIM_DB_PASSWORD:-stellaops_bsim_dev}
JAVA_HOME: /opt/java/openjdk
MAXMEM: 4G
volumes:
- ghidra-projects:/projects
- ghidra-scripts:/scripts
- ghidra-output:/output
networks:
- stellaops-bsim
deploy:
resources:
limits:
cpus: '4'
memory: 8G
# Keep container running for ad-hoc analysis
entrypoint: ["tail", "-f", "/dev/null"]
restart: unless-stopped
volumes:
bsim-data:
driver: local
ghidra-projects:
ghidra-scripts:
ghidra-output:
networks:
stellaops-bsim:
driver: bridge

View File

@@ -0,0 +1,140 @@
-- BSim PostgreSQL Schema Initialization
-- Copyright (c) StellaOps. All rights reserved.
-- Licensed under AGPL-3.0-or-later.
--
-- This script creates the core BSim schema structure.
-- Note: Full Ghidra BSim schema is auto-created by Ghidra tools.
-- This provides a minimal functional schema for integration testing.
-- Create schema comment
COMMENT ON DATABASE bsim_corpus IS 'Ghidra BSim function signature database for StellaOps BinaryIndex';
-- Enable required extensions
CREATE EXTENSION IF NOT EXISTS "uuid-ossp";
CREATE EXTENSION IF NOT EXISTS "pg_trgm";
-- BSim executables table
CREATE TABLE IF NOT EXISTS bsim_executables (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
name TEXT NOT NULL,
architecture TEXT NOT NULL,
library_name TEXT,
library_version TEXT,
md5_hash BYTEA,
sha256_hash BYTEA,
date_added TIMESTAMPTZ NOT NULL DEFAULT now(),
UNIQUE (sha256_hash)
);
-- BSim functions table
CREATE TABLE IF NOT EXISTS bsim_functions (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
executable_id UUID NOT NULL REFERENCES bsim_executables(id) ON DELETE CASCADE,
name TEXT NOT NULL,
address BIGINT NOT NULL,
flags INTEGER DEFAULT 0,
UNIQUE (executable_id, address)
);
-- BSim function vectors (feature vectors for similarity)
CREATE TABLE IF NOT EXISTS bsim_vectors (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
function_id UUID NOT NULL REFERENCES bsim_functions(id) ON DELETE CASCADE,
lsh_hash BYTEA NOT NULL, -- Locality-sensitive hash
feature_count INTEGER NOT NULL,
vector_data BYTEA NOT NULL, -- Serialized feature vector
UNIQUE (function_id)
);
-- BSim function signatures (compact fingerprints)
CREATE TABLE IF NOT EXISTS bsim_signatures (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
function_id UUID NOT NULL REFERENCES bsim_functions(id) ON DELETE CASCADE,
signature_type TEXT NOT NULL, -- 'basic', 'weighted', 'full'
signature_hash BYTEA NOT NULL,
significance REAL NOT NULL DEFAULT 0.0,
created_at TIMESTAMPTZ NOT NULL DEFAULT now(),
UNIQUE (function_id, signature_type)
);
-- BSim clusters (similar function groups)
CREATE TABLE IF NOT EXISTS bsim_clusters (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
name TEXT,
function_count INTEGER NOT NULL DEFAULT 0,
centroid_vector BYTEA,
created_at TIMESTAMPTZ NOT NULL DEFAULT now()
);
-- Cluster membership
CREATE TABLE IF NOT EXISTS bsim_cluster_members (
cluster_id UUID NOT NULL REFERENCES bsim_clusters(id) ON DELETE CASCADE,
function_id UUID NOT NULL REFERENCES bsim_functions(id) ON DELETE CASCADE,
similarity REAL NOT NULL,
PRIMARY KEY (cluster_id, function_id)
);
-- Ingestion tracking
CREATE TABLE IF NOT EXISTS bsim_ingest_log (
id UUID PRIMARY KEY DEFAULT uuid_generate_v4(),
executable_id UUID REFERENCES bsim_executables(id),
library_name TEXT NOT NULL,
library_version TEXT,
functions_ingested INTEGER NOT NULL DEFAULT 0,
status TEXT NOT NULL DEFAULT 'pending',
error_message TEXT,
started_at TIMESTAMPTZ,
completed_at TIMESTAMPTZ,
ingested_at TIMESTAMPTZ NOT NULL DEFAULT now()
);
-- Indexes for efficient querying
CREATE INDEX IF NOT EXISTS idx_bsim_functions_executable ON bsim_functions(executable_id);
CREATE INDEX IF NOT EXISTS idx_bsim_functions_name ON bsim_functions(name);
CREATE INDEX IF NOT EXISTS idx_bsim_vectors_lsh ON bsim_vectors USING hash (lsh_hash);
CREATE INDEX IF NOT EXISTS idx_bsim_signatures_hash ON bsim_signatures USING hash (signature_hash);
CREATE INDEX IF NOT EXISTS idx_bsim_executables_library ON bsim_executables(library_name, library_version);
CREATE INDEX IF NOT EXISTS idx_bsim_ingest_log_status ON bsim_ingest_log(status);
-- Views for common queries
CREATE OR REPLACE VIEW bsim_function_summary AS
SELECT
f.id AS function_id,
f.name AS function_name,
f.address,
e.name AS executable_name,
e.library_name,
e.library_version,
e.architecture,
s.significance
FROM bsim_functions f
JOIN bsim_executables e ON f.executable_id = e.id
LEFT JOIN bsim_signatures s ON f.id = s.function_id AND s.signature_type = 'basic';
CREATE OR REPLACE VIEW bsim_library_stats AS
SELECT
e.library_name,
e.library_version,
COUNT(DISTINCT e.id) AS executable_count,
COUNT(DISTINCT f.id) AS function_count,
MAX(l.ingested_at) AS last_ingested
FROM bsim_executables e
LEFT JOIN bsim_functions f ON e.id = f.executable_id
LEFT JOIN bsim_ingest_log l ON e.id = l.executable_id
WHERE e.library_name IS NOT NULL
GROUP BY e.library_name, e.library_version
ORDER BY e.library_name, e.library_version;
-- Grant permissions
GRANT ALL ON ALL TABLES IN SCHEMA public TO bsim_user;
GRANT ALL ON ALL SEQUENCES IN SCHEMA public TO bsim_user;
-- Insert schema version marker
INSERT INTO bsim_ingest_log (library_name, functions_ingested, status, completed_at)
VALUES ('_schema_init', 0, 'completed', now());
-- Log successful initialization
DO $$
BEGIN
RAISE NOTICE 'BSim schema initialized successfully';
END $$;

View File

@@ -0,0 +1,49 @@
# devops/docker/schema-versions/Dockerfile
# Versioned PostgreSQL container for schema evolution testing
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-008
#
# USAGE:
# ======
# Build for specific module and version:
# docker build --build-arg MODULE=scanner --build-arg SCHEMA_VERSION=v1.2.0 \
# -t stellaops/schema-test:scanner-v1.2.0 .
#
# Run for testing:
# docker run -d -p 5432:5432 stellaops/schema-test:scanner-v1.2.0
ARG POSTGRES_VERSION=16
FROM postgres:${POSTGRES_VERSION}-alpine
# Build arguments
ARG MODULE=scanner
ARG SCHEMA_VERSION=latest
ARG SCHEMA_DATE=""
# Labels for identification
LABEL org.opencontainers.image.title="StellaOps Schema Test - ${MODULE}"
LABEL org.opencontainers.image.description="PostgreSQL with ${MODULE} schema version ${SCHEMA_VERSION}"
LABEL org.opencontainers.image.version="${SCHEMA_VERSION}"
LABEL org.stellaops.module="${MODULE}"
LABEL org.stellaops.schema.version="${SCHEMA_VERSION}"
LABEL org.stellaops.schema.date="${SCHEMA_DATE}"
# Environment variables
ENV POSTGRES_USER=stellaops_test
ENV POSTGRES_PASSWORD=test_password
ENV POSTGRES_DB=stellaops_schema_test
ENV STELLAOPS_MODULE=${MODULE}
ENV STELLAOPS_SCHEMA_VERSION=${SCHEMA_VERSION}
# Copy initialization scripts
COPY docker-entrypoint-initdb.d/ /docker-entrypoint-initdb.d/
# Copy module-specific schema
COPY schemas/${MODULE}/ /schemas/${MODULE}/
# Health check
HEALTHCHECK --interval=10s --timeout=5s --start-period=30s --retries=3 \
CMD pg_isready -U ${POSTGRES_USER} -d ${POSTGRES_DB} || exit 1
# Expose PostgreSQL port
EXPOSE 5432

View File

@@ -0,0 +1,179 @@
#!/bin/bash
# build-schema-images.sh
# Build versioned PostgreSQL images for schema evolution testing
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-008
#
# USAGE:
# ======
# Build all versions for a module:
# ./build-schema-images.sh scanner
#
# Build specific version:
# ./build-schema-images.sh scanner v1.2.0
#
# Build all modules:
# ./build-schema-images.sh --all
set -e
SCRIPT_DIR="$(cd "$(dirname "${BASH_SOURCE[0]}")" && pwd)"
REPO_ROOT="$(cd "$SCRIPT_DIR/../../.." && pwd)"
REGISTRY="${SCHEMA_REGISTRY:-ghcr.io/stellaops}"
POSTGRES_VERSION="${POSTGRES_VERSION:-16}"
# Modules with schema evolution support
MODULES=("scanner" "concelier" "evidencelocker" "authority" "sbomservice" "policy")
usage() {
echo "Usage: $0 <module|--all> [version]"
echo ""
echo "Arguments:"
echo " module Module name (scanner, concelier, evidencelocker, authority, sbomservice, policy)"
echo " --all Build all modules"
echo " version Optional specific version to build (default: all versions)"
echo ""
echo "Environment variables:"
echo " SCHEMA_REGISTRY Container registry (default: ghcr.io/stellaops)"
echo " POSTGRES_VERSION PostgreSQL version (default: 16)"
echo " PUSH_IMAGES Set to 'true' to push images after build"
exit 1
}
# Get schema versions from git tags or migration files
get_schema_versions() {
local module=$1
local versions=()
# Check for version tags
local tags=$(git tag -l "${module}-schema-v*" 2>/dev/null | sed "s/${module}-schema-//" | sort -V)
if [ -n "$tags" ]; then
versions=($tags)
else
# Fall back to migration file count
local migration_dir="$REPO_ROOT/docs/db/migrations/${module}"
if [ -d "$migration_dir" ]; then
local count=$(ls -1 "$migration_dir"/*.sql 2>/dev/null | wc -l)
for i in $(seq 1 $count); do
versions+=("v1.0.$i")
done
fi
fi
# Always include 'latest'
versions+=("latest")
echo "${versions[@]}"
}
# Copy schema files to build context
prepare_schema_context() {
local module=$1
local version=$2
local build_dir="$SCRIPT_DIR/.build/${module}/${version}"
mkdir -p "$build_dir/schemas/${module}"
mkdir -p "$build_dir/docker-entrypoint-initdb.d"
# Copy entrypoint scripts
cp "$SCRIPT_DIR/docker-entrypoint-initdb.d/"*.sh "$build_dir/docker-entrypoint-initdb.d/"
# Copy base schema
local base_schema="$REPO_ROOT/docs/db/schemas/${module}.sql"
if [ -f "$base_schema" ]; then
cp "$base_schema" "$build_dir/schemas/${module}/base.sql"
fi
# Copy migrations directory
local migrations_dir="$REPO_ROOT/docs/db/migrations/${module}"
if [ -d "$migrations_dir" ]; then
mkdir -p "$build_dir/schemas/${module}/migrations"
cp "$migrations_dir"/*.sql "$build_dir/schemas/${module}/migrations/" 2>/dev/null || true
fi
echo "$build_dir"
}
# Build image for module and version
build_image() {
local module=$1
local version=$2
echo "Building ${module} schema version ${version}..."
local build_dir=$(prepare_schema_context "$module" "$version")
local image_tag="${REGISTRY}/schema-test:${module}-${version}"
local schema_date=$(date -u +%Y-%m-%dT%H:%M:%SZ)
# Copy Dockerfile to build context
cp "$SCRIPT_DIR/Dockerfile" "$build_dir/"
# Build the image
docker build \
--build-arg MODULE="$module" \
--build-arg SCHEMA_VERSION="$version" \
--build-arg SCHEMA_DATE="$schema_date" \
--build-arg POSTGRES_VERSION="$POSTGRES_VERSION" \
-t "$image_tag" \
"$build_dir"
echo "Built: $image_tag"
# Push if requested
if [ "$PUSH_IMAGES" = "true" ]; then
echo "Pushing: $image_tag"
docker push "$image_tag"
fi
# Cleanup build directory
rm -rf "$build_dir"
}
# Build all versions for a module
build_module() {
local module=$1
local target_version=$2
echo "========================================"
echo "Building schema images for: $module"
echo "========================================"
if [ -n "$target_version" ]; then
build_image "$module" "$target_version"
else
local versions=$(get_schema_versions "$module")
for version in $versions; do
build_image "$module" "$version"
done
fi
}
# Main
if [ $# -lt 1 ]; then
usage
fi
case "$1" in
--all)
for module in "${MODULES[@]}"; do
build_module "$module" "$2"
done
;;
--help|-h)
usage
;;
*)
if [[ " ${MODULES[*]} " =~ " $1 " ]]; then
build_module "$1" "$2"
else
echo "Error: Unknown module '$1'"
echo "Valid modules: ${MODULES[*]}"
exit 1
fi
;;
esac
echo ""
echo "Build complete!"
echo "To push images, run with PUSH_IMAGES=true"

View File

@@ -0,0 +1,70 @@
#!/bin/bash
# 00-init-schema.sh
# Initialize PostgreSQL with module schema for testing
# Sprint: SPRINT_20260105_002_005_TEST_cross_cutting
# Task: CCUT-008
set -e
echo "Initializing schema for module: ${STELLAOPS_MODULE}"
echo "Schema version: ${STELLAOPS_SCHEMA_VERSION}"
# Create extensions
psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" <<-EOSQL
CREATE EXTENSION IF NOT EXISTS "uuid-ossp";
CREATE EXTENSION IF NOT EXISTS "pgcrypto";
CREATE EXTENSION IF NOT EXISTS "btree_gist";
EOSQL
# Apply base schema if exists
BASE_SCHEMA="/schemas/${STELLAOPS_MODULE}/base.sql"
if [ -f "$BASE_SCHEMA" ]; then
echo "Applying base schema: $BASE_SCHEMA"
psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" -f "$BASE_SCHEMA"
fi
# Apply versioned schema if exists
VERSION_SCHEMA="/schemas/${STELLAOPS_MODULE}/${STELLAOPS_SCHEMA_VERSION}.sql"
if [ -f "$VERSION_SCHEMA" ]; then
echo "Applying version schema: $VERSION_SCHEMA"
psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" -f "$VERSION_SCHEMA"
fi
# Apply all migrations up to version
MIGRATIONS_DIR="/schemas/${STELLAOPS_MODULE}/migrations"
if [ -d "$MIGRATIONS_DIR" ]; then
echo "Applying migrations from: $MIGRATIONS_DIR"
# Get version number for comparison
VERSION_NUM=$(echo "$STELLAOPS_SCHEMA_VERSION" | sed 's/v//' | sed 's/\.//g')
for migration in $(ls -1 "$MIGRATIONS_DIR"/*.sql 2>/dev/null | sort -V); do
MIGRATION_VERSION=$(basename "$migration" .sql | sed 's/[^0-9]//g')
if [ -n "$VERSION_NUM" ] && [ "$MIGRATION_VERSION" -gt "$VERSION_NUM" ]; then
echo "Skipping migration $migration (version $MIGRATION_VERSION > $VERSION_NUM)"
continue
fi
echo "Applying migration: $migration"
psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" -f "$migration"
done
fi
# Record schema version in metadata table
psql -v ON_ERROR_STOP=1 --username "$POSTGRES_USER" --dbname "$POSTGRES_DB" <<-EOSQL
CREATE TABLE IF NOT EXISTS _schema_metadata (
key TEXT PRIMARY KEY,
value TEXT NOT NULL,
updated_at TIMESTAMPTZ DEFAULT NOW()
);
INSERT INTO _schema_metadata (key, value)
VALUES
('module', '${STELLAOPS_MODULE}'),
('schema_version', '${STELLAOPS_SCHEMA_VERSION}'),
('initialized_at', NOW()::TEXT)
ON CONFLICT (key) DO UPDATE SET value = EXCLUDED.value, updated_at = NOW();
EOSQL
echo "Schema initialization complete for ${STELLAOPS_MODULE} version ${STELLAOPS_SCHEMA_VERSION}"

View File

@@ -0,0 +1,63 @@
# StellaOps Timeline Service
# Multi-stage build for optimized production image
FROM mcr.microsoft.com/dotnet/sdk:10.0-preview AS build
WORKDIR /src
# Copy solution and project files for restore
COPY ["src/Timeline/StellaOps.Timeline.WebService/StellaOps.Timeline.WebService.csproj", "src/Timeline/StellaOps.Timeline.WebService/"]
COPY ["src/Timeline/__Libraries/StellaOps.Timeline.Core/StellaOps.Timeline.Core.csproj", "src/Timeline/__Libraries/StellaOps.Timeline.Core/"]
COPY ["src/__Libraries/StellaOps.Eventing/StellaOps.Eventing.csproj", "src/__Libraries/StellaOps.Eventing/"]
COPY ["src/__Libraries/StellaOps.HybridLogicalClock/StellaOps.HybridLogicalClock.csproj", "src/__Libraries/StellaOps.HybridLogicalClock/"]
COPY ["src/__Libraries/StellaOps.Microservice/StellaOps.Microservice.csproj", "src/__Libraries/StellaOps.Microservice/"]
COPY ["src/__Libraries/StellaOps.Replay.Core/StellaOps.Replay.Core.csproj", "src/__Libraries/StellaOps.Replay.Core/"]
COPY ["nuget.config", "."]
COPY ["Directory.Build.props", "."]
COPY ["Directory.Packages.props", "."]
# Restore dependencies
RUN dotnet restore "src/Timeline/StellaOps.Timeline.WebService/StellaOps.Timeline.WebService.csproj"
# Copy source code
COPY ["src/", "src/"]
# Build
WORKDIR /src/src/Timeline/StellaOps.Timeline.WebService
RUN dotnet build -c Release -o /app/build --no-restore
# Publish
FROM build AS publish
RUN dotnet publish -c Release -o /app/publish --no-build /p:UseAppHost=false
# Runtime image
FROM mcr.microsoft.com/dotnet/aspnet:10.0-preview AS runtime
WORKDIR /app
# Create non-root user
RUN addgroup --system --gid 1000 stellaops && \
adduser --system --uid 1000 --ingroup stellaops stellaops
# Copy published files
COPY --from=publish /app/publish .
# Set ownership
RUN chown -R stellaops:stellaops /app
# Switch to non-root user
USER stellaops
# Environment configuration
ENV ASPNETCORE_URLS=http://+:8080 \
ASPNETCORE_ENVIRONMENT=Production \
DOTNET_EnableDiagnostics=0 \
DOTNET_SYSTEM_GLOBALIZATION_INVARIANT=false
# Health check
HEALTHCHECK --interval=30s --timeout=3s --start-period=5s --retries=3 \
CMD curl -f http://localhost:8080/health || exit 1
# Expose port
EXPOSE 8080
# Entry point
ENTRYPOINT ["dotnet", "StellaOps.Timeline.WebService.dll"]

View File

@@ -0,0 +1,119 @@
# HLC Queue Alerting Rules
# Sprint: SPRINT_20260105_002_004_BE_hlc_integration_tests
# Task: INT-018 - Create alerts for HLC anomalies
groups:
- name: hlc_alerts
interval: 1m
rules:
# Critical: Chain verification failures indicate tampering or corruption
- alert: HlcChainVerificationFailure
expr: increase(scheduler_chain_verification_failures_total[5m]) > 0
for: 1m
labels:
severity: critical
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#chain-verification-failure
annotations:
summary: "HLC chain verification failure detected"
description: "Chain verification failure on node {{ $labels.node_id }} for tenant {{ $labels.tenant_id }}. This may indicate data tampering or corruption."
impact: "Audit trail integrity compromised. Investigation required."
action: "1. Check scheduler_log table for gaps. 2. Verify no unauthorized changes. 3. Review chain head consistency."
# Critical: Clock skew exceeds tolerance - can cause ordering issues
- alert: HlcClockSkewExceedsTolerance
expr: increase(hlc_clock_skew_rejections_total[5m]) > 5
for: 2m
labels:
severity: critical
team: infrastructure
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#clock-skew
annotations:
summary: "HLC clock skew rejections on {{ $labels.node_id }}"
description: "Node {{ $labels.node_id }} is rejecting HLC updates due to clock skew. {{ $value }} rejections in last 5 minutes."
impact: "Job ordering may be inconsistent. Distributed consistency at risk."
action: "1. Check NTP synchronization on affected node. 2. Verify time sources. 3. Consider increasing skew tolerance temporarily."
# Warning: Physical time offset is drifting
- alert: HlcPhysicalTimeOffset
expr: abs(hlc_physical_time_offset_seconds) > 0.5
for: 5m
labels:
severity: warning
team: infrastructure
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#time-offset
annotations:
summary: "HLC physical time offset on {{ $labels.node_id }}"
description: "HLC physical time is {{ $value }}s offset from wall clock on {{ $labels.node_id }}."
impact: "May cause timestamp ordering anomalies in logs and diagnostics."
action: "Monitor NTP status and consider clock synchronization."
# Warning: High merge conflict rate in air-gap sync
- alert: HlcMergeConflictRateHigh
expr: increase(airgap_merge_conflicts_total[1h]) > 100
for: 10m
labels:
severity: warning
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#merge-conflicts
annotations:
summary: "High HLC merge conflict rate during air-gap sync"
description: "{{ $value }} merge conflicts detected in the last hour for conflict type {{ $labels.conflict_type }}."
impact: "Air-gap sync may be producing unexpected results or dropping jobs."
action: "1. Review conflict resolution logs. 2. Check for duplicate job submissions. 3. Verify offline node clocks."
# Warning: Air-gap sync duration increasing
- alert: HlcSyncDurationHigh
expr: histogram_quantile(0.95, sum(rate(airgap_sync_duration_seconds_bucket[15m])) by (le)) > 30
for: 10m
labels:
severity: warning
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#slow-sync
annotations:
summary: "Air-gap sync duration is high"
description: "95th percentile sync duration is {{ $value }}s, exceeding 30s threshold."
impact: "Air-gap import operations are slow, may delay job processing."
action: "1. Check bundle sizes. 2. Verify database performance. 3. Consider chunking large bundles."
# Info: HLC enqueue rate is zero (may be expected in some deployments)
- alert: HlcEnqueueRateZero
expr: sum(rate(scheduler_hlc_enqueues_total[10m])) == 0
for: 30m
labels:
severity: info
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#no-enqueues
annotations:
summary: "No HLC enqueues in last 30 minutes"
description: "No jobs have been enqueued with HLC timestamps in the last 30 minutes."
impact: "May be expected if no jobs are scheduled, or may indicate HLC ordering is disabled."
action: "Verify EnableHlcOrdering configuration if HLC ordering is expected."
# Warning: Batch snapshot creation failing
- alert: HlcBatchSnapshotFailures
expr: increase(scheduler_batch_snapshot_failures_total[5m]) > 0
for: 2m
labels:
severity: warning
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#batch-snapshot-failure
annotations:
summary: "Batch snapshot creation failures"
description: "{{ $value }} batch snapshot creation failures in the last 5 minutes."
impact: "DSSE-signed batch proofs may be missing for affected time ranges."
action: "1. Check signing key availability. 2. Verify database connectivity. 3. Review batch size limits."
# Critical: Multiple nodes with same node ID (configuration error)
- alert: HlcDuplicateNodeId
expr: count by (node_id) (group by (node_id, instance) (hlc_ticks_total)) > 1
for: 5m
labels:
severity: critical
team: scheduler
runbook: https://docs.stellaops.internal/operations/runbooks/hlc-troubleshooting#duplicate-node-id
annotations:
summary: "Duplicate HLC node ID detected"
description: "Multiple instances are using node_id={{ $labels.node_id }}. This will cause ordering conflicts."
impact: "Critical: Job ordering and chain integrity will be compromised."
action: "Immediately reconfigure affected instances with unique node IDs."

View File

@@ -0,0 +1,290 @@
{
"dashboard": {
"id": null,
"uid": "stellaops-hlc-metrics",
"title": "StellaOps HLC Queue Metrics",
"description": "Hybrid Logical Clock ordering metrics for the Scheduler queue",
"tags": ["stellaops", "hlc", "scheduler", "audit"],
"timezone": "utc",
"schemaVersion": 39,
"version": 1,
"refresh": "30s",
"time": {
"from": "now-1h",
"to": "now"
},
"panels": [
{
"id": 1,
"title": "HLC Tick Rate",
"description": "Rate of HLC tick operations per second",
"type": "timeseries",
"gridPos": { "h": 8, "w": 12, "x": 0, "y": 0 },
"fieldConfig": {
"defaults": {
"unit": "ops",
"custom": { "drawStyle": "line", "lineInterpolation": "smooth" }
}
},
"targets": [
{
"expr": "rate(hlc_ticks_total[1m])",
"legendFormat": "{{node_id}}",
"refId": "A"
}
]
},
{
"id": 2,
"title": "Clock Skew Rejections",
"description": "HLC rejections due to clock skew exceeding tolerance",
"type": "stat",
"gridPos": { "h": 4, "w": 6, "x": 12, "y": 0 },
"fieldConfig": {
"defaults": {
"unit": "short",
"thresholds": {
"mode": "absolute",
"steps": [
{ "color": "green", "value": null },
{ "color": "yellow", "value": 1 },
{ "color": "red", "value": 10 }
]
}
}
},
"targets": [
{
"expr": "sum(increase(hlc_clock_skew_rejections_total[1h]))",
"refId": "A"
}
]
},
{
"id": 3,
"title": "Physical Time Offset",
"description": "Difference between HLC physical time and wall clock",
"type": "gauge",
"gridPos": { "h": 4, "w": 6, "x": 18, "y": 0 },
"fieldConfig": {
"defaults": {
"unit": "ms",
"thresholds": {
"mode": "absolute",
"steps": [
{ "color": "green", "value": null },
{ "color": "yellow", "value": 100 },
{ "color": "red", "value": 1000 }
]
},
"max": 5000
}
},
"targets": [
{
"expr": "max(hlc_physical_time_offset_seconds) * 1000",
"refId": "A"
}
]
},
{
"id": 4,
"title": "Scheduler HLC Enqueues",
"description": "Rate of jobs enqueued with HLC timestamps",
"type": "timeseries",
"gridPos": { "h": 8, "w": 12, "x": 12, "y": 4 },
"fieldConfig": {
"defaults": {
"unit": "ops",
"custom": { "drawStyle": "bars", "fillOpacity": 50 }
}
},
"targets": [
{
"expr": "rate(scheduler_hlc_enqueues_total[5m])",
"legendFormat": "{{tenant_id}}",
"refId": "A"
}
]
},
{
"id": 5,
"title": "Chain Verifications",
"description": "Chain verification operations by result",
"type": "timeseries",
"gridPos": { "h": 8, "w": 12, "x": 0, "y": 8 },
"fieldConfig": {
"defaults": {
"unit": "ops"
},
"overrides": [
{
"matcher": { "id": "byName", "options": "valid" },
"properties": [{ "id": "color", "value": { "fixedColor": "green", "mode": "fixed" } }]
},
{
"matcher": { "id": "byName", "options": "invalid" },
"properties": [{ "id": "color", "value": { "fixedColor": "red", "mode": "fixed" } }]
}
]
},
"targets": [
{
"expr": "rate(scheduler_chain_verifications_total[5m])",
"legendFormat": "{{result}}",
"refId": "A"
}
]
},
{
"id": 6,
"title": "Verification Failures",
"description": "Chain verification failures - indicates tampering or corruption",
"type": "stat",
"gridPos": { "h": 4, "w": 6, "x": 12, "y": 8 },
"fieldConfig": {
"defaults": {
"unit": "short",
"thresholds": {
"mode": "absolute",
"steps": [
{ "color": "green", "value": null },
{ "color": "red", "value": 1 }
]
}
}
},
"targets": [
{
"expr": "sum(increase(scheduler_chain_verification_failures_total[1h]))",
"refId": "A"
}
]
},
{
"id": 7,
"title": "Batch Snapshots",
"description": "Batch snapshot creation rate",
"type": "stat",
"gridPos": { "h": 4, "w": 6, "x": 18, "y": 8 },
"fieldConfig": {
"defaults": {
"unit": "short"
}
},
"targets": [
{
"expr": "sum(increase(scheduler_batch_snapshots_total[1h]))",
"refId": "A"
}
]
},
{
"id": 8,
"title": "Air-Gap Bundle Exports",
"description": "Rate of air-gap bundles exported",
"type": "timeseries",
"gridPos": { "h": 8, "w": 8, "x": 0, "y": 16 },
"fieldConfig": {
"defaults": {
"unit": "ops"
}
},
"targets": [
{
"expr": "rate(airgap_bundles_exported_total[5m])",
"legendFormat": "{{node_id}}",
"refId": "A"
}
]
},
{
"id": 9,
"title": "Air-Gap Bundle Imports",
"description": "Rate of air-gap bundles imported",
"type": "timeseries",
"gridPos": { "h": 8, "w": 8, "x": 8, "y": 16 },
"fieldConfig": {
"defaults": {
"unit": "ops"
}
},
"targets": [
{
"expr": "rate(airgap_bundles_imported_total[5m])",
"legendFormat": "imported",
"refId": "A"
}
]
},
{
"id": 10,
"title": "Air-Gap Merge Conflicts",
"description": "Merge conflicts by type during air-gap sync",
"type": "stat",
"gridPos": { "h": 4, "w": 8, "x": 16, "y": 16 },
"fieldConfig": {
"defaults": {
"unit": "short",
"thresholds": {
"mode": "absolute",
"steps": [
{ "color": "green", "value": null },
{ "color": "yellow", "value": 1 },
{ "color": "red", "value": 10 }
]
}
}
},
"targets": [
{
"expr": "sum by (conflict_type) (increase(airgap_merge_conflicts_total[1h]))",
"legendFormat": "{{conflict_type}}",
"refId": "A"
}
]
},
{
"id": 11,
"title": "Sync Duration",
"description": "Air-gap sync operation duration percentiles",
"type": "timeseries",
"gridPos": { "h": 8, "w": 8, "x": 16, "y": 20 },
"fieldConfig": {
"defaults": {
"unit": "s"
}
},
"targets": [
{
"expr": "histogram_quantile(0.50, sum(rate(airgap_sync_duration_seconds_bucket[5m])) by (le))",
"legendFormat": "p50",
"refId": "A"
},
{
"expr": "histogram_quantile(0.95, sum(rate(airgap_sync_duration_seconds_bucket[5m])) by (le))",
"legendFormat": "p95",
"refId": "B"
},
{
"expr": "histogram_quantile(0.99, sum(rate(airgap_sync_duration_seconds_bucket[5m])) by (le))",
"legendFormat": "p99",
"refId": "C"
}
]
}
],
"annotations": {
"list": [
{
"name": "Deployments",
"datasource": "-- Grafana --",
"enable": true,
"iconColor": "blue"
}
]
}
},
"folderId": 0,
"overwrite": true
}

View File

@@ -4,6 +4,7 @@
<Nullable>enable</Nullable>
<ImplicitUsings>enable</ImplicitUsings>
<LangVersion>preview</LangVersion>
<TreatWarningsAsErrors>true</TreatWarningsAsErrors>
<AssetTargetFallback></AssetTargetFallback>
</PropertyGroup>
</Project>

View File

@@ -5,6 +5,7 @@
<ImplicitUsings>enable</ImplicitUsings>
<Nullable>enable</Nullable>
<LangVersion>preview</LangVersion>
<TreatWarningsAsErrors>true</TreatWarningsAsErrors>
<AssetTargetFallback></AssetTargetFallback>
</PropertyGroup>
</Project>

View File

@@ -8,5 +8,6 @@
<RuntimeIdentifier>linux-x64</RuntimeIdentifier>
<InvariantGlobalization>true</InvariantGlobalization>
<EnableTrimAnalyzer>false</EnableTrimAnalyzer>
<TreatWarningsAsErrors>true</TreatWarningsAsErrors>
</PropertyGroup>
</Project>

View File

@@ -0,0 +1,14 @@
# AirGap Docs Index
> **Note:** This directory contains **operational guides** for air-gap workflows. For module architecture and implementation details, see [docs/modules/airgap/](../modules/airgap/).
## Operational Guides
- Time anchors & staleness: `staleness-and-time.md`, `time-config-sample.json`, `time-api.md`, `time-anchor-verification-gap.md`
- Import pipeline: `importer.md`, `bundle-repositories.md`
- Controller/diagnostics: `controller.md`, `sealed-startup-diagnostics.md`
- Portable evidence flows: `portable-evidence.md`
- Offline bundle formats: `offline-bundle-format.md`
- Parity verification: `offline-parity-verification.md`
Use these as the front door for AirGap operational work; update alongside code changes.

View File

@@ -1,50 +1,50 @@
# StellaOps BOM Index (`bom-index@1`)
The BOM index is a deterministic, offline-friendly sidecar that accelerates queries for
layer-to-component membership and entrypoint usage. It is emitted alongside CycloneDX
SBOMs and consumed by Scheduler/Notify services.
## File Layout
Binary little-endian encoding, organised as the following sections:
1. **Header**
- `magic` (`byte[7]`): ASCII `"BOMIDX1"` identifier.
- `version` (`uint16`): current value `1`.
- `flags` (`uint16`): bit `0` set when entrypoint usage bitmaps are present.
- `imageDigestLength` (`uint16`) + UTF-8 digest string (e.g. `sha256:...`).
- `generatedAt` (`int64`): microseconds since Unix epoch.
- `layerCount` (`uint32`), `componentCount` (`uint32`), `entrypointCount` (`uint32`).
2. **Layer Table**
- For each layer: `length` (`uint16`) + UTF-8 layer digest (canonical order, base image → top layer).
3. **Component Table**
- For each component: `length` (`uint16`) + UTF-8 identity (CycloneDX purl when available, otherwise canonical key).
4. **Component ↦ Layer Bitmaps**
- For each component (matching table order):
- `bitmapLength` (`uint32`).
- Roaring bitmap payload (`Collections.Special.RoaringBitmap.Serialize`) encoding layer indexes that introduce or retain the component.
5. **Entrypoint Table** *(optional; present when `flags & 0x1 == 1`)*
- For each unique entrypoint/launcher string: `length` (`uint16`) + UTF-8 value (sorted ordinally).
6. **Component ↦ Entrypoint Bitmaps** *(optional)*
- For each component: roaring bitmap whose set bits reference entrypoint indexes used by EntryTrace. Empty bitmap (`length == 0`) indicates the component is not part of any resolved entrypoint closure.
## Determinism Guarantees
* Layer, component, and entrypoint tables are strictly ordered (base → top layer, lexicographically for components and entrypoints).
* Roaring bitmaps are optimised prior to serialisation and always produced from sorted indexes.
* Header timestamp is normalised to microsecond precision using UTC.
## Sample
`sample-index.bin` is generated from the integration fixture used in unit tests. It contains:
* 2 layers: `sha256:layer1`, `sha256:layer2`.
* 3 components: `pkg:npm/a`, `pkg:npm/b`, `pkg:npm/c`.
* Entrypoint bitmaps for `/app/start.sh` and `/app/init.sh`.
The sample can be decoded with the `BomIndexBuilder` unit tests or any RoaringBitmap implementation compatible with `Collections.Special.RoaringBitmap`.
# StellaOps BOM Index (`bom-index@1`)
The BOM index is a deterministic, offline-friendly sidecar that accelerates queries for
layer-to-component membership and entrypoint usage. It is emitted alongside CycloneDX
SBOMs and consumed by Scheduler/Notify services.
## File Layout
Binary little-endian encoding, organised as the following sections:
1. **Header**
- `magic` (`byte[7]`): ASCII `"BOMIDX1"` identifier.
- `version` (`uint16`): current value `1`.
- `flags` (`uint16`): bit `0` set when entrypoint usage bitmaps are present.
- `imageDigestLength` (`uint16`) + UTF-8 digest string (e.g. `sha256:...`).
- `generatedAt` (`int64`): microseconds since Unix epoch.
- `layerCount` (`uint32`), `componentCount` (`uint32`), `entrypointCount` (`uint32`).
2. **Layer Table**
- For each layer: `length` (`uint16`) + UTF-8 layer digest (canonical order, base image → top layer).
3. **Component Table**
- For each component: `length` (`uint16`) + UTF-8 identity (CycloneDX purl when available, otherwise canonical key).
4. **Component ↦ Layer Bitmaps**
- For each component (matching table order):
- `bitmapLength` (`uint32`).
- Roaring bitmap payload (`Collections.Special.RoaringBitmap.Serialize`) encoding layer indexes that introduce or retain the component.
5. **Entrypoint Table** *(optional; present when `flags & 0x1 == 1`)*
- For each unique entrypoint/launcher string: `length` (`uint16`) + UTF-8 value (sorted ordinally).
6. **Component ↦ Entrypoint Bitmaps** *(optional)*
- For each component: roaring bitmap whose set bits reference entrypoint indexes used by EntryTrace. Empty bitmap (`length == 0`) indicates the component is not part of any resolved entrypoint closure.
## Determinism Guarantees
* Layer, component, and entrypoint tables are strictly ordered (base → top layer, lexicographically for components and entrypoints).
* Roaring bitmaps are optimised prior to serialisation and always produced from sorted indexes.
* Header timestamp is normalised to microsecond precision using UTC.
## Sample
`sample-index.bin` is generated from the integration fixture used in unit tests. It contains:
* 2 layers: `sha256:layer1`, `sha256:layer2`.
* 3 components: `pkg:npm/a`, `pkg:npm/b`, `pkg:npm/c`.
* Entrypoint bitmaps for `/app/start.sh` and `/app/init.sh`.
The sample can be decoded with the `BomIndexBuilder` unit tests or any RoaringBitmap implementation compatible with `Collections.Special.RoaringBitmap`.

Some files were not shown because too many files have changed in this diff Show More