Compare commits
4 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| be4b439804 | |||
| 749fc8a94d | |||
| 6112094d45 | |||
| e9a2bc9f9d |
@@ -71,6 +71,8 @@ func run(args []string, stdout, stderr io.Writer) (exitCode int) {
|
|||||||
return runSAT(args[1:], stdout, stderr)
|
return runSAT(args[1:], stdout, stderr)
|
||||||
case "benchmark":
|
case "benchmark":
|
||||||
return runBenchmark(args[1:], stdout, stderr)
|
return runBenchmark(args[1:], stdout, stderr)
|
||||||
|
case "bee-worker":
|
||||||
|
return runBeeWorker(args[1:], stdout, stderr)
|
||||||
case "version", "--version", "-version":
|
case "version", "--version", "-version":
|
||||||
fmt.Fprintln(stdout, Version)
|
fmt.Fprintln(stdout, Version)
|
||||||
return 0
|
return 0
|
||||||
@@ -90,6 +92,7 @@ func printRootUsage(w io.Writer) {
|
|||||||
bee web --listen :80 [--audit-path `+app.DefaultAuditJSONPath+`]
|
bee web --listen :80 [--audit-path `+app.DefaultAuditJSONPath+`]
|
||||||
bee sat nvidia|memory|storage|cpu [--duration <seconds>]
|
bee sat nvidia|memory|storage|cpu [--duration <seconds>]
|
||||||
bee benchmark nvidia [--profile standard|stability|overnight]
|
bee benchmark nvidia [--profile standard|stability|overnight]
|
||||||
|
bee bee-worker --export-dir `+app.DefaultExportDir+` --task-id TASK-001
|
||||||
bee version
|
bee version
|
||||||
bee help [command]`)
|
bee help [command]`)
|
||||||
}
|
}
|
||||||
@@ -110,6 +113,8 @@ func runHelp(args []string, stdout, stderr io.Writer) int {
|
|||||||
return runSAT([]string{"--help"}, stdout, stderr)
|
return runSAT([]string{"--help"}, stdout, stderr)
|
||||||
case "benchmark":
|
case "benchmark":
|
||||||
return runBenchmark([]string{"--help"}, stdout, stderr)
|
return runBenchmark([]string{"--help"}, stdout, stderr)
|
||||||
|
case "bee-worker":
|
||||||
|
return runBeeWorker([]string{"--help"}, stdout, stderr)
|
||||||
case "version":
|
case "version":
|
||||||
fmt.Fprintln(stdout, "usage: bee version")
|
fmt.Fprintln(stdout, "usage: bee version")
|
||||||
return 0
|
return 0
|
||||||
@@ -462,6 +467,28 @@ func runBenchmark(args []string, stdout, stderr io.Writer) int {
|
|||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func runBeeWorker(args []string, stdout, stderr io.Writer) int {
|
||||||
|
fs := flag.NewFlagSet("bee-worker", flag.ContinueOnError)
|
||||||
|
fs.SetOutput(stderr)
|
||||||
|
exportDir := fs.String("export-dir", app.DefaultExportDir, "directory with task state and artifacts")
|
||||||
|
taskID := fs.String("task-id", "", "task identifier, e.g. TASK-001")
|
||||||
|
fs.Usage = func() {
|
||||||
|
fmt.Fprintf(stderr, "usage: bee bee-worker --export-dir %s --task-id TASK-001\n", app.DefaultExportDir)
|
||||||
|
fs.PrintDefaults()
|
||||||
|
}
|
||||||
|
if err := fs.Parse(args); err != nil {
|
||||||
|
if err == flag.ErrHelp {
|
||||||
|
return 0
|
||||||
|
}
|
||||||
|
return 2
|
||||||
|
}
|
||||||
|
if fs.NArg() != 0 {
|
||||||
|
fs.Usage()
|
||||||
|
return 2
|
||||||
|
}
|
||||||
|
return webui.RunPersistedTask(*exportDir, *taskID, stdout, stderr)
|
||||||
|
}
|
||||||
|
|
||||||
func parseBenchmarkIndexCSV(raw string) ([]int, error) {
|
func parseBenchmarkIndexCSV(raw string) ([]int, error) {
|
||||||
raw = strings.TrimSpace(raw)
|
raw = strings.TrimSpace(raw)
|
||||||
if raw == "" {
|
if raw == "" {
|
||||||
|
|||||||
@@ -105,6 +105,7 @@ var (
|
|||||||
benchmarkSkippedPattern = regexp.MustCompile(`^([a-z0-9_]+)(?:\[\d+\])?=SKIPPED (.+)$`)
|
benchmarkSkippedPattern = regexp.MustCompile(`^([a-z0-9_]+)(?:\[\d+\])?=SKIPPED (.+)$`)
|
||||||
benchmarkIterationsPattern = regexp.MustCompile(`^([a-z0-9_]+)_iterations=(\d+)$`)
|
benchmarkIterationsPattern = regexp.MustCompile(`^([a-z0-9_]+)_iterations=(\d+)$`)
|
||||||
benchmarkGeteuid = os.Geteuid
|
benchmarkGeteuid = os.Geteuid
|
||||||
|
benchmarkResetNvidiaGPU = resetNvidiaGPU
|
||||||
benchmarkSleep = time.Sleep
|
benchmarkSleep = time.Sleep
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -249,6 +250,35 @@ func setBenchmarkPowerLimit(ctx context.Context, verboseLog string, gpuIndex, po
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func resetBenchmarkGPU(ctx context.Context, verboseLog string, gpuIndex int, logFunc func(string)) error {
|
||||||
|
if logFunc != nil {
|
||||||
|
logFunc(fmt.Sprintf("power benchmark pre-flight: GPU %d reset via shared NVIDIA recover path", gpuIndex))
|
||||||
|
}
|
||||||
|
out, err := benchmarkResetNvidiaGPU(gpuIndex)
|
||||||
|
appendSATVerboseLog(verboseLog,
|
||||||
|
fmt.Sprintf("[%s] start power-preflight-gpu-%d-reset.log", time.Now().UTC().Format(time.RFC3339), gpuIndex),
|
||||||
|
"cmd: bee-nvidia-recover reset-gpu "+strconv.Itoa(gpuIndex),
|
||||||
|
)
|
||||||
|
if trimmed := strings.TrimSpace(out); trimmed != "" && logFunc != nil {
|
||||||
|
for _, line := range strings.Split(trimmed, "\n") {
|
||||||
|
line = strings.TrimSpace(line)
|
||||||
|
if line != "" {
|
||||||
|
logFunc(line)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
rc := 0
|
||||||
|
if err != nil {
|
||||||
|
rc = 1
|
||||||
|
}
|
||||||
|
appendSATVerboseLog(verboseLog,
|
||||||
|
fmt.Sprintf("[%s] finish power-preflight-gpu-%d-reset.log", time.Now().UTC().Format(time.RFC3339), gpuIndex),
|
||||||
|
fmt.Sprintf("rc: %d", rc),
|
||||||
|
"",
|
||||||
|
)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
func resetBenchmarkGPUs(ctx context.Context, verboseLog string, gpuIndices []int, logFunc func(string)) []int {
|
func resetBenchmarkGPUs(ctx context.Context, verboseLog string, gpuIndices []int, logFunc func(string)) []int {
|
||||||
if len(gpuIndices) == 0 {
|
if len(gpuIndices) == 0 {
|
||||||
return nil
|
return nil
|
||||||
@@ -266,8 +296,7 @@ func resetBenchmarkGPUs(ctx context.Context, verboseLog string, gpuIndices []int
|
|||||||
}
|
}
|
||||||
var failed []int
|
var failed []int
|
||||||
for _, idx := range gpuIndices {
|
for _, idx := range gpuIndices {
|
||||||
name := fmt.Sprintf("power-preflight-gpu-%d-reset.log", idx)
|
if err := resetBenchmarkGPU(ctx, verboseLog, idx, logFunc); err != nil {
|
||||||
if _, err := runSATCommandCtx(ctx, verboseLog, name, []string{"nvidia-smi", "-i", strconv.Itoa(idx), "-r"}, nil, logFunc); err != nil {
|
|
||||||
failed = append(failed, idx)
|
failed = append(failed, idx)
|
||||||
if logFunc != nil {
|
if logFunc != nil {
|
||||||
logFunc(fmt.Sprintf("power benchmark pre-flight: GPU %d reset failed: %v", idx, err))
|
logFunc(fmt.Sprintf("power benchmark pre-flight: GPU %d reset failed: %v", idx, err))
|
||||||
@@ -4440,8 +4469,7 @@ func (s *System) RunNvidiaPowerBench(ctx context.Context, baseDir string, opts N
|
|||||||
_ = os.MkdirAll(singleDir, 0755)
|
_ = os.MkdirAll(singleDir, 0755)
|
||||||
singleInfo := cloneBenchmarkGPUInfoMap(infoByIndex)
|
singleInfo := cloneBenchmarkGPUInfoMap(infoByIndex)
|
||||||
if failed := resetBenchmarkGPUs(ctx, verboseLog, []int{idx}, logFunc); len(failed) > 0 {
|
if failed := resetBenchmarkGPUs(ctx, verboseLog, []int{idx}, logFunc); len(failed) > 0 {
|
||||||
result.Findings = append(result.Findings,
|
return "", fmt.Errorf("power benchmark pre-flight: failed to reset GPU %d; benchmark aborted to keep measurements clean", idx)
|
||||||
fmt.Sprintf("GPU %d reset pre-flight did not complete before its first power test; throttle counters may contain stale state.", idx))
|
|
||||||
}
|
}
|
||||||
logFunc(fmt.Sprintf("power calibration: GPU %d single-card baseline", idx))
|
logFunc(fmt.Sprintf("power calibration: GPU %d single-card baseline", idx))
|
||||||
singlePowerStopCh := make(chan struct{})
|
singlePowerStopCh := make(chan struct{})
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package platform
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"os"
|
"fmt"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"strings"
|
"strings"
|
||||||
@@ -188,18 +188,16 @@ func TestBenchmarkCalibrationThrottleReasonIgnoresPowerReasons(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestResetBenchmarkGPUsSkipsWithoutRoot(t *testing.T) {
|
func TestResetBenchmarkGPUsSkipsWithoutRoot(t *testing.T) {
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
oldGeteuid := benchmarkGeteuid
|
oldGeteuid := benchmarkGeteuid
|
||||||
oldExec := satExecCommand
|
oldReset := benchmarkResetNvidiaGPU
|
||||||
benchmarkGeteuid = func() int { return 1000 }
|
benchmarkGeteuid = func() int { return 1000 }
|
||||||
satExecCommand = func(name string, args ...string) *exec.Cmd {
|
benchmarkResetNvidiaGPU = func(int) (string, error) {
|
||||||
t.Fatalf("unexpected command: %s %v", name, args)
|
t.Fatal("unexpected reset call")
|
||||||
return nil
|
return "", nil
|
||||||
}
|
}
|
||||||
t.Cleanup(func() {
|
t.Cleanup(func() {
|
||||||
benchmarkGeteuid = oldGeteuid
|
benchmarkGeteuid = oldGeteuid
|
||||||
satExecCommand = oldExec
|
benchmarkResetNvidiaGPU = oldReset
|
||||||
})
|
})
|
||||||
|
|
||||||
var logs []string
|
var logs []string
|
||||||
@@ -215,44 +213,52 @@ func TestResetBenchmarkGPUsSkipsWithoutRoot(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestResetBenchmarkGPUsResetsEachGPU(t *testing.T) {
|
func TestResetBenchmarkGPUsResetsEachGPU(t *testing.T) {
|
||||||
t.Parallel()
|
|
||||||
|
|
||||||
dir := t.TempDir()
|
|
||||||
script := filepath.Join(dir, "nvidia-smi")
|
|
||||||
argsLog := filepath.Join(dir, "args.log")
|
|
||||||
if err := os.WriteFile(script, []byte("#!/bin/sh\nprintf '%s\\n' \"$*\" >> "+argsLog+"\nprintf 'ok\\n'\n"), 0755); err != nil {
|
|
||||||
t.Fatalf("write script: %v", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
oldGeteuid := benchmarkGeteuid
|
oldGeteuid := benchmarkGeteuid
|
||||||
oldSleep := benchmarkSleep
|
oldSleep := benchmarkSleep
|
||||||
oldLookPath := satLookPath
|
oldReset := benchmarkResetNvidiaGPU
|
||||||
benchmarkGeteuid = func() int { return 0 }
|
benchmarkGeteuid = func() int { return 0 }
|
||||||
benchmarkSleep = func(time.Duration) {}
|
benchmarkSleep = func(time.Duration) {}
|
||||||
satLookPath = func(file string) (string, error) {
|
var calls []int
|
||||||
if file == "nvidia-smi" {
|
benchmarkResetNvidiaGPU = func(index int) (string, error) {
|
||||||
return script, nil
|
calls = append(calls, index)
|
||||||
}
|
return "ok\n", nil
|
||||||
return exec.LookPath(file)
|
|
||||||
}
|
}
|
||||||
t.Cleanup(func() {
|
t.Cleanup(func() {
|
||||||
benchmarkGeteuid = oldGeteuid
|
benchmarkGeteuid = oldGeteuid
|
||||||
benchmarkSleep = oldSleep
|
benchmarkSleep = oldSleep
|
||||||
satLookPath = oldLookPath
|
benchmarkResetNvidiaGPU = oldReset
|
||||||
})
|
})
|
||||||
|
|
||||||
failed := resetBenchmarkGPUs(context.Background(), filepath.Join(dir, "verbose.log"), []int{2, 5}, nil)
|
failed := resetBenchmarkGPUs(context.Background(), filepath.Join(t.TempDir(), "verbose.log"), []int{2, 5}, nil)
|
||||||
if len(failed) != 0 {
|
if len(failed) != 0 {
|
||||||
t.Fatalf("failed=%v want no failures", failed)
|
t.Fatalf("failed=%v want no failures", failed)
|
||||||
}
|
}
|
||||||
raw, err := os.ReadFile(argsLog)
|
if got, want := fmt.Sprint(calls), "[2 5]"; got != want {
|
||||||
if err != nil {
|
t.Fatalf("calls=%v want %s", calls, want)
|
||||||
t.Fatalf("read args log: %v", err)
|
|
||||||
}
|
}
|
||||||
got := strings.Fields(string(raw))
|
}
|
||||||
want := []string{"-i", "2", "-r", "-i", "5", "-r"}
|
|
||||||
if strings.Join(got, " ") != strings.Join(want, " ") {
|
func TestResetBenchmarkGPUsTracksFailuresFromSharedReset(t *testing.T) {
|
||||||
t.Fatalf("args=%v want %v", got, want)
|
oldGeteuid := benchmarkGeteuid
|
||||||
|
oldSleep := benchmarkSleep
|
||||||
|
oldReset := benchmarkResetNvidiaGPU
|
||||||
|
benchmarkGeteuid = func() int { return 0 }
|
||||||
|
benchmarkSleep = func(time.Duration) {}
|
||||||
|
benchmarkResetNvidiaGPU = func(index int) (string, error) {
|
||||||
|
if index == 5 {
|
||||||
|
return "busy\n", exec.ErrNotFound
|
||||||
|
}
|
||||||
|
return "ok\n", nil
|
||||||
|
}
|
||||||
|
t.Cleanup(func() {
|
||||||
|
benchmarkGeteuid = oldGeteuid
|
||||||
|
benchmarkSleep = oldSleep
|
||||||
|
benchmarkResetNvidiaGPU = oldReset
|
||||||
|
})
|
||||||
|
|
||||||
|
failed := resetBenchmarkGPUs(context.Background(), filepath.Join(t.TempDir(), "verbose.log"), []int{2, 5}, nil)
|
||||||
|
if got, want := fmt.Sprint(failed), "[5]"; got != want {
|
||||||
|
t.Fatalf("failed=%v want %s", failed, want)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -18,6 +18,7 @@ var workerPatterns = []string{
|
|||||||
"stress-ng",
|
"stress-ng",
|
||||||
"stressapptest",
|
"stressapptest",
|
||||||
"memtester",
|
"memtester",
|
||||||
|
"nvbandwidth",
|
||||||
// DCGM diagnostic workers — nvvs is spawned by dcgmi diag and survives
|
// DCGM diagnostic workers — nvvs is spawned by dcgmi diag and survives
|
||||||
// if dcgmi is killed mid-run, leaving the GPU occupied (DCGM_ST_IN_USE).
|
// if dcgmi is killed mid-run, leaving the GPU occupied (DCGM_ST_IN_USE).
|
||||||
"nvvs",
|
"nvvs",
|
||||||
@@ -71,13 +72,19 @@ func KillTestWorkers() []KilledProcess {
|
|||||||
if idx := strings.LastIndexByte(exe, '/'); idx >= 0 {
|
if idx := strings.LastIndexByte(exe, '/'); idx >= 0 {
|
||||||
base = exe[idx+1:]
|
base = exe[idx+1:]
|
||||||
}
|
}
|
||||||
for _, pat := range workerPatterns {
|
if shouldKillWorkerProcess(exe, base) {
|
||||||
if strings.Contains(base, pat) || strings.Contains(exe, pat) {
|
|
||||||
_ = syscall.Kill(pid, syscall.SIGKILL)
|
_ = syscall.Kill(pid, syscall.SIGKILL)
|
||||||
killed = append(killed, KilledProcess{PID: pid, Name: base})
|
killed = append(killed, KilledProcess{PID: pid, Name: base})
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return killed
|
return killed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func shouldKillWorkerProcess(exe, base string) bool {
|
||||||
|
for _, pat := range workerPatterns {
|
||||||
|
if strings.Contains(base, pat) || strings.Contains(exe, pat) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|||||||
39
audit/internal/platform/kill_workers_test.go
Normal file
39
audit/internal/platform/kill_workers_test.go
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
package platform
|
||||||
|
|
||||||
|
import "testing"
|
||||||
|
|
||||||
|
func TestShouldKillWorkerProcess(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
name string
|
||||||
|
exe string
|
||||||
|
base string
|
||||||
|
want bool
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
name: "nvbandwidth executable",
|
||||||
|
exe: "/usr/libexec/datacenter-gpu-manager-4/plugins/cuda13/nvbandwidth",
|
||||||
|
base: "nvbandwidth",
|
||||||
|
want: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "dcgmi executable",
|
||||||
|
exe: "/usr/bin/dcgmi",
|
||||||
|
base: "dcgmi",
|
||||||
|
want: true,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: "unrelated process",
|
||||||
|
exe: "/usr/bin/bash",
|
||||||
|
base: "bash",
|
||||||
|
want: false,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, tt := range tests {
|
||||||
|
t.Run(tt.name, func(t *testing.T) {
|
||||||
|
if got := shouldKillWorkerProcess(tt.exe, tt.base); got != tt.want {
|
||||||
|
t.Fatalf("shouldKillWorkerProcess(%q, %q)=%v want %v", tt.exe, tt.base, got, tt.want)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -3,6 +3,8 @@ package platform
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -28,3 +30,22 @@ func runNvidiaRecover(args ...string) (string, error) {
|
|||||||
raw, err := exec.Command("sudo", helperArgs...).CombinedOutput()
|
raw, err := exec.Command("sudo", helperArgs...).CombinedOutput()
|
||||||
return string(raw), err
|
return string(raw), err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func resetNvidiaGPU(index int) (string, error) {
|
||||||
|
if index < 0 {
|
||||||
|
return "", fmt.Errorf("gpu index must be >= 0")
|
||||||
|
}
|
||||||
|
out, err := runNvidiaRecover("reset-gpu", strconv.Itoa(index))
|
||||||
|
if strings.TrimSpace(out) == "" && err == nil {
|
||||||
|
out = "GPU reset completed.\n"
|
||||||
|
}
|
||||||
|
return out, err
|
||||||
|
}
|
||||||
|
|
||||||
|
func restartNvidiaDrivers() (string, error) {
|
||||||
|
out, err := runNvidiaRecover("restart-drivers")
|
||||||
|
if strings.TrimSpace(out) == "" && err == nil {
|
||||||
|
out = "NVIDIA drivers restarted.\n"
|
||||||
|
}
|
||||||
|
return out, err
|
||||||
|
}
|
||||||
|
|||||||
@@ -404,14 +404,7 @@ func normalizeNvidiaBusID(v string) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (s *System) ResetNvidiaGPU(index int) (string, error) {
|
func (s *System) ResetNvidiaGPU(index int) (string, error) {
|
||||||
if index < 0 {
|
return resetNvidiaGPU(index)
|
||||||
return "", fmt.Errorf("gpu index must be >= 0")
|
|
||||||
}
|
|
||||||
out, err := runNvidiaRecover("reset-gpu", strconv.Itoa(index))
|
|
||||||
if strings.TrimSpace(out) == "" && err == nil {
|
|
||||||
out = "GPU reset completed.\n"
|
|
||||||
}
|
|
||||||
return out, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// RunNCCLTests runs nccl-tests all_reduce_perf across the selected NVIDIA GPUs.
|
// RunNCCLTests runs nccl-tests all_reduce_perf across the selected NVIDIA GPUs.
|
||||||
|
|||||||
@@ -62,7 +62,7 @@ func (s *System) ServiceState(name string) string {
|
|||||||
|
|
||||||
func (s *System) ServiceDo(name string, action ServiceAction) (string, error) {
|
func (s *System) ServiceDo(name string, action ServiceAction) (string, error) {
|
||||||
if name == "bee-nvidia" && action == ServiceRestart {
|
if name == "bee-nvidia" && action == ServiceRestart {
|
||||||
return runNvidiaRecover("restart-drivers")
|
return restartNvidiaDrivers()
|
||||||
}
|
}
|
||||||
// bee-web runs as the bee user; sudo is required to control system services.
|
// bee-web runs as the bee user; sudo is required to control system services.
|
||||||
// /etc/sudoers.d/bee grants bee NOPASSWD:ALL.
|
// /etc/sudoers.d/bee grants bee NOPASSWD:ALL.
|
||||||
|
|||||||
@@ -806,15 +806,14 @@ func (h *handler) handleAPISATAbort(w http.ResponseWriter, r *http.Request) {
|
|||||||
now := time.Now()
|
now := time.Now()
|
||||||
t.DoneAt = &now
|
t.DoneAt = &now
|
||||||
case TaskRunning:
|
case TaskRunning:
|
||||||
if t.job != nil {
|
if t.job == nil || !t.job.abort() {
|
||||||
t.job.abort()
|
globalQueue.mu.Unlock()
|
||||||
|
writeJSON(w, map[string]string{"status": "not_running"})
|
||||||
|
return
|
||||||
}
|
}
|
||||||
if taskMayLeaveOrphanWorkers(t.Target) {
|
globalQueue.mu.Unlock()
|
||||||
platform.KillTestWorkers()
|
writeJSON(w, map[string]string{"status": "aborting"})
|
||||||
}
|
return
|
||||||
t.Status = TaskCancelled
|
|
||||||
now := time.Now()
|
|
||||||
t.DoneAt = &now
|
|
||||||
}
|
}
|
||||||
globalQueue.mu.Unlock()
|
globalQueue.mu.Unlock()
|
||||||
writeJSON(w, map[string]string{"status": "aborted"})
|
writeJSON(w, map[string]string{"status": "aborted"})
|
||||||
|
|||||||
@@ -53,13 +53,21 @@ func (j *jobState) abort() bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (j *jobState) append(line string) {
|
func (j *jobState) append(line string) {
|
||||||
|
j.appendWithOptions(line, true, true)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (j *jobState) appendFromLog(line string) {
|
||||||
|
j.appendWithOptions(line, false, false)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (j *jobState) appendWithOptions(line string, persistLog, serialMirror bool) {
|
||||||
j.mu.Lock()
|
j.mu.Lock()
|
||||||
defer j.mu.Unlock()
|
defer j.mu.Unlock()
|
||||||
j.lines = append(j.lines, line)
|
j.lines = append(j.lines, line)
|
||||||
if j.logPath != "" {
|
if persistLog && j.logPath != "" {
|
||||||
j.writeLogLineLocked(line)
|
j.writeLogLineLocked(line)
|
||||||
}
|
}
|
||||||
if j.serialPrefix != "" {
|
if serialMirror && j.serialPrefix != "" {
|
||||||
taskSerialWriteLine(j.serialPrefix + line)
|
taskSerialWriteLine(j.serialPrefix + line)
|
||||||
}
|
}
|
||||||
for _, ch := range j.subs {
|
for _, ch := range j.subs {
|
||||||
|
|||||||
@@ -207,7 +207,7 @@ func renderInstall() string {
|
|||||||
func renderTasks() string {
|
func renderTasks() string {
|
||||||
return `<div style="display:flex;align-items:center;gap:12px;margin-bottom:16px;flex-wrap:wrap">
|
return `<div style="display:flex;align-items:center;gap:12px;margin-bottom:16px;flex-wrap:wrap">
|
||||||
<button class="btn btn-danger btn-sm" onclick="cancelAll()">Cancel All</button>
|
<button class="btn btn-danger btn-sm" onclick="cancelAll()">Cancel All</button>
|
||||||
<button class="btn btn-sm" style="background:#b45309;color:#fff" onclick="killWorkers()" title="Send SIGKILL to all running test processes (bee-gpu-burn, stress-ng, stressapptest, memtester)">Kill Workers</button>
|
<button class="btn btn-sm" style="background:#b45309;color:#fff" onclick="killWorkers()" title="Abort running tasks and kill orphaned test processes (bee-gpu-burn, dcgmi, nvvs, nvbandwidth, stress-ng, stressapptest, memtester)">Abort Tasks And Kill Orphans</button>
|
||||||
<span id="kill-toast" style="font-size:12px;color:var(--muted);display:none"></span>
|
<span id="kill-toast" style="font-size:12px;color:var(--muted);display:none"></span>
|
||||||
<span style="font-size:12px;color:var(--muted)">Open a task to view its saved logs and charts.</span>
|
<span style="font-size:12px;color:var(--muted)">Open a task to view its saved logs and charts.</span>
|
||||||
</div>
|
</div>
|
||||||
@@ -289,7 +289,7 @@ function cancelAll() {
|
|||||||
fetch('/api/tasks/cancel-all',{method:'POST'}).then(()=>loadTasks());
|
fetch('/api/tasks/cancel-all',{method:'POST'}).then(()=>loadTasks());
|
||||||
}
|
}
|
||||||
function killWorkers() {
|
function killWorkers() {
|
||||||
if (!confirm('Send SIGKILL to all running test workers (bee-gpu-burn, stress-ng, stressapptest, memtester)?\n\nThis will also cancel all queued and running tasks.')) return;
|
if (!confirm('Abort all queued/running tasks and kill orphaned test workers (bee-gpu-burn, dcgmi, nvvs, nvbandwidth, stress-ng, stressapptest, memtester)?\n\nRunning bee-worker processes will first be asked to stop gracefully; orphaned test processes will then be killed.')) return;
|
||||||
fetch('/api/tasks/kill-workers',{method:'POST'})
|
fetch('/api/tasks/kill-workers',{method:'POST'})
|
||||||
.then(r=>r.json())
|
.then(r=>r.json())
|
||||||
.then(d=>{
|
.then(d=>{
|
||||||
|
|||||||
505
audit/internal/webui/task_runner.go
Normal file
505
audit/internal/webui/task_runner.go
Normal file
@@ -0,0 +1,505 @@
|
|||||||
|
package webui
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"log/slog"
|
||||||
|
"os"
|
||||||
|
"os/signal"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
"syscall"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"bee/audit/internal/app"
|
||||||
|
"bee/audit/internal/platform"
|
||||||
|
"bee/audit/internal/runtimeenv"
|
||||||
|
)
|
||||||
|
|
||||||
|
type taskRunnerState struct {
|
||||||
|
PID int `json:"pid"`
|
||||||
|
Status string `json:"status"`
|
||||||
|
Error string `json:"error,omitempty"`
|
||||||
|
UpdatedAt time.Time `json:"updated_at"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func taskRunnerStatePath(t *Task) string {
|
||||||
|
if t == nil || strings.TrimSpace(t.ArtifactsDir) == "" {
|
||||||
|
return ""
|
||||||
|
}
|
||||||
|
return filepath.Join(t.ArtifactsDir, "runner-state.json")
|
||||||
|
}
|
||||||
|
|
||||||
|
func writeTaskRunnerState(t *Task, state taskRunnerState) error {
|
||||||
|
path := taskRunnerStatePath(t)
|
||||||
|
if path == "" {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
if err := os.MkdirAll(filepath.Dir(path), 0755); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
data, err := json.MarshalIndent(state, "", " ")
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
tmp := path + ".tmp"
|
||||||
|
if err := os.WriteFile(tmp, data, 0644); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
return os.Rename(tmp, path)
|
||||||
|
}
|
||||||
|
|
||||||
|
func readTaskRunnerState(t *Task) (taskRunnerState, bool) {
|
||||||
|
path := taskRunnerStatePath(t)
|
||||||
|
if path == "" {
|
||||||
|
return taskRunnerState{}, false
|
||||||
|
}
|
||||||
|
data, err := os.ReadFile(path)
|
||||||
|
if err != nil || len(data) == 0 {
|
||||||
|
return taskRunnerState{}, false
|
||||||
|
}
|
||||||
|
var state taskRunnerState
|
||||||
|
if err := json.Unmarshal(data, &state); err != nil {
|
||||||
|
return taskRunnerState{}, false
|
||||||
|
}
|
||||||
|
return state, true
|
||||||
|
}
|
||||||
|
|
||||||
|
func processAlive(pid int) bool {
|
||||||
|
if pid <= 0 {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
err := syscall.Kill(pid, 0)
|
||||||
|
return err == nil || err == syscall.EPERM
|
||||||
|
}
|
||||||
|
|
||||||
|
func finalizeTaskForResult(t *Task, errMsg string, cancelled bool) {
|
||||||
|
now := time.Now()
|
||||||
|
t.DoneAt = &now
|
||||||
|
switch {
|
||||||
|
case cancelled:
|
||||||
|
t.Status = TaskCancelled
|
||||||
|
t.ErrMsg = "aborted"
|
||||||
|
case strings.TrimSpace(errMsg) != "":
|
||||||
|
t.Status = TaskFailed
|
||||||
|
t.ErrMsg = errMsg
|
||||||
|
default:
|
||||||
|
t.Status = TaskDone
|
||||||
|
t.ErrMsg = ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func executeTaskWithOptions(opts *HandlerOptions, t *Task, j *jobState, ctx context.Context) {
|
||||||
|
if opts == nil {
|
||||||
|
j.append("ERROR: handler options not configured")
|
||||||
|
j.finish("handler options not configured")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
a := opts.App
|
||||||
|
|
||||||
|
recovered := len(j.lines) > 0
|
||||||
|
j.append(fmt.Sprintf("Starting %s...", t.Name))
|
||||||
|
if recovered {
|
||||||
|
j.append(fmt.Sprintf("Recovered after bee-web restart at %s", time.Now().UTC().Format(time.RFC3339)))
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
archive string
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
|
||||||
|
switch t.Target {
|
||||||
|
case "nvidia":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
diagLevel := 2
|
||||||
|
if t.params.StressMode {
|
||||||
|
diagLevel = 3
|
||||||
|
}
|
||||||
|
if len(t.params.GPUIndices) > 0 || diagLevel > 0 {
|
||||||
|
result, e := a.RunNvidiaAcceptancePackWithOptions(ctx, "", diagLevel, t.params.GPUIndices, j.append)
|
||||||
|
if e != nil {
|
||||||
|
err = e
|
||||||
|
} else {
|
||||||
|
archive = result.Body
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
archive, err = a.RunNvidiaAcceptancePack("", j.append)
|
||||||
|
}
|
||||||
|
case "nvidia-targeted-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if dur <= 0 {
|
||||||
|
dur = 300
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaTargetedStressValidatePack(ctx, "", dur, t.params.GPUIndices, j.append)
|
||||||
|
case "nvidia-bench-perf":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaBenchmarkCtx(ctx, "", platform.NvidiaBenchmarkOptions{
|
||||||
|
Profile: t.params.BenchmarkProfile,
|
||||||
|
SizeMB: t.params.SizeMB,
|
||||||
|
GPUIndices: t.params.GPUIndices,
|
||||||
|
ExcludeGPUIndices: t.params.ExcludeGPUIndices,
|
||||||
|
RunNCCL: t.params.RunNCCL,
|
||||||
|
ParallelGPUs: t.params.ParallelGPUs,
|
||||||
|
RampStep: t.params.RampStep,
|
||||||
|
RampTotal: t.params.RampTotal,
|
||||||
|
RampRunID: t.params.RampRunID,
|
||||||
|
}, j.append)
|
||||||
|
case "nvidia-bench-power":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaPowerBenchCtx(ctx, app.DefaultBeeBenchPowerDir, platform.NvidiaBenchmarkOptions{
|
||||||
|
Profile: t.params.BenchmarkProfile,
|
||||||
|
GPUIndices: t.params.GPUIndices,
|
||||||
|
ExcludeGPUIndices: t.params.ExcludeGPUIndices,
|
||||||
|
RampStep: t.params.RampStep,
|
||||||
|
RampTotal: t.params.RampTotal,
|
||||||
|
RampRunID: t.params.RampRunID,
|
||||||
|
}, j.append)
|
||||||
|
case "nvidia-bench-autotune":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaPowerSourceAutotuneCtx(ctx, app.DefaultBeeBenchAutotuneDir, platform.NvidiaBenchmarkOptions{
|
||||||
|
Profile: t.params.BenchmarkProfile,
|
||||||
|
SizeMB: t.params.SizeMB,
|
||||||
|
}, t.params.BenchmarkKind, j.append)
|
||||||
|
case "nvidia-compute":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
rampPlan, planErr := resolveNvidiaRampPlan(t.params.BurnProfile, t.params.StaggerGPUStart, t.params.GPUIndices)
|
||||||
|
if planErr != nil {
|
||||||
|
err = planErr
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if t.params.BurnProfile != "" && t.params.StaggerGPUStart && dur <= 0 {
|
||||||
|
dur = rampPlan.DurationSec
|
||||||
|
}
|
||||||
|
if rampPlan.StaggerSeconds > 0 {
|
||||||
|
j.append(fmt.Sprintf("NVIDIA staggered ramp-up enabled: %ds per GPU; post-ramp hold: %ds; total runtime: %ds", rampPlan.StaggerSeconds, dur, rampPlan.TotalDurationSec))
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaOfficialComputePack(ctx, "", dur, t.params.GPUIndices, rampPlan.StaggerSeconds, j.append)
|
||||||
|
case "nvidia-targeted-power":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaTargetedPowerPack(ctx, "", dur, t.params.GPUIndices, j.append)
|
||||||
|
case "nvidia-pulse":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaPulseTestPack(ctx, "", dur, t.params.GPUIndices, j.append)
|
||||||
|
case "nvidia-bandwidth":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = a.RunNvidiaBandwidthPack(ctx, "", t.params.GPUIndices, j.append)
|
||||||
|
case "nvidia-interconnect":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = a.RunNCCLTests(ctx, "", t.params.GPUIndices, j.append)
|
||||||
|
case "nvidia-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
rampPlan, planErr := resolveNvidiaRampPlan(t.params.BurnProfile, t.params.StaggerGPUStart, t.params.GPUIndices)
|
||||||
|
if planErr != nil {
|
||||||
|
err = planErr
|
||||||
|
break
|
||||||
|
}
|
||||||
|
if t.params.BurnProfile != "" && t.params.StaggerGPUStart && dur <= 0 {
|
||||||
|
dur = rampPlan.DurationSec
|
||||||
|
}
|
||||||
|
if rampPlan.StaggerSeconds > 0 {
|
||||||
|
j.append(fmt.Sprintf("NVIDIA staggered ramp-up enabled: %ds per GPU; post-ramp hold: %ds; total runtime: %ds", rampPlan.StaggerSeconds, dur, rampPlan.TotalDurationSec))
|
||||||
|
}
|
||||||
|
archive, err = runNvidiaStressPackCtx(a, ctx, "", platform.NvidiaStressOptions{
|
||||||
|
DurationSec: dur,
|
||||||
|
Loader: t.params.Loader,
|
||||||
|
GPUIndices: t.params.GPUIndices,
|
||||||
|
ExcludeGPUIndices: t.params.ExcludeGPUIndices,
|
||||||
|
StaggerSeconds: rampPlan.StaggerSeconds,
|
||||||
|
}, j.append)
|
||||||
|
case "memory":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
sizeMB, passes := resolveMemoryValidatePreset(t.params.BurnProfile, t.params.StressMode)
|
||||||
|
j.append(fmt.Sprintf("Memory validate preset: %d MB x %d pass(es)", sizeMB, passes))
|
||||||
|
archive, err = runMemoryAcceptancePackCtx(a, ctx, "", sizeMB, passes, j.append)
|
||||||
|
case "storage":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = runStorageAcceptancePackCtx(a, ctx, "", t.params.StressMode, j.append)
|
||||||
|
case "cpu":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
if dur <= 0 {
|
||||||
|
if t.params.StressMode {
|
||||||
|
dur = 1800
|
||||||
|
} else {
|
||||||
|
dur = 60
|
||||||
|
}
|
||||||
|
}
|
||||||
|
j.append(fmt.Sprintf("CPU stress duration: %ds", dur))
|
||||||
|
archive, err = runCPUAcceptancePackCtx(a, ctx, "", dur, j.append)
|
||||||
|
case "amd":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = runAMDAcceptancePackCtx(a, ctx, "", j.append)
|
||||||
|
case "amd-mem":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = runAMDMemIntegrityPackCtx(a, ctx, "", j.append)
|
||||||
|
case "amd-bandwidth":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
archive, err = runAMDMemBandwidthPackCtx(a, ctx, "", j.append)
|
||||||
|
case "amd-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
archive, err = runAMDStressPackCtx(a, ctx, "", dur, j.append)
|
||||||
|
case "memory-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
archive, err = runMemoryStressPackCtx(a, ctx, "", dur, j.append)
|
||||||
|
case "sat-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
dur := t.params.Duration
|
||||||
|
if t.params.BurnProfile != "" && dur <= 0 {
|
||||||
|
dur = resolveBurnPreset(t.params.BurnProfile).DurationSec
|
||||||
|
}
|
||||||
|
archive, err = runSATStressPackCtx(a, ctx, "", dur, j.append)
|
||||||
|
case "platform-stress":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
runOpts := resolvePlatformStressPreset(t.params.BurnProfile)
|
||||||
|
runOpts.Components = t.params.PlatformComponents
|
||||||
|
archive, err = a.RunPlatformStress(ctx, "", runOpts, j.append)
|
||||||
|
case "audit":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
result, e := a.RunAuditNow(opts.RuntimeMode)
|
||||||
|
if e != nil {
|
||||||
|
err = e
|
||||||
|
} else {
|
||||||
|
for _, line := range splitLines(result.Body) {
|
||||||
|
j.append(line)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case "support-bundle":
|
||||||
|
j.append("Building support bundle...")
|
||||||
|
archive, err = buildSupportBundle(opts.ExportDir)
|
||||||
|
case "install":
|
||||||
|
if strings.TrimSpace(t.params.Device) == "" {
|
||||||
|
err = fmt.Errorf("device is required")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
installLogPath := platform.InstallLogPath(t.params.Device)
|
||||||
|
j.append("Install log: " + installLogPath)
|
||||||
|
err = streamCmdJob(j, installCommand(ctx, t.params.Device, installLogPath))
|
||||||
|
case "install-to-ram":
|
||||||
|
if a == nil {
|
||||||
|
err = fmt.Errorf("app not configured")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
err = a.RunInstallToRAM(ctx, j.append)
|
||||||
|
default:
|
||||||
|
j.append("ERROR: unknown target: " + t.Target)
|
||||||
|
j.finish("unknown target")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
if archive != "" {
|
||||||
|
archivePath := app.ExtractArchivePath(archive)
|
||||||
|
if err == nil && app.ReadSATOverallStatus(archivePath) == "FAILED" {
|
||||||
|
err = fmt.Errorf("SAT overall_status=FAILED (see summary.txt)")
|
||||||
|
}
|
||||||
|
if opts.App != nil && opts.App.StatusDB != nil {
|
||||||
|
app.ApplySATResultToDB(opts.App.StatusDB, t.Target, archivePath)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
if ctx.Err() != nil {
|
||||||
|
j.append("Aborted.")
|
||||||
|
j.finish("aborted")
|
||||||
|
} else {
|
||||||
|
j.append("ERROR: " + err.Error())
|
||||||
|
j.finish(err.Error())
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if archive != "" {
|
||||||
|
j.append("Archive: " + archive)
|
||||||
|
}
|
||||||
|
j.finish("")
|
||||||
|
}
|
||||||
|
|
||||||
|
func loadPersistedTask(statePath, taskID string) (*Task, error) {
|
||||||
|
data, err := os.ReadFile(statePath)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
var persisted []persistedTask
|
||||||
|
if err := json.Unmarshal(data, &persisted); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
for _, pt := range persisted {
|
||||||
|
if pt.ID != taskID {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
t := &Task{
|
||||||
|
ID: pt.ID,
|
||||||
|
Name: pt.Name,
|
||||||
|
Target: pt.Target,
|
||||||
|
Priority: pt.Priority,
|
||||||
|
Status: pt.Status,
|
||||||
|
CreatedAt: pt.CreatedAt,
|
||||||
|
StartedAt: pt.StartedAt,
|
||||||
|
DoneAt: pt.DoneAt,
|
||||||
|
ErrMsg: pt.ErrMsg,
|
||||||
|
LogPath: pt.LogPath,
|
||||||
|
ArtifactsDir: pt.ArtifactsDir,
|
||||||
|
ReportJSONPath: pt.ReportJSONPath,
|
||||||
|
ReportHTMLPath: pt.ReportHTMLPath,
|
||||||
|
params: pt.Params,
|
||||||
|
}
|
||||||
|
ensureTaskReportPaths(t)
|
||||||
|
return t, nil
|
||||||
|
}
|
||||||
|
return nil, fmt.Errorf("task %s not found", taskID)
|
||||||
|
}
|
||||||
|
|
||||||
|
func RunPersistedTask(exportDir, taskID string, stdout, stderr io.Writer) int {
|
||||||
|
if strings.TrimSpace(exportDir) == "" || strings.TrimSpace(taskID) == "" {
|
||||||
|
fmt.Fprintln(stderr, "bee task-run: --export-dir and --task-id are required")
|
||||||
|
return 2
|
||||||
|
}
|
||||||
|
|
||||||
|
runtimeInfo, err := runtimeenv.Detect("auto")
|
||||||
|
if err != nil {
|
||||||
|
slog.Warn("resolve runtime for task-run", "err", err)
|
||||||
|
}
|
||||||
|
opts := &HandlerOptions{
|
||||||
|
ExportDir: exportDir,
|
||||||
|
App: app.New(platform.New()),
|
||||||
|
RuntimeMode: runtimeInfo.Mode,
|
||||||
|
}
|
||||||
|
statePath := filepath.Join(exportDir, "tasks-state.json")
|
||||||
|
task, err := loadPersistedTask(statePath, taskID)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Fprintln(stderr, err.Error())
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
if task.StartedAt == nil || task.StartedAt.IsZero() {
|
||||||
|
now := time.Now()
|
||||||
|
task.StartedAt = &now
|
||||||
|
}
|
||||||
|
if task.Status == "" {
|
||||||
|
task.Status = TaskRunning
|
||||||
|
}
|
||||||
|
if err := writeTaskRunnerState(task, taskRunnerState{
|
||||||
|
PID: os.Getpid(),
|
||||||
|
Status: TaskRunning,
|
||||||
|
UpdatedAt: time.Now().UTC(),
|
||||||
|
}); err != nil {
|
||||||
|
fmt.Fprintln(stderr, err.Error())
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
ctx, cancel := signal.NotifyContext(context.Background(), syscall.SIGINT, syscall.SIGTERM)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
j := newTaskJobState(task.LogPath, taskSerialPrefix(task))
|
||||||
|
executeTaskWithOptions(opts, task, j, ctx)
|
||||||
|
finalizeTaskForResult(task, j.err, ctx.Err() != nil)
|
||||||
|
if err := writeTaskReportArtifacts(task); err != nil {
|
||||||
|
appendJobLog(task.LogPath, "WARN: task report generation failed: "+err.Error())
|
||||||
|
}
|
||||||
|
j.closeLog()
|
||||||
|
if err := writeTaskRunnerState(task, taskRunnerState{
|
||||||
|
PID: os.Getpid(),
|
||||||
|
Status: task.Status,
|
||||||
|
Error: task.ErrMsg,
|
||||||
|
UpdatedAt: time.Now().UTC(),
|
||||||
|
}); err != nil {
|
||||||
|
fmt.Fprintln(stderr, err.Error())
|
||||||
|
}
|
||||||
|
if task.ErrMsg != "" {
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
return 0
|
||||||
|
}
|
||||||
@@ -4,6 +4,7 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"log/slog"
|
"log/slog"
|
||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
@@ -13,6 +14,7 @@ import (
|
|||||||
"sort"
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"bee/audit/internal/app"
|
"bee/audit/internal/app"
|
||||||
@@ -111,6 +113,7 @@ type Task struct {
|
|||||||
|
|
||||||
// runtime fields (not serialised)
|
// runtime fields (not serialised)
|
||||||
job *jobState
|
job *jobState
|
||||||
|
runnerPID int
|
||||||
params taskParams
|
params taskParams
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -328,6 +331,13 @@ var (
|
|||||||
installCommand = func(ctx context.Context, device string, logPath string) *exec.Cmd {
|
installCommand = func(ctx context.Context, device string, logPath string) *exec.Cmd {
|
||||||
return exec.CommandContext(ctx, "bee-install", device, logPath)
|
return exec.CommandContext(ctx, "bee-install", device, logPath)
|
||||||
}
|
}
|
||||||
|
externalTaskRunnerCommand = func(exportDir, taskID string) (*exec.Cmd, error) {
|
||||||
|
exe, err := os.Executable()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return exec.Command(exe, "bee-worker", "--export-dir", exportDir, "--task-id", taskID), nil
|
||||||
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
// enqueue adds a task to the queue and notifies the worker.
|
// enqueue adds a task to the queue and notifies the worker.
|
||||||
@@ -365,6 +375,11 @@ func (q *taskQueue) prune() {
|
|||||||
|
|
||||||
// nextPending returns the highest-priority pending task (nil if none).
|
// nextPending returns the highest-priority pending task (nil if none).
|
||||||
func (q *taskQueue) nextPending() *Task {
|
func (q *taskQueue) nextPending() *Task {
|
||||||
|
for _, t := range q.tasks {
|
||||||
|
if t.Status == TaskRunning {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
var best *Task
|
var best *Task
|
||||||
for _, t := range q.tasks {
|
for _, t := range q.tasks {
|
||||||
if t.Status != TaskPending {
|
if t.Status != TaskPending {
|
||||||
@@ -484,6 +499,7 @@ func (q *taskQueue) startWorker(opts *HandlerOptions) {
|
|||||||
if !q.started {
|
if !q.started {
|
||||||
q.loadLocked()
|
q.loadLocked()
|
||||||
q.started = true
|
q.started = true
|
||||||
|
q.resumeRunningTasksLocked()
|
||||||
goRecoverLoop("task worker", 2*time.Second, q.worker)
|
goRecoverLoop("task worker", 2*time.Second, q.worker)
|
||||||
}
|
}
|
||||||
hasPending := q.nextPending() != nil
|
hasPending := q.nextPending() != nil
|
||||||
@@ -517,15 +533,12 @@ func (q *taskQueue) worker() {
|
|||||||
t.StartedAt = &now
|
t.StartedAt = &now
|
||||||
t.DoneAt = nil
|
t.DoneAt = nil
|
||||||
t.ErrMsg = ""
|
t.ErrMsg = ""
|
||||||
j := newTaskJobState(t.LogPath, taskSerialPrefix(t))
|
j := newTaskJobState(t.LogPath)
|
||||||
t.job = j
|
t.job = j
|
||||||
q.persistLocked()
|
q.persistLocked()
|
||||||
q.mu.Unlock()
|
q.mu.Unlock()
|
||||||
|
|
||||||
taskCtx, taskCancel := context.WithCancel(context.Background())
|
q.runTaskExternal(t, j)
|
||||||
j.cancel = taskCancel
|
|
||||||
q.executeTask(t, j, taskCtx)
|
|
||||||
taskCancel()
|
|
||||||
|
|
||||||
q.mu.Lock()
|
q.mu.Lock()
|
||||||
q.prune()
|
q.prune()
|
||||||
@@ -537,6 +550,207 @@ func (q *taskQueue) worker() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) resumeRunningTasksLocked() {
|
||||||
|
for _, t := range q.tasks {
|
||||||
|
if t.Status != TaskRunning {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if t.job == nil {
|
||||||
|
t.job = newTaskJobState(t.LogPath)
|
||||||
|
}
|
||||||
|
q.attachExternalTaskControlsLocked(t, t.job)
|
||||||
|
q.startRecoveredTaskMonitorLocked(t, t.job)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) attachExternalTaskControlsLocked(t *Task, j *jobState) {
|
||||||
|
if t == nil || j == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
j.cancel = func() {
|
||||||
|
pid := t.runnerPID
|
||||||
|
if pid <= 0 {
|
||||||
|
if state, ok := readTaskRunnerState(t); ok {
|
||||||
|
pid = state.PID
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if pid > 0 {
|
||||||
|
_ = syscall.Kill(pid, syscall.SIGTERM)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) startRecoveredTaskMonitorLocked(t *Task, j *jobState) {
|
||||||
|
if t == nil || j == nil || t.runnerPID <= 0 {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
goRecoverOnce("task runner monitor", func() {
|
||||||
|
stopTail := make(chan struct{})
|
||||||
|
doneTail := make(chan struct{})
|
||||||
|
go q.followTaskLog(t, j, stopTail, doneTail)
|
||||||
|
for processAlive(t.runnerPID) {
|
||||||
|
time.Sleep(500 * time.Millisecond)
|
||||||
|
}
|
||||||
|
close(stopTail)
|
||||||
|
<-doneTail
|
||||||
|
q.finishExternalTask(t, j, nil)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) runTaskExternal(t *Task, j *jobState) {
|
||||||
|
stopTail := make(chan struct{})
|
||||||
|
doneTail := make(chan struct{})
|
||||||
|
defer func() {
|
||||||
|
close(stopTail)
|
||||||
|
<-doneTail
|
||||||
|
}()
|
||||||
|
go q.followTaskLog(t, j, stopTail, doneTail)
|
||||||
|
|
||||||
|
cmd, err := externalTaskRunnerCommand(q.opts.ExportDir, t.ID)
|
||||||
|
if err != nil {
|
||||||
|
j.appendFromLog("ERROR: " + err.Error())
|
||||||
|
q.finishExternalTask(t, j, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
if err := cmd.Start(); err != nil {
|
||||||
|
j.appendFromLog("ERROR: " + err.Error())
|
||||||
|
q.finishExternalTask(t, j, err)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
q.mu.Lock()
|
||||||
|
t.runnerPID = cmd.Process.Pid
|
||||||
|
q.attachExternalTaskControlsLocked(t, j)
|
||||||
|
q.persistLocked()
|
||||||
|
q.mu.Unlock()
|
||||||
|
|
||||||
|
waitErr := cmd.Wait()
|
||||||
|
time.Sleep(200 * time.Millisecond)
|
||||||
|
q.finishExternalTask(t, j, waitErr)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) followTaskLog(t *Task, j *jobState, stop <-chan struct{}, done chan<- struct{}) {
|
||||||
|
defer close(done)
|
||||||
|
path := ""
|
||||||
|
if t != nil {
|
||||||
|
path = t.LogPath
|
||||||
|
}
|
||||||
|
if strings.TrimSpace(path) == "" {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
offset := int64(0)
|
||||||
|
if info, err := os.Stat(path); err == nil {
|
||||||
|
offset = info.Size()
|
||||||
|
}
|
||||||
|
var partial string
|
||||||
|
ticker := time.NewTicker(250 * time.Millisecond)
|
||||||
|
defer ticker.Stop()
|
||||||
|
flush := func() {
|
||||||
|
data, newOffset, err := readTaskLogDelta(path, offset)
|
||||||
|
if err != nil || len(data) == 0 {
|
||||||
|
offset = newOffset
|
||||||
|
return
|
||||||
|
}
|
||||||
|
offset = newOffset
|
||||||
|
text := partial + strings.ReplaceAll(string(data), "\r\n", "\n")
|
||||||
|
lines := strings.Split(text, "\n")
|
||||||
|
partial = lines[len(lines)-1]
|
||||||
|
for _, line := range lines[:len(lines)-1] {
|
||||||
|
if line == "" {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
j.appendFromLog(line)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-ticker.C:
|
||||||
|
flush()
|
||||||
|
case <-stop:
|
||||||
|
flush()
|
||||||
|
if strings.TrimSpace(partial) != "" {
|
||||||
|
j.appendFromLog(partial)
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func readTaskLogDelta(path string, offset int64) ([]byte, int64, error) {
|
||||||
|
f, err := os.Open(path)
|
||||||
|
if err != nil {
|
||||||
|
return nil, offset, err
|
||||||
|
}
|
||||||
|
defer f.Close()
|
||||||
|
info, err := f.Stat()
|
||||||
|
if err != nil {
|
||||||
|
return nil, offset, err
|
||||||
|
}
|
||||||
|
if info.Size() < offset {
|
||||||
|
offset = 0
|
||||||
|
}
|
||||||
|
if _, err := f.Seek(offset, io.SeekStart); err != nil {
|
||||||
|
return nil, offset, err
|
||||||
|
}
|
||||||
|
data, err := io.ReadAll(io.LimitReader(f, 1<<20))
|
||||||
|
return data, offset + int64(len(data)), err
|
||||||
|
}
|
||||||
|
|
||||||
|
func (q *taskQueue) finishExternalTask(t *Task, j *jobState, waitErr error) {
|
||||||
|
q.mu.Lock()
|
||||||
|
defer q.mu.Unlock()
|
||||||
|
if t.Status == TaskDone || t.Status == TaskFailed || t.Status == TaskCancelled {
|
||||||
|
if j != nil && !j.isDone() {
|
||||||
|
j.finish(t.ErrMsg)
|
||||||
|
j.closeLog()
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case q.trigger <- struct{}{}:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
state, ok := readTaskRunnerState(t)
|
||||||
|
switch {
|
||||||
|
case ok && state.Status != TaskRunning:
|
||||||
|
t.Status = state.Status
|
||||||
|
t.ErrMsg = state.Error
|
||||||
|
now := state.UpdatedAt
|
||||||
|
if now.IsZero() {
|
||||||
|
now = time.Now()
|
||||||
|
}
|
||||||
|
t.DoneAt = &now
|
||||||
|
case waitErr != nil:
|
||||||
|
now := time.Now()
|
||||||
|
t.Status = TaskFailed
|
||||||
|
t.ErrMsg = waitErr.Error()
|
||||||
|
t.DoneAt = &now
|
||||||
|
default:
|
||||||
|
now := time.Now()
|
||||||
|
t.Status = TaskFailed
|
||||||
|
t.ErrMsg = "task runner exited without final state"
|
||||||
|
t.DoneAt = &now
|
||||||
|
}
|
||||||
|
t.runnerPID = 0
|
||||||
|
q.finalizeTaskArtifactPathsLocked(t)
|
||||||
|
q.persistLocked()
|
||||||
|
|
||||||
|
if j != nil && !j.isDone() {
|
||||||
|
j.finish(t.ErrMsg)
|
||||||
|
j.closeLog()
|
||||||
|
}
|
||||||
|
if t.ErrMsg != "" {
|
||||||
|
taskSerialEvent(t, "finished with status="+t.Status+" error="+t.ErrMsg)
|
||||||
|
} else {
|
||||||
|
taskSerialEvent(t, "finished with status="+t.Status)
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case q.trigger <- struct{}{}:
|
||||||
|
default:
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (q *taskQueue) executeTask(t *Task, j *jobState, ctx context.Context) {
|
func (q *taskQueue) executeTask(t *Task, j *jobState, ctx context.Context) {
|
||||||
startedKmsgWatch := false
|
startedKmsgWatch := false
|
||||||
defer q.finalizeTaskRun(t, j)
|
defer q.finalizeTaskRun(t, j)
|
||||||
@@ -985,15 +1199,11 @@ func (h *handler) handleAPITasksCancel(w http.ResponseWriter, r *http.Request) {
|
|||||||
taskSerialEvent(t, "finished with status="+t.Status)
|
taskSerialEvent(t, "finished with status="+t.Status)
|
||||||
writeJSON(w, map[string]string{"status": "cancelled"})
|
writeJSON(w, map[string]string{"status": "cancelled"})
|
||||||
case TaskRunning:
|
case TaskRunning:
|
||||||
if t.job != nil {
|
if t.job == nil || !t.job.abort() {
|
||||||
t.job.abort()
|
writeError(w, http.StatusConflict, "task is not cancellable")
|
||||||
|
return
|
||||||
}
|
}
|
||||||
t.Status = TaskCancelled
|
writeJSON(w, map[string]string{"status": "aborting"})
|
||||||
now := time.Now()
|
|
||||||
t.DoneAt = &now
|
|
||||||
globalQueue.persistLocked()
|
|
||||||
taskSerialEvent(t, "finished with status="+t.Status)
|
|
||||||
writeJSON(w, map[string]string{"status": "cancelled"})
|
|
||||||
default:
|
default:
|
||||||
writeError(w, http.StatusConflict, "task is not running or pending")
|
writeError(w, http.StatusConflict, "task is not running or pending")
|
||||||
}
|
}
|
||||||
@@ -1039,12 +1249,6 @@ func (h *handler) handleAPITasksCancelAll(w http.ResponseWriter, _ *http.Request
|
|||||||
if t.job != nil {
|
if t.job != nil {
|
||||||
t.job.abort()
|
t.job.abort()
|
||||||
}
|
}
|
||||||
if taskMayLeaveOrphanWorkers(t.Target) {
|
|
||||||
platform.KillTestWorkers()
|
|
||||||
}
|
|
||||||
t.Status = TaskCancelled
|
|
||||||
t.DoneAt = &now
|
|
||||||
taskSerialEvent(t, "finished with status="+t.Status)
|
|
||||||
n++
|
n++
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -1175,11 +1379,21 @@ func (q *taskQueue) loadLocked() {
|
|||||||
}
|
}
|
||||||
q.assignTaskLogPathLocked(t)
|
q.assignTaskLogPathLocked(t)
|
||||||
if t.Status == TaskRunning {
|
if t.Status == TaskRunning {
|
||||||
// The task was interrupted by a bee-web restart. Child processes
|
state, ok := readTaskRunnerState(t)
|
||||||
// (e.g. bee-gpu-burn-worker, dcgmi/nvvs) survive the restart in
|
switch {
|
||||||
// their own process groups. Kill any matching stale workers before
|
case ok && state.Status == TaskRunning && processAlive(state.PID):
|
||||||
// marking the task failed so the next GPU test does not inherit a
|
t.runnerPID = state.PID
|
||||||
// busy DCGM slot or duplicate workers.
|
t.job = newTaskJobState(t.LogPath)
|
||||||
|
case ok && state.Status != TaskRunning:
|
||||||
|
t.runnerPID = state.PID
|
||||||
|
t.Status = state.Status
|
||||||
|
t.ErrMsg = state.Error
|
||||||
|
now := state.UpdatedAt
|
||||||
|
if now.IsZero() {
|
||||||
|
now = time.Now()
|
||||||
|
}
|
||||||
|
t.DoneAt = &now
|
||||||
|
default:
|
||||||
if taskMayLeaveOrphanWorkers(t.Target) {
|
if taskMayLeaveOrphanWorkers(t.Target) {
|
||||||
_ = platform.KillTestWorkers()
|
_ = platform.KillTestWorkers()
|
||||||
}
|
}
|
||||||
@@ -1187,6 +1401,7 @@ func (q *taskQueue) loadLocked() {
|
|||||||
t.Status = TaskFailed
|
t.Status = TaskFailed
|
||||||
t.DoneAt = &now
|
t.DoneAt = &now
|
||||||
t.ErrMsg = "interrupted by bee-web restart"
|
t.ErrMsg = "interrupted by bee-web restart"
|
||||||
|
}
|
||||||
} else if t.Status == TaskPending {
|
} else if t.Status == TaskPending {
|
||||||
t.StartedAt = nil
|
t.StartedAt = nil
|
||||||
t.DoneAt = nil
|
t.DoneAt = nil
|
||||||
|
|||||||
2
bible
2
bible
Submodule bible updated: 98448c993f...1d89a4918e
@@ -23,9 +23,9 @@ insmod serial
|
|||||||
serial --unit=0 --speed=115200 --word=8 --parity=no --stop=1
|
serial --unit=0 --speed=115200 --word=8 --parity=no --stop=1
|
||||||
|
|
||||||
insmod gfxterm
|
insmod gfxterm
|
||||||
insmod png
|
|
||||||
|
|
||||||
source /boot/grub/theme.cfg
|
|
||||||
|
|
||||||
terminal_input console serial
|
terminal_input console serial
|
||||||
terminal_output gfxterm serial
|
terminal_output gfxterm serial
|
||||||
|
|
||||||
|
insmod png
|
||||||
|
source /boot/grub/theme.cfg
|
||||||
|
|||||||
@@ -1,15 +1,5 @@
|
|||||||
source /boot/grub/config.cfg
|
source /boot/grub/config.cfg
|
||||||
|
|
||||||
echo ""
|
|
||||||
echo " ███████╗ █████╗ ███████╗██╗ ██╗ ██████╗ ███████╗███████╗"
|
|
||||||
echo " ██╔════╝██╔══██╗██╔════╝╚██╗ ██╔╝ ██╔══██╗██╔════╝██╔════╝"
|
|
||||||
echo " █████╗ ███████║███████╗ ╚████╔╝ █████╗██████╔╝█████╗ █████╗"
|
|
||||||
echo " ██╔══╝ ██╔══██║╚════██║ ╚██╔╝ ╚════╝██╔══██╗██╔══╝ ██╔══╝"
|
|
||||||
echo " ███████╗██║ ██║███████║ ██║ ██████╔╝███████╗███████╗"
|
|
||||||
echo " ╚══════╝╚═╝ ╚═╝╚══════╝ ╚═╝ ╚═════╝ ╚══════╝╚══════╝"
|
|
||||||
echo " Hardware Audit LiveCD"
|
|
||||||
echo ""
|
|
||||||
|
|
||||||
menuentry "EASY-BEE" {
|
menuentry "EASY-BEE" {
|
||||||
linux @KERNEL_LIVE@ @APPEND_LIVE@ bee.display=kms bee.nvidia.mode=normal pci=realloc net.ifnames=0 biosdevname=0 mitigations=off transparent_hugepage=always numa_balancing=disable pcie_aspm=off intel_idle.max_cstate=1 processor.max_cstate=1 nowatchdog nosoftlockup
|
linux @KERNEL_LIVE@ @APPEND_LIVE@ bee.display=kms bee.nvidia.mode=normal pci=realloc net.ifnames=0 biosdevname=0 mitigations=off transparent_hugepage=always numa_balancing=disable pcie_aspm=off intel_idle.max_cstate=1 processor.max_cstate=1 nowatchdog nosoftlockup
|
||||||
initrd @INITRD_LIVE@
|
initrd @INITRD_LIVE@
|
||||||
|
|||||||
Binary file not shown.
|
Before Width: | Height: | Size: 70 KiB After Width: | Height: | Size: 78 KiB |
@@ -34,11 +34,11 @@ terminal-font: "Unifont Regular 16"
|
|||||||
item_font = "Unifont Regular 16"
|
item_font = "Unifont Regular 16"
|
||||||
selected_item_color= "#f5a800"
|
selected_item_color= "#f5a800"
|
||||||
selected_item_font = "Unifont Regular 16"
|
selected_item_font = "Unifont Regular 16"
|
||||||
item_height = 16
|
item_height = 20
|
||||||
item_padding = 0
|
item_padding = 2
|
||||||
item_spacing = 4
|
item_spacing = 4
|
||||||
icon_width = 0
|
icon_width = 0
|
||||||
icon_heigh = 0
|
icon_height = 0
|
||||||
item_icon_space = 0
|
item_icon_space = 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -60,35 +60,129 @@ wait_for_process_exit() {
|
|||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
kill_pattern() {
|
log_pid_details() {
|
||||||
pattern="$1"
|
pid="$1"
|
||||||
if pgrep -f "$pattern" >/dev/null 2>&1; then
|
line=$(ps -p "$pid" -o pid=,comm=,args= 2>/dev/null | sed 's/^[[:space:]]*//')
|
||||||
pgrep -af "$pattern" 2>/dev/null | while IFS= read -r line; do
|
if [ -n "$line" ]; then
|
||||||
|
log_blocker "$line"
|
||||||
|
else
|
||||||
|
log_blocker "pid $pid"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
collect_gpu_compute_pids() {
|
||||||
|
index="$1"
|
||||||
|
if ! command -v nvidia-smi >/dev/null 2>&1; then
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
nvidia-smi --id="$index" \
|
||||||
|
--query-compute-apps=pid \
|
||||||
|
--format=csv,noheader,nounits 2>/dev/null \
|
||||||
|
| sed 's/^[[:space:]]*//;s/[[:space:]]*$//' \
|
||||||
|
| grep -E '^[0-9]+$' || true
|
||||||
|
}
|
||||||
|
|
||||||
|
collect_gpu_device_pids() {
|
||||||
|
index="$1"
|
||||||
|
dev="/dev/nvidia$index"
|
||||||
|
[ -e "$dev" ] || return 0
|
||||||
|
if command -v fuser >/dev/null 2>&1; then
|
||||||
|
fuser "$dev" 2>/dev/null \
|
||||||
|
| tr ' ' '\n' \
|
||||||
|
| sed 's/[^0-9].*$//' \
|
||||||
|
| grep -E '^[0-9]+$' || true
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
collect_gpu_holder_pids() {
|
||||||
|
index="$1"
|
||||||
|
{
|
||||||
|
collect_gpu_compute_pids "$index"
|
||||||
|
collect_gpu_device_pids "$index"
|
||||||
|
} | awk 'NF' | sort -u
|
||||||
|
}
|
||||||
|
|
||||||
|
kill_pid_list() {
|
||||||
|
pids="$1"
|
||||||
|
[ -n "$pids" ] || return 0
|
||||||
|
|
||||||
|
for pid in $pids; do
|
||||||
|
log_pid_details "$pid"
|
||||||
|
done
|
||||||
|
log "terminating GPU holder PIDs: $(echo "$pids" | tr '\n' ' ' | sed 's/[[:space:]]*$//')"
|
||||||
|
for pid in $pids; do
|
||||||
|
kill -TERM "$pid" >/dev/null 2>&1 || true
|
||||||
|
done
|
||||||
|
sleep 1
|
||||||
|
for pid in $pids; do
|
||||||
|
if kill -0 "$pid" >/dev/null 2>&1; then
|
||||||
|
log "forcing GPU holder PID $pid to exit"
|
||||||
|
kill -KILL "$pid" >/dev/null 2>&1 || true
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
}
|
||||||
|
|
||||||
|
gpu_has_display_holders() {
|
||||||
|
index="$1"
|
||||||
|
holders=$(collect_gpu_device_pids "$index")
|
||||||
|
[ -n "$holders" ] || return 1
|
||||||
|
for pid in $holders; do
|
||||||
|
comm=$(ps -p "$pid" -o comm= 2>/dev/null | sed 's/^[[:space:]]*//;s/[[:space:]]*$//')
|
||||||
|
case "$comm" in
|
||||||
|
Xorg|Xwayland|X|gnome-shell)
|
||||||
|
return 0
|
||||||
|
;;
|
||||||
|
esac
|
||||||
|
done
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
stop_nv_hostengine_if_running() {
|
||||||
|
if pgrep -x nv-hostengine >/dev/null 2>&1; then
|
||||||
|
pgrep -af "^nv-hostengine$" 2>/dev/null | while IFS= read -r line; do
|
||||||
[ -n "$line" ] || continue
|
[ -n "$line" ] || continue
|
||||||
log_blocker "$line"
|
log_blocker "$line"
|
||||||
done
|
done
|
||||||
log "killing processes matching: $pattern"
|
log "stopping nv-hostengine"
|
||||||
pkill -TERM -f "$pattern" >/dev/null 2>&1 || true
|
pkill -TERM -x nv-hostengine >/dev/null 2>&1 || true
|
||||||
sleep 1
|
wait_for_process_exit nv-hostengine || pkill -KILL -x nv-hostengine >/dev/null 2>&1 || true
|
||||||
pkill -KILL -f "$pattern" >/dev/null 2>&1 || true
|
hostengine_was_active=1
|
||||||
|
return 0
|
||||||
fi
|
fi
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
stop_fabricmanager_if_active() {
|
||||||
|
if unit_exists nvidia-fabricmanager.service && stop_unit_if_active nvidia-fabricmanager.service; then
|
||||||
|
log_blocker "service nvidia-fabricmanager.service"
|
||||||
|
fabric_was_active=1
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
return 1
|
||||||
|
}
|
||||||
|
|
||||||
|
stop_display_stack_if_active() {
|
||||||
|
stopped=1
|
||||||
|
for unit in display-manager.service lightdm.service; do
|
||||||
|
if unit_exists "$unit" && stop_unit_if_active "$unit"; then
|
||||||
|
log_blocker "service $unit"
|
||||||
|
display_was_active=1
|
||||||
|
stopped=0
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
return "$stopped"
|
||||||
|
}
|
||||||
|
|
||||||
|
try_gpu_reset() {
|
||||||
|
index="$1"
|
||||||
|
log "resetting GPU $index"
|
||||||
|
nvidia-smi -r -i "$index"
|
||||||
}
|
}
|
||||||
|
|
||||||
drain_gpu_clients() {
|
drain_gpu_clients() {
|
||||||
display_was_active=0
|
display_was_active=0
|
||||||
fabric_was_active=0
|
fabric_was_active=0
|
||||||
|
hostengine_was_active=0
|
||||||
for unit in display-manager.service lightdm.service; do
|
|
||||||
if unit_exists "$unit" && stop_unit_if_active "$unit"; then
|
|
||||||
log_blocker "service $unit"
|
|
||||||
display_was_active=1
|
|
||||||
fi
|
|
||||||
done
|
|
||||||
|
|
||||||
if unit_exists nvidia-fabricmanager.service && stop_unit_if_active nvidia-fabricmanager.service; then
|
|
||||||
log_blocker "service nvidia-fabricmanager.service"
|
|
||||||
fabric_was_active=1
|
|
||||||
fi
|
|
||||||
|
|
||||||
if pgrep -x nv-hostengine >/dev/null 2>&1; then
|
if pgrep -x nv-hostengine >/dev/null 2>&1; then
|
||||||
pgrep -af "^nv-hostengine$" 2>/dev/null | while IFS= read -r line; do
|
pgrep -af "^nv-hostengine$" 2>/dev/null | while IFS= read -r line; do
|
||||||
@@ -98,21 +192,25 @@ drain_gpu_clients() {
|
|||||||
log "stopping nv-hostengine"
|
log "stopping nv-hostengine"
|
||||||
pkill -TERM -x nv-hostengine >/dev/null 2>&1 || true
|
pkill -TERM -x nv-hostengine >/dev/null 2>&1 || true
|
||||||
wait_for_process_exit nv-hostengine || pkill -KILL -x nv-hostengine >/dev/null 2>&1 || true
|
wait_for_process_exit nv-hostengine || pkill -KILL -x nv-hostengine >/dev/null 2>&1 || true
|
||||||
|
hostengine_was_active=1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
for pattern in \
|
if unit_exists nvidia-fabricmanager.service && stop_unit_if_active nvidia-fabricmanager.service; then
|
||||||
"nvidia-smi" \
|
log_blocker "service nvidia-fabricmanager.service"
|
||||||
"dcgmi" \
|
fabric_was_active=1
|
||||||
"nvvs" \
|
fi
|
||||||
"dcgmproftester" \
|
|
||||||
"all_reduce_perf" \
|
for unit in display-manager.service lightdm.service; do
|
||||||
"nvtop" \
|
if unit_exists "$unit" && stop_unit_if_active "$unit"; then
|
||||||
"bee-gpu-burn" \
|
log_blocker "service $unit"
|
||||||
"bee-john-gpu-stress" \
|
display_was_active=1
|
||||||
"bee-nccl-gpu-stress" \
|
fi
|
||||||
"Xorg" \
|
done
|
||||||
"Xwayland"; do
|
|
||||||
kill_pattern "$pattern"
|
for dev in /dev/nvidia[0-9]*; do
|
||||||
|
[ -e "$dev" ] || continue
|
||||||
|
holders=$(collect_gpu_device_pids "${dev#/dev/nvidia}")
|
||||||
|
kill_pid_list "$holders"
|
||||||
done
|
done
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -125,7 +223,7 @@ restore_gpu_clients() {
|
|||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if command -v nv-hostengine >/dev/null 2>&1 && ! pgrep -x nv-hostengine >/dev/null 2>&1; then
|
if [ "${hostengine_was_active:-0}" = "1" ] && command -v nv-hostengine >/dev/null 2>&1 && ! pgrep -x nv-hostengine >/dev/null 2>&1; then
|
||||||
log "starting nv-hostengine"
|
log "starting nv-hostengine"
|
||||||
nv-hostengine
|
nv-hostengine
|
||||||
fi
|
fi
|
||||||
@@ -153,10 +251,60 @@ restart_drivers() {
|
|||||||
|
|
||||||
reset_gpu() {
|
reset_gpu() {
|
||||||
index="$1"
|
index="$1"
|
||||||
drain_gpu_clients
|
display_was_active=0
|
||||||
log "resetting GPU $index"
|
fabric_was_active=0
|
||||||
nvidia-smi -r -i "$index"
|
hostengine_was_active=0
|
||||||
|
|
||||||
|
holders=$(collect_gpu_holder_pids "$index")
|
||||||
|
if [ -n "$holders" ]; then
|
||||||
|
kill_pid_list "$holders"
|
||||||
|
fi
|
||||||
|
if try_gpu_reset "$index"; then
|
||||||
restore_gpu_clients
|
restore_gpu_clients
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
stop_nv_hostengine_if_running || true
|
||||||
|
holders=$(collect_gpu_holder_pids "$index")
|
||||||
|
if [ -n "$holders" ]; then
|
||||||
|
kill_pid_list "$holders"
|
||||||
|
fi
|
||||||
|
if try_gpu_reset "$index"; then
|
||||||
|
restore_gpu_clients
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
stop_fabricmanager_if_active || true
|
||||||
|
holders=$(collect_gpu_holder_pids "$index")
|
||||||
|
if [ -n "$holders" ]; then
|
||||||
|
kill_pid_list "$holders"
|
||||||
|
fi
|
||||||
|
if try_gpu_reset "$index"; then
|
||||||
|
restore_gpu_clients
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
if gpu_has_display_holders "$index"; then
|
||||||
|
stop_display_stack_if_active || true
|
||||||
|
holders=$(collect_gpu_holder_pids "$index")
|
||||||
|
if [ -n "$holders" ]; then
|
||||||
|
kill_pid_list "$holders"
|
||||||
|
fi
|
||||||
|
if try_gpu_reset "$index"; then
|
||||||
|
restore_gpu_clients
|
||||||
|
return 0
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
holders=$(collect_gpu_holder_pids "$index")
|
||||||
|
if [ -n "$holders" ]; then
|
||||||
|
log "GPU $index still has holders after targeted drain"
|
||||||
|
kill_pid_list "$holders"
|
||||||
|
fi
|
||||||
|
try_gpu_reset "$index"
|
||||||
|
rc=$?
|
||||||
|
restore_gpu_clients
|
||||||
|
return "$rc"
|
||||||
}
|
}
|
||||||
|
|
||||||
cmd="${1:-}"
|
cmd="${1:-}"
|
||||||
|
|||||||
@@ -47,6 +47,13 @@ echo "==> Сборка бинарника..."
|
|||||||
)
|
)
|
||||||
echo " OK: $(ls -lh "${LOCAL_BIN}" | awk '{print $5, $9}')"
|
echo " OK: $(ls -lh "${LOCAL_BIN}" | awk '{print $5, $9}')"
|
||||||
|
|
||||||
|
LOCAL_SHA="$(shasum -a 256 "${LOCAL_BIN}" | awk '{print $1}')"
|
||||||
|
REMOTE_SHA="$("${SSH_CMD[@]}" "$REMOTE" "if [ -f '${REMOTE_BIN}' ] && command -v sha256sum >/dev/null 2>&1; then sha256sum '${REMOTE_BIN}' | awk '{print \\$1}'; fi" 2>/dev/null || true)"
|
||||||
|
if [[ -n "${REMOTE_SHA}" && "${LOCAL_SHA}" == "${REMOTE_SHA}" ]]; then
|
||||||
|
echo "==> Бинарник не изменился (${LOCAL_SHA}); копирование и перезапуск сервисов пропущены."
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
# --- Deploy ---
|
# --- Deploy ---
|
||||||
echo "==> Копирование на ${REMOTE}..."
|
echo "==> Копирование на ${REMOTE}..."
|
||||||
"${SCP_CMD[@]}" "${LOCAL_BIN}" "${REMOTE}:/tmp/bee-new"
|
"${SCP_CMD[@]}" "${LOCAL_BIN}" "${REMOTE}:/tmp/bee-new"
|
||||||
|
|||||||
Reference in New Issue
Block a user