github.com/smithx10/nomad@v0.9.1-rc1/e2e/clientstate/clientstate.go (about) 1 package clientstate 2 3 import ( 4 "bytes" 5 "fmt" 6 "io/ioutil" 7 "math/rand" 8 "net/http" 9 "os" 10 "os/exec" 11 "path/filepath" 12 "strconv" 13 "syscall" 14 "time" 15 16 "github.com/hashicorp/nomad/api" 17 "github.com/hashicorp/nomad/client/state" 18 "github.com/hashicorp/nomad/e2e/e2eutil" 19 "github.com/hashicorp/nomad/e2e/execagent" 20 "github.com/hashicorp/nomad/e2e/framework" 21 "github.com/hashicorp/nomad/helper/discover" 22 "github.com/hashicorp/nomad/helper/testlog" 23 "github.com/hashicorp/nomad/helper/uuid" 24 "github.com/hashicorp/nomad/testutil" 25 ) 26 27 func init() { 28 framework.AddSuites(&framework.TestSuite{ 29 Component: "clientstate", 30 CanRunLocal: true, 31 Cases: []framework.TestCase{ 32 &ClientStateTC{}, 33 }, 34 }) 35 } 36 37 type ClientStateTC struct { 38 framework.TC 39 40 // bin is the path to Nomad binary 41 bin string 42 } 43 44 func (tc *ClientStateTC) BeforeAll(f *framework.F) { 45 if os.Getenv("NOMAD_TEST_STATE") == "" { 46 f.T().Skip("Skipping very slow state corruption test unless NOMAD_TEST_STATE=1") 47 } 48 49 bin, err := discover.NomadExecutable() 50 f.NoError(err) 51 tc.bin = bin 52 } 53 54 func getPID(client *api.Client, alloc *api.Allocation, path string) (int, error) { 55 allocfs := client.AllocFS() 56 r, err := allocfs.Cat(alloc, path, nil) 57 if err != nil { 58 return 0, err 59 } 60 defer r.Close() 61 62 out, err := ioutil.ReadAll(r) 63 if err != nil { 64 return 0, err 65 } 66 67 lines := bytes.SplitN(out, []byte{'\n'}, 2) 68 if len(lines) != 2 || len(lines[1]) > 0 { 69 return 0, fmt.Errorf("expected 1 line not %q", string(out)) 70 } 71 72 // Capture pid 73 pid, err := strconv.Atoi(string(lines[0])) 74 if err != nil { 75 return 0, err 76 } 77 78 return pid, nil 79 } 80 81 // TestClientState_Kill force kills Nomad agents and restarts them in a tight 82 // loop to assert Nomad is crash safe. 83 func (tc *ClientStateTC) TestClientState_Kill(f *framework.F) { 84 t := f.T() 85 t.Parallel() 86 87 serverOut := testlog.NewPrefixWriter(t, "SERVER: ") 88 clientOut := testlog.NewPrefixWriter(t, "CLIENT: ") 89 serverAgent, clientAgent, err := execagent.NewClientServerPair(tc.bin, serverOut, clientOut) 90 f.NoError(err) 91 92 f.NoError(serverAgent.Start()) 93 defer serverAgent.Destroy() 94 f.NoError(clientAgent.Start()) 95 defer clientAgent.Destroy() 96 97 // Get a client for the server agent to use even while the client is 98 // down. 99 client, err := serverAgent.Client() 100 f.NoError(err) 101 102 jobID := "sleeper-" + uuid.Generate()[:8] 103 allocs := e2eutil.RegisterAndWaitForAllocs(t, client, "clientstate/sleeper.nomad", jobID) 104 f.Len(allocs, 1) 105 106 alloc, _, err := client.Allocations().Info(allocs[0].ID, nil) 107 f.NoError(err) 108 109 defer func() { 110 if _, _, err := client.Jobs().Deregister(jobID, false, nil); err != nil { 111 t.Logf("error stopping job: %v", err) 112 } 113 114 testutil.WaitForResult(func() (bool, error) { 115 sum, _, err := client.Jobs().Summary(jobID, nil) 116 if err != nil { 117 return false, err 118 } 119 if r := sum.Summary["sleeper"].Running; r > 0 { 120 return false, fmt.Errorf("still running: %d", r) 121 } 122 return true, nil 123 }, func(err error) { 124 f.NoError(err) 125 }) 126 127 //XXX Must use client agent for gc'ing allocs? 128 clientAPI, err := clientAgent.Client() 129 f.NoError(err) 130 if err := clientAPI.Allocations().GC(alloc, nil); err != nil { 131 t.Logf("error garbage collecting alloc: %v", err) 132 } 133 134 if err := client.System().GarbageCollect(); err != nil { 135 t.Logf("error doing full gc: %v", err) 136 } 137 138 //HACK to wait until things have GC'd 139 time.Sleep(time.Second) 140 }() 141 142 assertHealthy := func() { 143 t.Helper() 144 testutil.WaitForResult(func() (bool, error) { 145 alloc, _, err = client.Allocations().Info(alloc.ID, nil) 146 f.NoError(err) // should never error 147 148 if len(alloc.TaskStates) == 0 { 149 return false, fmt.Errorf("waiting for tasks to start") 150 } 151 152 if s := alloc.TaskStates["sleeper"].State; s != "running" { 153 return false, fmt.Errorf("task should be running: %q", s) 154 } 155 156 // Restarts should never happen 157 f.Zero(alloc.TaskStates["sleeper"].Restarts) 158 return true, nil 159 }, func(err error) { 160 f.NoError(err) 161 }) 162 } 163 assertHealthy() 164 165 // Find pid 166 pid := 0 167 testutil.WaitForResult(func() (bool, error) { 168 pid, err = getPID(client, alloc, "sleeper/pid") 169 return pid > 0, err 170 }, func(err error) { 171 f.NoError(err) 172 }) 173 174 // Kill and restart a few times 175 tries := 10 176 for i := 0; i < tries; i++ { 177 t.Logf("TEST RUN %d/%d", i+1, tries) 178 179 // Kill -9 the Agent 180 agentPid := clientAgent.Cmd.Process.Pid 181 f.NoError(clientAgent.Cmd.Process.Signal(os.Kill)) 182 183 state, err := clientAgent.Cmd.Process.Wait() 184 f.NoError(err) 185 f.False(state.Exited()) // kill signal != exited 186 f.False(state.Success()) 187 188 // Assert sleeper is still running 189 f.NoError(syscall.Kill(pid, 0)) 190 assertHealthy() 191 192 // Should not be able to reach its filesystem 193 _, err = getPID(client, alloc, "sleeper/pid") 194 f.Error(err) 195 196 // Restart the agent (have to create a new Cmd) 197 clientAgent.Cmd = exec.Command(clientAgent.BinPath, "agent", 198 "-config", clientAgent.ConfFile, 199 "-data-dir", clientAgent.DataDir, 200 "-servers", fmt.Sprintf("127.0.0.1:%d", serverAgent.Vars.RPC), 201 ) 202 clientAgent.Cmd.Stdout = clientOut 203 clientAgent.Cmd.Stderr = clientOut 204 f.NoError(clientAgent.Start()) 205 206 // Assert a new process did start 207 f.NotEqual(clientAgent.Cmd.Process.Pid, agentPid) 208 209 // Retrieving the pid should work once it restarts 210 testutil.WaitForResult(func() (bool, error) { 211 newPid, err := getPID(client, alloc, "sleeper/pid") 212 return newPid == pid, err 213 }, func(err error) { 214 f.NoError(err) 215 }) 216 217 // Alloc should still be running 218 assertHealthy() 219 } 220 } 221 222 // TestClientState_KillDuringRestart force kills Nomad agents and restarts them 223 // in a tight loop to assert Nomad is crash safe while a task is restarting. 224 func (tc *ClientStateTC) TestClientState_KillDuringRestart(f *framework.F) { 225 t := f.T() 226 t.Parallel() 227 228 serverOut := testlog.NewPrefixWriter(t, "SERVER: ") 229 clientOut := testlog.NewPrefixWriter(t, "CLIENT: ") 230 serverAgent, clientAgent, err := execagent.NewClientServerPair(tc.bin, serverOut, clientOut) 231 f.NoError(err) 232 233 f.NoError(serverAgent.Start()) 234 defer serverAgent.Destroy() 235 236 f.NoError(clientAgent.Start()) 237 defer clientAgent.Destroy() 238 239 // Get a client for the server agent to use even while the client is 240 // down. 241 client, err := serverAgent.Client() 242 f.NoError(err) 243 244 jobID := "restarter-" + uuid.Generate()[:8] 245 allocs := e2eutil.RegisterAndWaitForAllocs(t, client, "clientstate/restarter.nomad", jobID) 246 f.Len(allocs, 1) 247 248 alloc, _, err := client.Allocations().Info(allocs[0].ID, nil) 249 f.NoError(err) 250 251 defer func() { 252 //FIXME(schmichael): this cleanup is insufficient, but I can't 253 // figure out how to fix it 254 client.Jobs().Deregister(jobID, false, nil) 255 client.System().GarbageCollect() 256 time.Sleep(time.Second) 257 }() 258 259 var restarts uint64 260 testutil.WaitForResult(func() (bool, error) { 261 alloc, _, err = client.Allocations().Info(alloc.ID, nil) 262 f.NoError(err) // should never error 263 264 if len(alloc.TaskStates) == 0 { 265 return false, fmt.Errorf("waiting for tasks to start") 266 } 267 268 n := alloc.TaskStates["restarter"].Restarts 269 if n < restarts { 270 // Restarts should never decrease; immediately fail 271 f.Failf("restarts decreased", "%d < %d", n, restarts) 272 } 273 274 // Capture current restarts 275 restarts = n 276 return true, nil 277 }, func(err error) { 278 f.NoError(err) 279 }) 280 281 dice := rand.New(rand.NewSource(time.Now().UnixNano())) 282 283 // Kill and restart agent a few times 284 i := 0 285 for deadline := time.Now().Add(5 * time.Minute); time.Now().Before(deadline); { 286 i++ 287 sleep := time.Duration(1500+dice.Int63n(6000)) * time.Millisecond 288 t.Logf("[TEST] ===> Run %d (pid: %d sleeping for %v; last restarts: %d)", i, clientAgent.Cmd.Process.Pid, sleep, restarts) 289 290 time.Sleep(sleep) 291 292 // Ensure restarts are progressing 293 alloc, _, err = client.Allocations().Info(alloc.ID, nil) 294 f.NoError(err) // should never error 295 n := alloc.TaskStates["restarter"].Restarts 296 if n < restarts { 297 // Restarts should never decrease; immediately fail 298 f.Failf("restarts decreased", "%d < %d", n, restarts) 299 } 300 if i > 5 && n == 0 { 301 // At least one restart should have happened by now 302 f.Failf("no restarts", "expected at least 1 restart after %d tries", i) 303 } 304 restarts = n 305 306 // Kill -9 Agent 307 agentPid := clientAgent.Cmd.Process.Pid 308 f.NoError(clientAgent.Cmd.Process.Signal(os.Kill)) 309 t.Logf("[TEST] ===> Killed %d", agentPid) 310 311 state, err := clientAgent.Cmd.Process.Wait() 312 f.NoError(err) 313 f.False(state.Exited()) // kill signal != exited 314 f.False(state.Success()) 315 316 // Restart the agent (have to create a new Cmd) 317 clientAgent.Cmd = exec.Command(clientAgent.BinPath, "agent", 318 "-config", clientAgent.ConfFile, 319 "-data-dir", clientAgent.DataDir, 320 "-servers", fmt.Sprintf("127.0.0.1:%d", serverAgent.Vars.RPC), 321 ) 322 clientAgent.Cmd.Stdout = clientOut 323 clientAgent.Cmd.Stderr = clientOut 324 f.NoError(clientAgent.Start()) 325 326 // Assert a new process did start 327 f.NotEqual(clientAgent.Cmd.Process.Pid, agentPid) 328 clientUrl := fmt.Sprintf("http://127.0.0.1:%d/v1/client/stats", clientAgent.Vars.HTTP) 329 testutil.WaitForResult(func() (bool, error) { 330 resp, err := http.Get(clientUrl) 331 if err != nil { 332 return false, err 333 } 334 resp.Body.Close() 335 return resp.StatusCode == 200, fmt.Errorf("%d != 200", resp.StatusCode) 336 }, func(err error) { 337 f.NoError(err) 338 }) 339 } 340 341 t.Logf("[TEST] ===> Final restarts: %d", restarts) 342 } 343 344 // TestClientState_Corrupt removes task state from the client's state db to 345 // assert it recovers. 346 func (tc *ClientStateTC) TestClientState_Corrupt(f *framework.F) { 347 t := f.T() 348 t.Parallel() 349 350 serverOut := testlog.NewPrefixWriter(t, "SERVER: ") 351 clientOut := testlog.NewPrefixWriter(t, "CLIENT: ") 352 serverAgent, clientAgent, err := execagent.NewClientServerPair(tc.bin, serverOut, clientOut) 353 f.NoError(err) 354 355 f.NoError(serverAgent.Start()) 356 defer serverAgent.Destroy() 357 f.NoError(clientAgent.Start()) 358 defer clientAgent.Destroy() 359 360 // Get a client for the server agent to use even while the client is 361 // down. 362 client, err := serverAgent.Client() 363 f.NoError(err) 364 365 jobID := "sleeper-" + uuid.Generate()[:8] 366 allocs := e2eutil.RegisterAndWaitForAllocs(t, client, "clientstate/sleeper.nomad", jobID) 367 f.Len(allocs, 1) 368 369 alloc, _, err := client.Allocations().Info(allocs[0].ID, nil) 370 f.NoError(err) 371 372 defer func() { 373 //FIXME(schmichael): this cleanup is insufficient, but I can't 374 // figure out how to fix it 375 client.Jobs().Deregister(jobID, false, nil) 376 client.System().GarbageCollect() 377 time.Sleep(time.Second) 378 }() 379 380 assertHealthy := func() { 381 t.Helper() 382 testutil.WaitForResult(func() (bool, error) { 383 alloc, _, err = client.Allocations().Info(alloc.ID, nil) 384 f.NoError(err) // should never error 385 386 if len(alloc.TaskStates) == 0 { 387 return false, fmt.Errorf("waiting for tasks to start") 388 } 389 390 if s := alloc.TaskStates["sleeper"].State; s != "running" { 391 return false, fmt.Errorf("task should be running: %q", s) 392 } 393 394 // Restarts should never happen 395 f.Zero(alloc.TaskStates["sleeper"].Restarts) 396 return true, nil 397 }, func(err error) { 398 f.NoError(err) 399 }) 400 } 401 assertHealthy() 402 403 // Find pid 404 pid := 0 405 testutil.WaitForResult(func() (bool, error) { 406 pid, err = getPID(client, alloc, "sleeper/pid") 407 return pid > 0, err 408 }, func(err error) { 409 f.NoError(err) 410 }) 411 412 // Kill and corrupt the state 413 agentPid := clientAgent.Cmd.Process.Pid 414 f.NoError(clientAgent.Cmd.Process.Signal(os.Interrupt)) 415 416 procState, err := clientAgent.Cmd.Process.Wait() 417 f.NoError(err) 418 f.True(procState.Exited()) 419 420 // Assert sleeper is still running 421 f.NoError(syscall.Kill(pid, 0)) 422 assertHealthy() 423 424 // Remove task bucket from client state 425 db, err := state.NewBoltStateDB(testlog.HCLogger(t), filepath.Join(clientAgent.DataDir, "client")) 426 f.NoError(err) 427 428 f.NoError(db.DeleteTaskBucket(alloc.ID, "sleeper")) 429 f.NoError(db.Close()) 430 431 // Restart the agent (have to create a new Cmd) 432 clientAgent.Cmd = exec.Command(clientAgent.BinPath, "agent", 433 "-config", clientAgent.ConfFile, 434 "-data-dir", clientAgent.DataDir, 435 "-servers", fmt.Sprintf("127.0.0.1:%d", serverAgent.Vars.RPC), 436 ) 437 clientAgent.Cmd.Stdout = clientOut 438 clientAgent.Cmd.Stderr = clientOut 439 f.NoError(clientAgent.Start()) 440 441 // Assert a new process did start 442 f.NotEqual(clientAgent.Cmd.Process.Pid, agentPid) 443 444 // Retrieving the pid should work once it restarts. 445 // Critically there are now 2 pids because the client task state was 446 // lost Nomad started a new copy. 447 testutil.WaitForResult(func() (bool, error) { 448 allocfs := client.AllocFS() 449 r, err := allocfs.Cat(alloc, "sleeper/pid", nil) 450 if err != nil { 451 return false, err 452 } 453 defer r.Close() 454 455 out, err := ioutil.ReadAll(r) 456 if err != nil { 457 return false, err 458 } 459 460 lines := bytes.SplitN(out, []byte{'\n'}, 3) 461 if len(lines) != 3 || len(lines[2]) > 0 { 462 return false, fmt.Errorf("expected 2 lines not %v", lines) 463 } 464 465 return true, nil 466 }, func(err error) { 467 f.NoError(err) 468 }) 469 470 // Alloc should still be running 471 assertHealthy() 472 }