286 lines
8.2 KiB
Go
286 lines
8.2 KiB
Go
// Copyright (c) HashiCorp, Inc.
|
|
// SPDX-License-Identifier: MPL-2.0
|
|
|
|
package raft
|
|
|
|
import (
|
|
"fmt"
|
|
"io"
|
|
"time"
|
|
|
|
hclog "github.com/hashicorp/go-hclog"
|
|
"github.com/hashicorp/go-metrics/compat"
|
|
)
|
|
|
|
// FSM is implemented by clients to make use of the replicated log.
|
|
type FSM interface {
|
|
// Apply is called once a log entry is committed by a majority of the cluster.
|
|
//
|
|
// Apply should apply the log to the FSM. Apply must be deterministic and
|
|
// produce the same result on all peers in the cluster.
|
|
//
|
|
// The returned value is returned to the client as the ApplyFuture.Response.
|
|
Apply(*Log) interface{}
|
|
|
|
// Snapshot returns an FSMSnapshot used to: support log compaction, to
|
|
// restore the FSM to a previous state, or to bring out-of-date followers up
|
|
// to a recent log index.
|
|
//
|
|
// The Snapshot implementation should return quickly, because Apply can not
|
|
// be called while Snapshot is running. Generally this means Snapshot should
|
|
// only capture a pointer to the state, and any expensive IO should happen
|
|
// as part of FSMSnapshot.Persist.
|
|
//
|
|
// Apply and Snapshot are always called from the same thread, but Apply will
|
|
// be called concurrently with FSMSnapshot.Persist. This means the FSM should
|
|
// be implemented to allow for concurrent updates while a snapshot is happening.
|
|
//
|
|
// Clients of this library should make no assumptions about whether a returned
|
|
// Snapshot() will actually be stored by Raft. In fact it's quite possible that
|
|
// any Snapshot returned by this call will be discarded, and that
|
|
// FSMSnapshot.Persist will never be called. Raft will always call
|
|
// FSMSnapshot.Release however.
|
|
Snapshot() (FSMSnapshot, error)
|
|
|
|
// Restore is used to restore an FSM from a snapshot. It is not called
|
|
// concurrently with any other command. The FSM must discard all previous
|
|
// state before restoring the snapshot.
|
|
Restore(snapshot io.ReadCloser) error
|
|
}
|
|
|
|
// BatchingFSM extends the FSM interface to add an ApplyBatch function. This can
|
|
// optionally be implemented by clients to enable multiple logs to be applied to
|
|
// the FSM in batches. Up to MaxAppendEntries could be sent in a batch.
|
|
type BatchingFSM interface {
|
|
// ApplyBatch is invoked once a batch of log entries has been committed and
|
|
// are ready to be applied to the FSM. ApplyBatch will take in an array of
|
|
// log entries. These log entries will be in the order they were committed,
|
|
// will not have gaps, and could be of a few log types. Clients should check
|
|
// the log type prior to attempting to decode the data attached. Presently
|
|
// the LogCommand and LogConfiguration types will be sent.
|
|
//
|
|
// The returned slice must be the same length as the input and each response
|
|
// should correlate to the log at the same index of the input. The returned
|
|
// values will be made available in the ApplyFuture returned by Raft.Apply
|
|
// method if that method was called on the same Raft node as the FSM.
|
|
ApplyBatch([]*Log) []interface{}
|
|
|
|
FSM
|
|
}
|
|
|
|
// FSMSnapshot is returned by an FSM in response to a Snapshot
|
|
// It must be safe to invoke FSMSnapshot methods with concurrent
|
|
// calls to Apply.
|
|
type FSMSnapshot interface {
|
|
// Persist should dump all necessary state to the WriteCloser 'sink',
|
|
// and call sink.Close() when finished or call sink.Cancel() on error.
|
|
Persist(sink SnapshotSink) error
|
|
|
|
// Release is invoked when we are finished with the snapshot.
|
|
Release()
|
|
}
|
|
|
|
// runFSM is a long running goroutine responsible for applying logs
|
|
// to the FSM. This is done async of other logs since we don't want
|
|
// the FSM to block our internal operations.
|
|
func (r *Raft) runFSM() {
|
|
var lastIndex, lastTerm uint64
|
|
|
|
batchingFSM, batchingEnabled := r.fsm.(BatchingFSM)
|
|
configStore, configStoreEnabled := r.fsm.(ConfigurationStore)
|
|
|
|
applySingle := func(req *commitTuple) {
|
|
// Apply the log if a command or config change
|
|
var resp interface{}
|
|
// Make sure we send a response
|
|
defer func() {
|
|
// Invoke the future if given
|
|
if req.future != nil {
|
|
req.future.response = resp
|
|
req.future.respond(nil)
|
|
}
|
|
}()
|
|
|
|
switch req.log.Type {
|
|
case LogCommand:
|
|
start := time.Now()
|
|
resp = r.fsm.Apply(req.log)
|
|
metrics.MeasureSince([]string{"raft", "fsm", "apply"}, start)
|
|
|
|
case LogConfiguration:
|
|
if !configStoreEnabled {
|
|
// Return early to avoid incrementing the index and term for
|
|
// an unimplemented operation.
|
|
return
|
|
}
|
|
|
|
start := time.Now()
|
|
configStore.StoreConfiguration(req.log.Index, DecodeConfiguration(req.log.Data))
|
|
metrics.MeasureSince([]string{"raft", "fsm", "store_config"}, start)
|
|
}
|
|
|
|
// Update the indexes
|
|
lastIndex = req.log.Index
|
|
lastTerm = req.log.Term
|
|
}
|
|
|
|
applyBatch := func(reqs []*commitTuple) {
|
|
if !batchingEnabled {
|
|
for _, ct := range reqs {
|
|
applySingle(ct)
|
|
}
|
|
return
|
|
}
|
|
|
|
// Only send LogCommand and LogConfiguration log types. LogBarrier types
|
|
// will not be sent to the FSM.
|
|
shouldSend := func(l *Log) bool {
|
|
switch l.Type {
|
|
case LogCommand, LogConfiguration:
|
|
return true
|
|
}
|
|
return false
|
|
}
|
|
|
|
var lastBatchIndex, lastBatchTerm uint64
|
|
sendLogs := make([]*Log, 0, len(reqs))
|
|
for _, req := range reqs {
|
|
if shouldSend(req.log) {
|
|
sendLogs = append(sendLogs, req.log)
|
|
}
|
|
lastBatchIndex = req.log.Index
|
|
lastBatchTerm = req.log.Term
|
|
}
|
|
|
|
var responses []interface{}
|
|
if len(sendLogs) > 0 {
|
|
start := time.Now()
|
|
responses = batchingFSM.ApplyBatch(sendLogs)
|
|
metrics.MeasureSince([]string{"raft", "fsm", "applyBatch"}, start)
|
|
metrics.AddSample([]string{"raft", "fsm", "applyBatchNum"}, float32(len(reqs)))
|
|
|
|
// Ensure we get the expected responses
|
|
if len(sendLogs) != len(responses) {
|
|
panic("invalid number of responses")
|
|
}
|
|
}
|
|
|
|
// Update the indexes
|
|
lastIndex = lastBatchIndex
|
|
lastTerm = lastBatchTerm
|
|
|
|
var i int
|
|
for _, req := range reqs {
|
|
var resp interface{}
|
|
// If the log was sent to the FSM, retrieve the response.
|
|
if shouldSend(req.log) {
|
|
resp = responses[i]
|
|
i++
|
|
}
|
|
|
|
if req.future != nil {
|
|
req.future.response = resp
|
|
req.future.respond(nil)
|
|
}
|
|
}
|
|
}
|
|
|
|
restore := func(req *restoreFuture) {
|
|
// Open the snapshot
|
|
meta, source, err := r.snapshots.Open(req.ID)
|
|
if err != nil {
|
|
req.respond(fmt.Errorf("failed to open snapshot %v: %v", req.ID, err))
|
|
return
|
|
}
|
|
defer source.Close()
|
|
|
|
snapLogger := r.logger.With(
|
|
"id", req.ID,
|
|
"last-index", meta.Index,
|
|
"last-term", meta.Term,
|
|
"size-in-bytes", meta.Size,
|
|
)
|
|
|
|
// Attempt to restore
|
|
if err := fsmRestoreAndMeasure(snapLogger, r.fsm, source, meta.Size); err != nil {
|
|
req.respond(fmt.Errorf("failed to restore snapshot %v: %v", req.ID, err))
|
|
return
|
|
}
|
|
|
|
// Update the last index and term
|
|
lastIndex = meta.Index
|
|
lastTerm = meta.Term
|
|
req.respond(nil)
|
|
}
|
|
|
|
snapshot := func(req *reqSnapshotFuture) {
|
|
// Is there something to snapshot?
|
|
if lastIndex == 0 {
|
|
req.respond(ErrNothingNewToSnapshot)
|
|
return
|
|
}
|
|
|
|
// Start a snapshot
|
|
start := time.Now()
|
|
snap, err := r.fsm.Snapshot()
|
|
metrics.MeasureSince([]string{"raft", "fsm", "snapshot"}, start)
|
|
|
|
// Respond to the request
|
|
req.index = lastIndex
|
|
req.term = lastTerm
|
|
req.snapshot = snap
|
|
req.respond(err)
|
|
}
|
|
|
|
saturation := newSaturationMetric([]string{"raft", "thread", "fsm", "saturation"}, 1*time.Second)
|
|
|
|
for {
|
|
saturation.sleeping()
|
|
|
|
select {
|
|
case ptr := <-r.fsmMutateCh:
|
|
saturation.working()
|
|
|
|
switch req := ptr.(type) {
|
|
case []*commitTuple:
|
|
applyBatch(req)
|
|
|
|
case *restoreFuture:
|
|
restore(req)
|
|
|
|
default:
|
|
panic(fmt.Errorf("bad type passed to fsmMutateCh: %#v", ptr))
|
|
}
|
|
|
|
case req := <-r.fsmSnapshotCh:
|
|
saturation.working()
|
|
|
|
snapshot(req)
|
|
|
|
case <-r.shutdownCh:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// fsmRestoreAndMeasure wraps the Restore call on an FSM to consistently measure
|
|
// and report timing metrics. The caller is still responsible for calling Close
|
|
// on the source in all cases.
|
|
func fsmRestoreAndMeasure(logger hclog.Logger, fsm FSM, source io.ReadCloser, snapshotSize int64) error {
|
|
start := time.Now()
|
|
|
|
crc := newCountingReadCloser(source)
|
|
|
|
monitor := startSnapshotRestoreMonitor(logger, crc, snapshotSize, false)
|
|
defer monitor.StopAndWait()
|
|
|
|
if err := fsm.Restore(crc); err != nil {
|
|
return err
|
|
}
|
|
metrics.MeasureSince([]string{"raft", "fsm", "restore"}, start)
|
|
metrics.SetGauge([]string{"raft", "fsm", "lastRestoreDuration"},
|
|
float32(time.Since(start).Milliseconds()))
|
|
|
|
return nil
|
|
}
|