mirror of
https://github.com/Oxalide/vsphere-influxdb-go.git
synced 2023-10-10 13:36:51 +02:00
138 lines
3.5 KiB
Go
138 lines
3.5 KiB
Go
// Package retention provides the retention policy enforcement service.
|
|
package retention // import "github.com/influxdata/influxdb/services/retention"
|
|
|
|
import (
|
|
"fmt"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/influxdata/influxdb/services/meta"
|
|
"github.com/uber-go/zap"
|
|
)
|
|
|
|
// Service represents the retention policy enforcement service.
|
|
type Service struct {
|
|
MetaClient interface {
|
|
Databases() []meta.DatabaseInfo
|
|
DeleteShardGroup(database, policy string, id uint64) error
|
|
PruneShardGroups() error
|
|
}
|
|
TSDBStore interface {
|
|
ShardIDs() []uint64
|
|
DeleteShard(shardID uint64) error
|
|
}
|
|
|
|
checkInterval time.Duration
|
|
wg sync.WaitGroup
|
|
done chan struct{}
|
|
|
|
logger zap.Logger
|
|
}
|
|
|
|
// NewService returns a configured retention policy enforcement service.
|
|
func NewService(c Config) *Service {
|
|
return &Service{
|
|
checkInterval: time.Duration(c.CheckInterval),
|
|
done: make(chan struct{}),
|
|
logger: zap.New(zap.NullEncoder()),
|
|
}
|
|
}
|
|
|
|
// Open starts retention policy enforcement.
|
|
func (s *Service) Open() error {
|
|
s.logger.Info(fmt.Sprint("Starting retention policy enforcement service with check interval of ", s.checkInterval))
|
|
s.wg.Add(2)
|
|
go s.deleteShardGroups()
|
|
go s.deleteShards()
|
|
return nil
|
|
}
|
|
|
|
// Close stops retention policy enforcement.
|
|
func (s *Service) Close() error {
|
|
s.logger.Info("retention policy enforcement terminating")
|
|
close(s.done)
|
|
s.wg.Wait()
|
|
return nil
|
|
}
|
|
|
|
// WithLogger sets the logger on the service.
|
|
func (s *Service) WithLogger(log zap.Logger) {
|
|
s.logger = log.With(zap.String("service", "retention"))
|
|
}
|
|
|
|
func (s *Service) deleteShardGroups() {
|
|
defer s.wg.Done()
|
|
|
|
ticker := time.NewTicker(s.checkInterval)
|
|
defer ticker.Stop()
|
|
for {
|
|
select {
|
|
case <-s.done:
|
|
return
|
|
|
|
case <-ticker.C:
|
|
dbs := s.MetaClient.Databases()
|
|
for _, d := range dbs {
|
|
for _, r := range d.RetentionPolicies {
|
|
for _, g := range r.ExpiredShardGroups(time.Now().UTC()) {
|
|
if err := s.MetaClient.DeleteShardGroup(d.Name, r.Name, g.ID); err != nil {
|
|
s.logger.Info(fmt.Sprintf("failed to delete shard group %d from database %s, retention policy %s: %s",
|
|
g.ID, d.Name, r.Name, err.Error()))
|
|
} else {
|
|
s.logger.Info(fmt.Sprintf("deleted shard group %d from database %s, retention policy %s",
|
|
g.ID, d.Name, r.Name))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func (s *Service) deleteShards() {
|
|
defer s.wg.Done()
|
|
|
|
ticker := time.NewTicker(s.checkInterval)
|
|
defer ticker.Stop()
|
|
for {
|
|
select {
|
|
case <-s.done:
|
|
return
|
|
|
|
case <-ticker.C:
|
|
s.logger.Info("retention policy shard deletion check commencing")
|
|
|
|
type deletionInfo struct {
|
|
db string
|
|
rp string
|
|
}
|
|
deletedShardIDs := make(map[uint64]deletionInfo, 0)
|
|
dbs := s.MetaClient.Databases()
|
|
for _, d := range dbs {
|
|
for _, r := range d.RetentionPolicies {
|
|
for _, g := range r.DeletedShardGroups() {
|
|
for _, sh := range g.Shards {
|
|
deletedShardIDs[sh.ID] = deletionInfo{db: d.Name, rp: r.Name}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
for _, id := range s.TSDBStore.ShardIDs() {
|
|
if info, ok := deletedShardIDs[id]; ok {
|
|
if err := s.TSDBStore.DeleteShard(id); err != nil {
|
|
s.logger.Error(fmt.Sprintf("failed to delete shard ID %d from database %s, retention policy %s: %s",
|
|
id, info.db, info.rp, err.Error()))
|
|
continue
|
|
}
|
|
s.logger.Info(fmt.Sprintf("shard ID %d from database %s, retention policy %s, deleted",
|
|
id, info.db, info.rp))
|
|
}
|
|
}
|
|
if err := s.MetaClient.PruneShardGroups(); err != nil {
|
|
s.logger.Info(fmt.Sprintf("error pruning shard groups: %s", err))
|
|
}
|
|
}
|
|
}
|
|
}
|