ncro: split into main.go entrypoint and root.go command; inject version via ldflags
Signed-off-by: NotAShelf <raf@notashelf.dev> Change-Id: I640d23c9fb4e53ee22b1c835b08b70896a6a6964
This commit is contained in:
parent
578e0b2d5e
commit
7765fea445
2 changed files with 217 additions and 206 deletions
207
cmd/ncro/main.go
207
cmd/ncro/main.go
|
|
@ -1,210 +1,5 @@
|
|||
package main
|
||||
|
||||
import (
|
||||
"context"
|
||||
"crypto/ed25519"
|
||||
"encoding/hex"
|
||||
"errors"
|
||||
"fmt"
|
||||
"log/slog"
|
||||
"net/http"
|
||||
"os"
|
||||
"os/signal"
|
||||
"syscall"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/spf13/cobra"
|
||||
"github.com/spf13/viper"
|
||||
"notashelf.dev/ncro/internal/cache"
|
||||
"notashelf.dev/ncro/internal/config"
|
||||
"notashelf.dev/ncro/internal/mesh"
|
||||
"notashelf.dev/ncro/internal/metrics"
|
||||
"notashelf.dev/ncro/internal/prober"
|
||||
"notashelf.dev/ncro/internal/router"
|
||||
"notashelf.dev/ncro/internal/server"
|
||||
)
|
||||
|
||||
func main() {
|
||||
if err := newRootCmd().Execute(); err != nil {
|
||||
os.Exit(1)
|
||||
}
|
||||
}
|
||||
|
||||
func newRootCmd() *cobra.Command {
|
||||
cmd := &cobra.Command{
|
||||
Use: "ncro",
|
||||
Short: "Nix Cache Route Optimizer",
|
||||
SilenceUsage: true,
|
||||
RunE: runServer,
|
||||
}
|
||||
|
||||
cmd.Flags().StringP("config", "c", "", "path to config YAML file (env: NCRO_CONFIG)")
|
||||
_ = viper.BindPFlag("config", cmd.Flags().Lookup("config"))
|
||||
viper.SetEnvPrefix("NCRO")
|
||||
viper.AutomaticEnv()
|
||||
|
||||
return cmd
|
||||
}
|
||||
|
||||
func runServer(_ *cobra.Command, _ []string) error {
|
||||
cfg, err := config.Load(viper.GetString("config"))
|
||||
if err != nil {
|
||||
return fmt.Errorf("load config: %w", err)
|
||||
}
|
||||
if err := cfg.Validate(); err != nil {
|
||||
return fmt.Errorf("invalid config: %w", err)
|
||||
}
|
||||
|
||||
level := slog.LevelInfo
|
||||
switch cfg.Logging.Level {
|
||||
case "debug":
|
||||
level = slog.LevelDebug
|
||||
case "warn":
|
||||
level = slog.LevelWarn
|
||||
case "error":
|
||||
level = slog.LevelError
|
||||
}
|
||||
var handler slog.Handler
|
||||
if cfg.Logging.Format == "text" {
|
||||
handler = slog.NewTextHandler(os.Stdout, &slog.HandlerOptions{Level: level})
|
||||
} else {
|
||||
handler = slog.NewJSONHandler(os.Stdout, &slog.HandlerOptions{Level: level})
|
||||
}
|
||||
slog.SetDefault(slog.New(handler))
|
||||
|
||||
metrics.Register(prometheus.DefaultRegisterer)
|
||||
|
||||
db, err := cache.Open(cfg.Cache.DBPath, cfg.Cache.MaxEntries)
|
||||
if err != nil {
|
||||
return fmt.Errorf("open database: %w", err)
|
||||
}
|
||||
defer db.Close()
|
||||
|
||||
expireDone := make(chan struct{})
|
||||
go func() {
|
||||
ticker := time.NewTicker(5 * time.Minute)
|
||||
defer ticker.Stop()
|
||||
for {
|
||||
select {
|
||||
case <-expireDone:
|
||||
return
|
||||
case <-ticker.C:
|
||||
if err := db.ExpireOldRoutes(); err != nil {
|
||||
slog.Warn("expire routes error", "error", err)
|
||||
}
|
||||
if err := db.ExpireNegatives(); err != nil {
|
||||
slog.Warn("expire negatives error", "error", err)
|
||||
}
|
||||
if count, err := db.RouteCount(); err == nil {
|
||||
metrics.RouteEntries.Set(float64(count))
|
||||
}
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
p := prober.New(cfg.Cache.LatencyAlpha)
|
||||
p.InitUpstreams(cfg.Upstreams)
|
||||
|
||||
if rows, err := db.LoadAllHealth(); err == nil {
|
||||
for _, row := range rows {
|
||||
p.Seed(row.URL, row.EMALatency, row.ConsecutiveFails, int64(row.TotalQueries))
|
||||
}
|
||||
} else {
|
||||
slog.Warn("failed to load persisted health data", "error", err)
|
||||
}
|
||||
|
||||
p.SetHealthPersistence(func(url string, ema float64, cf uint32, tq uint64) {
|
||||
if err := db.SaveHealth(url, ema, int(cf), int64(tq)); err != nil {
|
||||
slog.Warn("failed to save health", "url", url, "error", err)
|
||||
}
|
||||
})
|
||||
|
||||
for _, u := range cfg.Upstreams {
|
||||
go p.ProbeUpstream(u.URL)
|
||||
}
|
||||
|
||||
probeDone := make(chan struct{})
|
||||
go p.RunProbeLoop(30*time.Second, probeDone)
|
||||
|
||||
r := router.New(db, p, cfg.Cache.TTL.Duration, 5*time.Second, cfg.Cache.NegativeTTL.Duration)
|
||||
for _, u := range cfg.Upstreams {
|
||||
if u.PublicKey != "" {
|
||||
if err := r.SetUpstreamKey(u.URL, u.PublicKey); err != nil {
|
||||
return fmt.Errorf("invalid upstream public key for %s: %w", u.URL, err)
|
||||
}
|
||||
slog.Info("narinfo signature verification enabled", "upstream", u.URL)
|
||||
}
|
||||
}
|
||||
|
||||
var gossipDone chan struct{}
|
||||
if cfg.Mesh.Enabled {
|
||||
store := mesh.NewRouteStore()
|
||||
node, err := mesh.NewNode(cfg.Mesh.PrivateKeyPath, store)
|
||||
if err != nil {
|
||||
return fmt.Errorf("create mesh node: %w", err)
|
||||
}
|
||||
slog.Info("mesh node identity", "node_id", node.ID(),
|
||||
"public_key", hex.EncodeToString(node.PublicKey()))
|
||||
|
||||
allowedKeys := make([]ed25519.PublicKey, 0, len(cfg.Mesh.Peers))
|
||||
for _, peer := range cfg.Mesh.Peers {
|
||||
if peer.PublicKey != "" {
|
||||
b, _ := hex.DecodeString(peer.PublicKey)
|
||||
allowedKeys = append(allowedKeys, ed25519.PublicKey(b))
|
||||
}
|
||||
}
|
||||
|
||||
if err := mesh.ListenAndServe(cfg.Mesh.BindAddr, store, allowedKeys...); err != nil {
|
||||
return fmt.Errorf("start mesh listener: %w", err)
|
||||
}
|
||||
|
||||
peerAddrs := make([]string, len(cfg.Mesh.Peers))
|
||||
for i, p := range cfg.Mesh.Peers {
|
||||
peerAddrs[i] = p.Addr
|
||||
}
|
||||
|
||||
gossipDone = make(chan struct{})
|
||||
go mesh.RunGossipLoop(node, db, peerAddrs, cfg.Mesh.GossipInterval.Duration, gossipDone)
|
||||
slog.Info("mesh enabled", "addr", cfg.Mesh.BindAddr, "peers", len(cfg.Mesh.Peers))
|
||||
}
|
||||
|
||||
srv := &http.Server{
|
||||
Addr: cfg.Server.Listen,
|
||||
Handler: server.New(r, p, db, cfg.Upstreams, cfg.Server.CachePriority),
|
||||
ReadTimeout: cfg.Server.ReadTimeout.Duration,
|
||||
WriteTimeout: cfg.Server.WriteTimeout.Duration,
|
||||
}
|
||||
|
||||
stop := make(chan os.Signal, 1)
|
||||
signal.Notify(stop, syscall.SIGINT, syscall.SIGTERM)
|
||||
|
||||
serverErr := make(chan error, 1)
|
||||
go func() {
|
||||
slog.Info("ncro listening", "addr", cfg.Server.Listen, "upstreams", len(cfg.Upstreams))
|
||||
if err := srv.ListenAndServe(); err != nil && !errors.Is(err, http.ErrServerClosed) {
|
||||
serverErr <- err
|
||||
}
|
||||
close(serverErr)
|
||||
}()
|
||||
|
||||
select {
|
||||
case <-stop:
|
||||
slog.Info("shutting down")
|
||||
case err := <-serverErr:
|
||||
return fmt.Errorf("server: %w", err)
|
||||
}
|
||||
|
||||
close(expireDone)
|
||||
close(probeDone)
|
||||
if gossipDone != nil {
|
||||
close(gossipDone)
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)
|
||||
defer cancel()
|
||||
if err := srv.Shutdown(ctx); err != nil {
|
||||
slog.Warn("shutdown error", "error", err)
|
||||
}
|
||||
return nil
|
||||
Execute()
|
||||
}
|
||||
|
|
|
|||
216
cmd/ncro/root.go
Normal file
216
cmd/ncro/root.go
Normal file
|
|
@ -0,0 +1,216 @@
|
|||
package main
|
||||
|
||||
import (
|
||||
"context"
|
||||
"crypto/ed25519"
|
||||
"encoding/hex"
|
||||
"errors"
|
||||
"fmt"
|
||||
"log/slog"
|
||||
"net/http"
|
||||
"os"
|
||||
"os/signal"
|
||||
"syscall"
|
||||
"time"
|
||||
|
||||
"github.com/prometheus/client_golang/prometheus"
|
||||
"github.com/spf13/cobra"
|
||||
"github.com/spf13/viper"
|
||||
"notashelf.dev/ncro/internal/cache"
|
||||
"notashelf.dev/ncro/internal/config"
|
||||
"notashelf.dev/ncro/internal/mesh"
|
||||
"notashelf.dev/ncro/internal/metrics"
|
||||
"notashelf.dev/ncro/internal/prober"
|
||||
"notashelf.dev/ncro/internal/router"
|
||||
"notashelf.dev/ncro/internal/server"
|
||||
)
|
||||
|
||||
// Injected at build time via -ldflags "-X main.version=<ver>".
|
||||
var version = "dev"
|
||||
|
||||
// Execute is the entrypoint called by main.
|
||||
func Execute() {
|
||||
if err := newRootCmd().Execute(); err != nil {
|
||||
os.Exit(1)
|
||||
}
|
||||
}
|
||||
|
||||
func newRootCmd() *cobra.Command {
|
||||
cmd := &cobra.Command{
|
||||
Use: "ncro",
|
||||
Short: "Nix Cache Route Optimizer",
|
||||
Version: version,
|
||||
SilenceUsage: true,
|
||||
RunE: runServer,
|
||||
}
|
||||
|
||||
cmd.Flags().StringP("config", "c", "", "path to config YAML file (env: NCRO_CONFIG)")
|
||||
_ = viper.BindPFlag("config", cmd.Flags().Lookup("config"))
|
||||
viper.SetEnvPrefix("NCRO")
|
||||
viper.AutomaticEnv()
|
||||
|
||||
return cmd
|
||||
}
|
||||
|
||||
func runServer(_ *cobra.Command, _ []string) error {
|
||||
cfg, err := config.Load(viper.GetString("config"))
|
||||
if err != nil {
|
||||
return fmt.Errorf("load config: %w", err)
|
||||
}
|
||||
if err := cfg.Validate(); err != nil {
|
||||
return fmt.Errorf("invalid config: %w", err)
|
||||
}
|
||||
|
||||
level := slog.LevelInfo
|
||||
switch cfg.Logging.Level {
|
||||
case "debug":
|
||||
level = slog.LevelDebug
|
||||
case "warn":
|
||||
level = slog.LevelWarn
|
||||
case "error":
|
||||
level = slog.LevelError
|
||||
}
|
||||
var handler slog.Handler
|
||||
if cfg.Logging.Format == "text" {
|
||||
handler = slog.NewTextHandler(os.Stdout, &slog.HandlerOptions{Level: level})
|
||||
} else {
|
||||
handler = slog.NewJSONHandler(os.Stdout, &slog.HandlerOptions{Level: level})
|
||||
}
|
||||
slog.SetDefault(slog.New(handler))
|
||||
|
||||
metrics.Register(prometheus.DefaultRegisterer)
|
||||
|
||||
db, err := cache.Open(cfg.Cache.DBPath, cfg.Cache.MaxEntries)
|
||||
if err != nil {
|
||||
return fmt.Errorf("open database: %w", err)
|
||||
}
|
||||
defer db.Close()
|
||||
|
||||
expireDone := make(chan struct{})
|
||||
go func() {
|
||||
ticker := time.NewTicker(5 * time.Minute)
|
||||
defer ticker.Stop()
|
||||
for {
|
||||
select {
|
||||
case <-expireDone:
|
||||
return
|
||||
case <-ticker.C:
|
||||
if err := db.ExpireOldRoutes(); err != nil {
|
||||
slog.Warn("expire routes error", "error", err)
|
||||
}
|
||||
if err := db.ExpireNegatives(); err != nil {
|
||||
slog.Warn("expire negatives error", "error", err)
|
||||
}
|
||||
if count, err := db.RouteCount(); err == nil {
|
||||
metrics.RouteEntries.Set(float64(count))
|
||||
}
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
p := prober.New(cfg.Cache.LatencyAlpha)
|
||||
p.InitUpstreams(cfg.Upstreams)
|
||||
|
||||
if rows, err := db.LoadAllHealth(); err == nil {
|
||||
for _, row := range rows {
|
||||
p.Seed(row.URL, row.EMALatency, row.ConsecutiveFails, int64(row.TotalQueries))
|
||||
}
|
||||
} else {
|
||||
slog.Warn("failed to load persisted health data", "error", err)
|
||||
}
|
||||
|
||||
p.SetHealthPersistence(func(url string, ema float64, cf uint32, tq uint64) {
|
||||
if err := db.SaveHealth(url, ema, int(cf), int64(tq)); err != nil {
|
||||
slog.Warn("failed to save health", "url", url, "error", err)
|
||||
}
|
||||
})
|
||||
|
||||
for _, u := range cfg.Upstreams {
|
||||
go p.ProbeUpstream(u.URL)
|
||||
}
|
||||
|
||||
probeDone := make(chan struct{})
|
||||
go p.RunProbeLoop(30*time.Second, probeDone)
|
||||
|
||||
r := router.New(db, p, cfg.Cache.TTL.Duration, 5*time.Second, cfg.Cache.NegativeTTL.Duration)
|
||||
for _, u := range cfg.Upstreams {
|
||||
if u.PublicKey != "" {
|
||||
if err := r.SetUpstreamKey(u.URL, u.PublicKey); err != nil {
|
||||
return fmt.Errorf("invalid upstream public key for %s: %w", u.URL, err)
|
||||
}
|
||||
slog.Info("narinfo signature verification enabled", "upstream", u.URL)
|
||||
}
|
||||
}
|
||||
|
||||
var gossipDone chan struct{}
|
||||
if cfg.Mesh.Enabled {
|
||||
store := mesh.NewRouteStore()
|
||||
node, err := mesh.NewNode(cfg.Mesh.PrivateKeyPath, store)
|
||||
if err != nil {
|
||||
return fmt.Errorf("create mesh node: %w", err)
|
||||
}
|
||||
slog.Info("mesh node identity", "node_id", node.ID(),
|
||||
"public_key", hex.EncodeToString(node.PublicKey()))
|
||||
|
||||
allowedKeys := make([]ed25519.PublicKey, 0, len(cfg.Mesh.Peers))
|
||||
for _, peer := range cfg.Mesh.Peers {
|
||||
if peer.PublicKey != "" {
|
||||
b, _ := hex.DecodeString(peer.PublicKey)
|
||||
allowedKeys = append(allowedKeys, ed25519.PublicKey(b))
|
||||
}
|
||||
}
|
||||
|
||||
if err := mesh.ListenAndServe(cfg.Mesh.BindAddr, store, allowedKeys...); err != nil {
|
||||
return fmt.Errorf("start mesh listener: %w", err)
|
||||
}
|
||||
|
||||
peerAddrs := make([]string, len(cfg.Mesh.Peers))
|
||||
for i, p := range cfg.Mesh.Peers {
|
||||
peerAddrs[i] = p.Addr
|
||||
}
|
||||
|
||||
gossipDone = make(chan struct{})
|
||||
go mesh.RunGossipLoop(node, db, peerAddrs, cfg.Mesh.GossipInterval.Duration, gossipDone)
|
||||
slog.Info("mesh enabled", "addr", cfg.Mesh.BindAddr, "peers", len(cfg.Mesh.Peers))
|
||||
}
|
||||
|
||||
srv := &http.Server{
|
||||
Addr: cfg.Server.Listen,
|
||||
Handler: server.New(r, p, db, cfg.Upstreams, cfg.Server.CachePriority),
|
||||
ReadTimeout: cfg.Server.ReadTimeout.Duration,
|
||||
WriteTimeout: cfg.Server.WriteTimeout.Duration,
|
||||
}
|
||||
|
||||
stop := make(chan os.Signal, 1)
|
||||
signal.Notify(stop, syscall.SIGINT, syscall.SIGTERM)
|
||||
|
||||
serverErr := make(chan error, 1)
|
||||
go func() {
|
||||
slog.Info("ncro listening", "addr", cfg.Server.Listen,
|
||||
"upstreams", len(cfg.Upstreams), "version", version)
|
||||
if err := srv.ListenAndServe(); err != nil && !errors.Is(err, http.ErrServerClosed) {
|
||||
serverErr <- err
|
||||
}
|
||||
close(serverErr)
|
||||
}()
|
||||
|
||||
select {
|
||||
case <-stop:
|
||||
slog.Info("shutting down")
|
||||
case err := <-serverErr:
|
||||
return fmt.Errorf("server: %w", err)
|
||||
}
|
||||
|
||||
close(expireDone)
|
||||
close(probeDone)
|
||||
if gossipDone != nil {
|
||||
close(gossipDone)
|
||||
}
|
||||
|
||||
ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)
|
||||
defer cancel()
|
||||
if err := srv.Shutdown(ctx); err != nil {
|
||||
slog.Warn("shutdown error", "error", err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
Loading…
Add table
Add a link
Reference in a new issue