mirror of
https://github.com/gravitl/netmaker.git
synced 2025-12-19 09:08:58 +08:00
Merge pull request #649 from gravitl/feature_v0.10.0_server_response
Feature v0.10.0 server response
This commit is contained in:
commit
0409792c92
10 changed files with 195 additions and 76 deletions
|
|
@ -67,6 +67,16 @@ func (s *NodeServiceServer) CreateNode(ctx context.Context, req *nodepb.Object)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var serverNodes = logic.GetServerNodes(node.Network)
|
||||||
|
var serverAddrs = make([]models.ServerAddr, len(serverNodes))
|
||||||
|
for i, server := range serverNodes {
|
||||||
|
serverAddrs[i] = models.ServerAddr{
|
||||||
|
IsLeader: logic.IsLeader(&server),
|
||||||
|
Address: server.Address,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
node.NetworkSettings.DefaultServerAddrs = serverAddrs
|
||||||
|
|
||||||
err = logic.CreateNode(&node)
|
err = logic.CreateNode(&node)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
|
|
||||||
|
|
@ -6,6 +6,7 @@ import (
|
||||||
"sort"
|
"sort"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
@ -36,6 +37,9 @@ func ResetLogs() {
|
||||||
|
|
||||||
// Log - handles adding logs
|
// Log - handles adding logs
|
||||||
func Log(verbosity int, message ...string) {
|
func Log(verbosity int, message ...string) {
|
||||||
|
var mu sync.Mutex
|
||||||
|
mu.Lock()
|
||||||
|
defer mu.Unlock()
|
||||||
var currentTime = time.Now()
|
var currentTime = time.Now()
|
||||||
var currentMessage = makeString(message...)
|
var currentMessage = makeString(message...)
|
||||||
if int32(verbosity) <= getVerbose() && getVerbose() >= 0 {
|
if int32(verbosity) <= getVerbose() && getVerbose() >= 0 {
|
||||||
|
|
|
||||||
|
|
@ -58,6 +58,21 @@ func GetSortedNetworkServerNodes(network string) ([]models.Node, error) {
|
||||||
return nodes, nil
|
return nodes, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetServerNodes - gets the server nodes of a network
|
||||||
|
func GetServerNodes(network string) []models.Node {
|
||||||
|
var nodes, err = GetNetworkNodes(network)
|
||||||
|
var serverNodes = make([]models.Node, 0)
|
||||||
|
if err != nil {
|
||||||
|
return serverNodes
|
||||||
|
}
|
||||||
|
for _, node := range nodes {
|
||||||
|
if node.IsServer == "yes" {
|
||||||
|
serverNodes = append(serverNodes, node)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return serverNodes
|
||||||
|
}
|
||||||
|
|
||||||
// UncordonNode - approves a node to join a network
|
// UncordonNode - approves a node to join a network
|
||||||
func UncordonNode(nodeid string) (models.Node, error) {
|
func UncordonNode(nodeid string) (models.Node, error) {
|
||||||
node, err := GetNodeByID(nodeid)
|
node, err := GetNodeByID(nodeid)
|
||||||
|
|
|
||||||
|
|
@ -20,6 +20,7 @@ func GetPeerUpdate(node *models.Node) (models.PeerUpdate, error) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return models.PeerUpdate{}, err
|
return models.PeerUpdate{}, err
|
||||||
}
|
}
|
||||||
|
var serverNodeAddresses = []models.ServerAddr{}
|
||||||
for _, peer := range networkNodes {
|
for _, peer := range networkNodes {
|
||||||
if peer.ID == node.ID {
|
if peer.ID == node.ID {
|
||||||
//skip yourself
|
//skip yourself
|
||||||
|
|
@ -55,9 +56,13 @@ func GetPeerUpdate(node *models.Node) (models.PeerUpdate, error) {
|
||||||
PersistentKeepaliveInterval: &keepalive,
|
PersistentKeepaliveInterval: &keepalive,
|
||||||
}
|
}
|
||||||
peers = append(peers, peerData)
|
peers = append(peers, peerData)
|
||||||
|
if peer.IsServer == "yes" {
|
||||||
|
serverNodeAddresses = append(serverNodeAddresses, models.ServerAddr{IsLeader: IsLeader(&peer), Address: peer.Address})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
peerUpdate.Network = node.Network
|
peerUpdate.Network = node.Network
|
||||||
peerUpdate.Peers = peers
|
peerUpdate.Peers = peers
|
||||||
|
peerUpdate.ServerAddrs = serverNodeAddresses
|
||||||
return peerUpdate, nil
|
return peerUpdate, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -2,12 +2,15 @@ package models
|
||||||
|
|
||||||
import "golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
import "golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
||||||
|
|
||||||
|
// PeerUpdate - struct
|
||||||
type PeerUpdate struct {
|
type PeerUpdate struct {
|
||||||
Network string
|
Network string `json:"network" bson:"network" yaml:"network"`
|
||||||
Peers []wgtypes.PeerConfig
|
ServerAddrs []ServerAddr `json:"serveraddrs" bson:"serveraddrs" yaml:"serveraddrs"`
|
||||||
|
Peers []wgtypes.PeerConfig `json:"peers" bson:"peers" yaml:"peers"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// KeyUpdate - key update struct
|
||||||
type KeyUpdate struct {
|
type KeyUpdate struct {
|
||||||
Network string
|
Network string `json:"network" bson:"network"`
|
||||||
Interface string
|
Interface string `json:"interface" bson:"interface"`
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -34,10 +34,11 @@ type Network struct {
|
||||||
LocalRange string `json:"localrange" bson:"localrange" validate:"omitempty,cidr"`
|
LocalRange string `json:"localrange" bson:"localrange" validate:"omitempty,cidr"`
|
||||||
|
|
||||||
// checkin interval is depreciated at the network level. Set on server with CHECKIN_INTERVAL
|
// checkin interval is depreciated at the network level. Set on server with CHECKIN_INTERVAL
|
||||||
DefaultCheckInInterval int32 `json:"checkininterval,omitempty" bson:"checkininterval,omitempty" validate:"omitempty,numeric,min=2,max=100000"`
|
DefaultCheckInInterval int32 `json:"checkininterval,omitempty" bson:"checkininterval,omitempty" validate:"omitempty,numeric,min=2,max=100000"`
|
||||||
DefaultUDPHolePunch string `json:"defaultudpholepunch" bson:"defaultudpholepunch" validate:"checkyesorno"`
|
DefaultUDPHolePunch string `json:"defaultudpholepunch" bson:"defaultudpholepunch" validate:"checkyesorno"`
|
||||||
DefaultExtClientDNS string `json:"defaultextclientdns" bson:"defaultextclientdns"`
|
DefaultExtClientDNS string `json:"defaultextclientdns" bson:"defaultextclientdns"`
|
||||||
DefaultMTU int32 `json:"defaultmtu" bson:"defaultmtu"`
|
DefaultMTU int32 `json:"defaultmtu" bson:"defaultmtu"`
|
||||||
|
DefaultServerAddrs []ServerAddr `json:"defaultserveraddrs" bson:"defaultserveraddrs" yaml:"defaultserveraddrs"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// SaveData - sensitive fields of a network that should be kept the same
|
// SaveData - sensitive fields of a network that should be kept the same
|
||||||
|
|
|
||||||
|
|
@ -169,3 +169,9 @@ type Telemetry struct {
|
||||||
UUID string `json:"uuid" bson:"uuid"`
|
UUID string `json:"uuid" bson:"uuid"`
|
||||||
LastSend int64 `json:"lastsend" bson:"lastsend"`
|
LastSend int64 `json:"lastsend" bson:"lastsend"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ServerAddr - to pass to clients to tell server addresses and if it's the leader or not
|
||||||
|
type ServerAddr struct {
|
||||||
|
IsLeader bool `json:"isleader" bson:"isleader" yaml:"isleader"`
|
||||||
|
Address string `json:"address" bson:"address" yaml:"address"`
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -3,10 +3,12 @@ package functions
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
"log"
|
"log"
|
||||||
"os"
|
"os"
|
||||||
"os/signal"
|
"os/signal"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
"sync"
|
||||||
"syscall"
|
"syscall"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
|
@ -19,6 +21,22 @@ import (
|
||||||
"golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
"golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var messageCache = make(map[string]string, 20)
|
||||||
|
|
||||||
|
const lastNodeUpdate = "lnu"
|
||||||
|
const lastPeerUpdate = "lpu"
|
||||||
|
|
||||||
|
func insert(network, which, cache string) {
|
||||||
|
var mu sync.Mutex
|
||||||
|
mu.Lock()
|
||||||
|
defer mu.Unlock()
|
||||||
|
messageCache[fmt.Sprintf("%s%s", network, which)] = cache
|
||||||
|
}
|
||||||
|
|
||||||
|
func read(network, which string) string {
|
||||||
|
return messageCache[fmt.Sprintf("%s%s", network, which)]
|
||||||
|
}
|
||||||
|
|
||||||
// Daemon runs netclient daemon from command line
|
// Daemon runs netclient daemon from command line
|
||||||
func Daemon() error {
|
func Daemon() error {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
|
@ -41,8 +59,13 @@ func Daemon() error {
|
||||||
// SetupMQTT creates a connection to broker and return client
|
// SetupMQTT creates a connection to broker and return client
|
||||||
func SetupMQTT(cfg *config.ClientConfig) mqtt.Client {
|
func SetupMQTT(cfg *config.ClientConfig) mqtt.Client {
|
||||||
opts := mqtt.NewClientOptions()
|
opts := mqtt.NewClientOptions()
|
||||||
ncutils.Log("setting broker to " + cfg.Server.CoreDNSAddr + ":1883")
|
for _, server := range cfg.Node.NetworkSettings.DefaultServerAddrs {
|
||||||
opts.AddBroker(cfg.Server.CoreDNSAddr + ":1883")
|
if server.Address != "" && server.IsLeader {
|
||||||
|
ncutils.Log(fmt.Sprintf("adding server (%s) to listen on network %s \n", server.Address, cfg.Node.Network))
|
||||||
|
opts.AddBroker(server.Address + ":1883")
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
opts.SetDefaultPublishHandler(All)
|
opts.SetDefaultPublishHandler(All)
|
||||||
client := mqtt.NewClient(opts)
|
client := mqtt.NewClient(opts)
|
||||||
if token := client.Connect(); token.Wait() && token.Error() != nil {
|
if token := client.Connect(); token.Wait() && token.Error() != nil {
|
||||||
|
|
@ -65,13 +88,13 @@ func MessageQueue(ctx context.Context, network string) {
|
||||||
}
|
}
|
||||||
ncutils.Log("subscribed to all topics for debugging purposes")
|
ncutils.Log("subscribed to all topics for debugging purposes")
|
||||||
}
|
}
|
||||||
if token := client.Subscribe("update/"+cfg.Node.ID, 0, NodeUpdate); token.Wait() && token.Error() != nil {
|
if token := client.Subscribe("update/"+cfg.Node.ID, 0, mqtt.MessageHandler(NodeUpdate)); token.Wait() && token.Error() != nil {
|
||||||
log.Fatal(token.Error())
|
log.Fatal(token.Error())
|
||||||
}
|
}
|
||||||
if cfg.DebugOn {
|
if cfg.DebugOn {
|
||||||
ncutils.Log("subscribed to node updates for node " + cfg.Node.Name + " update/" + cfg.Node.ID)
|
ncutils.Log("subscribed to node updates for node " + cfg.Node.Name + " update/" + cfg.Node.ID)
|
||||||
}
|
}
|
||||||
if token := client.Subscribe("update/peers/"+cfg.Node.ID, 0, UpdatePeers); token.Wait() && token.Error() != nil {
|
if token := client.Subscribe("update/peers/"+cfg.Node.ID, 0, mqtt.MessageHandler(UpdatePeers)); token.Wait() && token.Error() != nil {
|
||||||
log.Fatal(token.Error())
|
log.Fatal(token.Error())
|
||||||
}
|
}
|
||||||
if cfg.DebugOn {
|
if cfg.DebugOn {
|
||||||
|
|
@ -91,8 +114,7 @@ var All mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodeUpdate -- mqtt message handler for /update/<NodeID> topic
|
// NodeUpdate -- mqtt message handler for /update/<NodeID> topic
|
||||||
var NodeUpdate mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {
|
func NodeUpdate(client mqtt.Client, msg mqtt.Message) {
|
||||||
ncutils.Log("received message to update node " + string(msg.Payload()))
|
|
||||||
//potentiall blocking i/o so do this in a go routine
|
//potentiall blocking i/o so do this in a go routine
|
||||||
go func() {
|
go func() {
|
||||||
var newNode models.Node
|
var newNode models.Node
|
||||||
|
|
@ -102,6 +124,13 @@ var NodeUpdate mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message)
|
||||||
ncutils.Log("error unmarshalling node update data" + err.Error())
|
ncutils.Log("error unmarshalling node update data" + err.Error())
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
ncutils.Log("received message to update node " + newNode.Name)
|
||||||
|
// see if cache hit, if so skip
|
||||||
|
var currentMessage = read(newNode.Network, lastNodeUpdate)
|
||||||
|
if currentMessage == string(msg.Payload()) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
insert(newNode.Network, lastNodeUpdate, string(msg.Payload()))
|
||||||
cfg.Network = newNode.Network
|
cfg.Network = newNode.Network
|
||||||
cfg.ReadConfig()
|
cfg.ReadConfig()
|
||||||
//check if interface name has changed if so delete.
|
//check if interface name has changed if so delete.
|
||||||
|
|
@ -169,7 +198,7 @@ var NodeUpdate mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message)
|
||||||
}
|
}
|
||||||
|
|
||||||
// UpdatePeers -- mqtt message handler for /update/peers/<NodeID> topic
|
// UpdatePeers -- mqtt message handler for /update/peers/<NodeID> topic
|
||||||
var UpdatePeers mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message) {
|
func UpdatePeers(client mqtt.Client, msg mqtt.Message) {
|
||||||
go func() {
|
go func() {
|
||||||
var peerUpdate models.PeerUpdate
|
var peerUpdate models.PeerUpdate
|
||||||
err := json.Unmarshal(msg.Payload(), &peerUpdate)
|
err := json.Unmarshal(msg.Payload(), &peerUpdate)
|
||||||
|
|
@ -177,10 +206,21 @@ var UpdatePeers mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message)
|
||||||
ncutils.Log("error unmarshalling peer data")
|
ncutils.Log("error unmarshalling peer data")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
// see if cache hit, if so skip
|
||||||
|
var currentMessage = read(peerUpdate.Network, lastPeerUpdate)
|
||||||
|
if currentMessage == string(msg.Payload()) {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
insert(peerUpdate.Network, lastPeerUpdate, string(msg.Payload()))
|
||||||
ncutils.Log("update peer handler")
|
ncutils.Log("update peer handler")
|
||||||
var cfg config.ClientConfig
|
var cfg config.ClientConfig
|
||||||
cfg.Network = peerUpdate.Network
|
cfg.Network = peerUpdate.Network
|
||||||
cfg.ReadConfig()
|
cfg.ReadConfig()
|
||||||
|
var shouldReSub = shouldResub(cfg.Node.NetworkSettings.DefaultServerAddrs, peerUpdate.ServerAddrs)
|
||||||
|
if shouldReSub {
|
||||||
|
Resubscribe(client, &cfg)
|
||||||
|
cfg.Node.NetworkSettings.DefaultServerAddrs = peerUpdate.ServerAddrs
|
||||||
|
}
|
||||||
file := ncutils.GetNetclientPathSpecific() + cfg.Node.Interface + ".conf"
|
file := ncutils.GetNetclientPathSpecific() + cfg.Node.Interface + ".conf"
|
||||||
err = wireguard.UpdateWgPeers(file, peerUpdate.Peers)
|
err = wireguard.UpdateWgPeers(file, peerUpdate.Peers)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|
@ -196,6 +236,29 @@ var UpdatePeers mqtt.MessageHandler = func(client mqtt.Client, msg mqtt.Message)
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Resubscribe --- handles resubscribing if needed
|
||||||
|
func Resubscribe(client mqtt.Client, cfg *config.ClientConfig) error {
|
||||||
|
if err := config.ModConfig(&cfg.Node); err == nil {
|
||||||
|
ncutils.Log("resubbing on network " + cfg.Node.Network)
|
||||||
|
client.Disconnect(250)
|
||||||
|
client = SetupMQTT(cfg)
|
||||||
|
if token := client.Subscribe("update/"+cfg.Node.ID, 0, NodeUpdate); token.Wait() && token.Error() != nil {
|
||||||
|
log.Fatal(token.Error())
|
||||||
|
}
|
||||||
|
if cfg.DebugOn {
|
||||||
|
ncutils.Log("subscribed to node updates for node " + cfg.Node.Name + " update/" + cfg.Node.ID)
|
||||||
|
}
|
||||||
|
if token := client.Subscribe("update/peers/"+cfg.Node.ID, 0, UpdatePeers); token.Wait() && token.Error() != nil {
|
||||||
|
log.Fatal(token.Error())
|
||||||
|
}
|
||||||
|
ncutils.Log("finished re subbing")
|
||||||
|
return nil
|
||||||
|
} else {
|
||||||
|
ncutils.Log("could not mod config when re-subbing")
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// UpdateKeys -- updates private key and returns new publickey
|
// UpdateKeys -- updates private key and returns new publickey
|
||||||
func UpdateKeys(cfg *config.ClientConfig, client mqtt.Client) error {
|
func UpdateKeys(cfg *config.ClientConfig, client mqtt.Client) error {
|
||||||
ncutils.Log("received message to update keys")
|
ncutils.Log("received message to update keys")
|
||||||
|
|
@ -291,3 +354,15 @@ func Hello(cfg *config.ClientConfig, network string) {
|
||||||
}
|
}
|
||||||
client.Disconnect(250)
|
client.Disconnect(250)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func shouldResub(currentServers, newServers []models.ServerAddr) bool {
|
||||||
|
if len(currentServers) != len(newServers) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
for _, srv := range currentServers {
|
||||||
|
if !ncutils.ServerAddrSliceContains(newServers, srv) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
|
||||||
|
|
@ -7,6 +7,7 @@ import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"log"
|
"log"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
|
"runtime"
|
||||||
|
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
nodepb "github.com/gravitl/netmaker/grpc"
|
nodepb "github.com/gravitl/netmaker/grpc"
|
||||||
|
|
@ -101,8 +102,7 @@ func JoinNetwork(cfg config.ClientConfig, privateKey string) error {
|
||||||
// make sure name is appropriate, if not, give blank name
|
// make sure name is appropriate, if not, give blank name
|
||||||
cfg.Node.Name = formatName(cfg.Node)
|
cfg.Node.Name = formatName(cfg.Node)
|
||||||
// differentiate between client/server here
|
// differentiate between client/server here
|
||||||
var node models.Node // fill this node with appropriate calls
|
var node = models.Node{
|
||||||
postnode := &models.Node{
|
|
||||||
Password: cfg.Node.Password,
|
Password: cfg.Node.Password,
|
||||||
ID: cfg.Node.ID,
|
ID: cfg.Node.ID,
|
||||||
MacAddress: cfg.Node.MacAddress,
|
MacAddress: cfg.Node.MacAddress,
|
||||||
|
|
@ -124,44 +124,17 @@ func JoinNetwork(cfg config.ClientConfig, privateKey string) error {
|
||||||
UDPHolePunch: cfg.Node.UDPHolePunch,
|
UDPHolePunch: cfg.Node.UDPHolePunch,
|
||||||
}
|
}
|
||||||
|
|
||||||
if cfg.Node.IsServer != "yes" {
|
ncutils.Log("joining " + cfg.Network + " at " + cfg.Server.GRPCAddress)
|
||||||
ncutils.Log("joining " + cfg.Network + " at " + cfg.Server.GRPCAddress)
|
var wcclient nodepb.NodeServiceClient
|
||||||
var wcclient nodepb.NodeServiceClient
|
|
||||||
|
|
||||||
conn, err := grpc.Dial(cfg.Server.GRPCAddress,
|
conn, err := grpc.Dial(cfg.Server.GRPCAddress,
|
||||||
ncutils.GRPCRequestOpts(cfg.Server.GRPCSSL))
|
ncutils.GRPCRequestOpts(cfg.Server.GRPCSSL))
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatalf("Unable to establish client connection to "+cfg.Server.GRPCAddress+": %v", err)
|
log.Fatalf("Unable to establish client connection to "+cfg.Server.GRPCAddress+": %v", err)
|
||||||
}
|
|
||||||
defer conn.Close()
|
|
||||||
wcclient = nodepb.NewNodeServiceClient(conn)
|
|
||||||
|
|
||||||
if err = config.ModConfig(postnode); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
data, err := json.Marshal(postnode)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// Create node on server
|
|
||||||
res, err := wcclient.CreateNode(
|
|
||||||
context.TODO(),
|
|
||||||
&nodepb.Object{
|
|
||||||
Data: string(data),
|
|
||||||
Type: nodepb.NODE_TYPE,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
ncutils.PrintLog("node created on remote server...updating configs", 1)
|
|
||||||
|
|
||||||
nodeData := res.Data
|
|
||||||
if err = json.Unmarshal([]byte(nodeData), &node); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
defer conn.Close()
|
||||||
|
wcclient = nodepb.NewNodeServiceClient(conn)
|
||||||
|
|
||||||
// get free port based on returned default listen port
|
// get free port based on returned default listen port
|
||||||
node.ListenPort, err = ncutils.GetFreePort(node.ListenPort)
|
node.ListenPort, err = ncutils.GetFreePort(node.ListenPort)
|
||||||
|
|
@ -182,32 +155,48 @@ func JoinNetwork(cfg config.ClientConfig, privateKey string) error {
|
||||||
cfg.Node.IsStatic = "yes"
|
cfg.Node.IsStatic = "yes"
|
||||||
}
|
}
|
||||||
|
|
||||||
if node.IsServer != "yes" { // == handle client side ==
|
err = wireguard.StorePrivKey(privateKey, cfg.Network)
|
||||||
err = config.ModConfig(&node)
|
if err != nil {
|
||||||
if err != nil {
|
return err
|
||||||
return err
|
}
|
||||||
}
|
if node.IsPending == "yes" {
|
||||||
err = wireguard.StorePrivKey(privateKey, cfg.Network)
|
ncutils.Log("Node is marked as PENDING.")
|
||||||
if err != nil {
|
ncutils.Log("Awaiting approval from Admin before configuring WireGuard.")
|
||||||
return err
|
if cfg.Daemon != "off" {
|
||||||
}
|
return daemon.InstallDaemon(cfg)
|
||||||
if node.IsPending == "yes" {
|
|
||||||
ncutils.Log("Node is marked as PENDING.")
|
|
||||||
ncutils.Log("Awaiting approval from Admin before configuring WireGuard.")
|
|
||||||
if cfg.Daemon != "off" {
|
|
||||||
return daemon.InstallDaemon(cfg)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// pushing any local changes to server before starting wireguard
|
|
||||||
err = Push(cfg.Network)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
// attempt to make backup
|
|
||||||
if err = config.SaveBackup(node.Network); err != nil {
|
|
||||||
ncutils.Log("failed to make backup, node will not auto restore if config is corrupted")
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
data, err := json.Marshal(&node)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// Create node on server
|
||||||
|
res, err := wcclient.CreateNode(
|
||||||
|
context.TODO(),
|
||||||
|
&nodepb.Object{
|
||||||
|
Data: string(data),
|
||||||
|
Type: nodepb.NODE_TYPE,
|
||||||
|
},
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
ncutils.PrintLog("node created on remote server...updating configs", 1)
|
||||||
|
|
||||||
|
nodeData := res.Data
|
||||||
|
if err = json.Unmarshal([]byte(nodeData), &node); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
node.OS = runtime.GOOS
|
||||||
|
cfg.Node = node
|
||||||
|
err = config.ModConfig(&node)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// attempt to make backup
|
||||||
|
if err = config.SaveBackup(node.Network); err != nil {
|
||||||
|
ncutils.Log("failed to make backup, node will not auto restore if config is corrupted")
|
||||||
|
}
|
||||||
|
|
||||||
ncutils.Log("retrieving peers")
|
ncutils.Log("retrieving peers")
|
||||||
peers, hasGateway, gateways, err := server.GetPeers(node.MacAddress, cfg.Network, cfg.Server.GRPCAddress, node.IsDualStack == "yes", node.IsIngressGateway == "yes", node.IsServer == "yes")
|
peers, hasGateway, gateways, err := server.GetPeers(node.MacAddress, cfg.Network, cfg.Server.GRPCAddress, node.IsDualStack == "yes", node.IsIngressGateway == "yes", node.IsServer == "yes")
|
||||||
|
|
|
||||||
|
|
@ -17,6 +17,7 @@ import (
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/gravitl/netmaker/models"
|
||||||
"golang.zx2c4.com/wireguard/wgctrl"
|
"golang.zx2c4.com/wireguard/wgctrl"
|
||||||
"golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
"golang.zx2c4.com/wireguard/wgctrl/wgtypes"
|
||||||
"google.golang.org/grpc"
|
"google.golang.org/grpc"
|
||||||
|
|
@ -532,3 +533,13 @@ func CheckWG() {
|
||||||
log.Println("running userspace WireGuard with " + uspace)
|
log.Println("running userspace WireGuard with " + uspace)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ServerAddrSliceContains - sees if a string slice contains a string element
|
||||||
|
func ServerAddrSliceContains(slice []models.ServerAddr, item models.ServerAddr) bool {
|
||||||
|
for _, s := range slice {
|
||||||
|
if s.Address == item.Address && s.IsLeader == item.IsLeader {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue