kube.go 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352
  1. // Copyright 2016 flannel authors
  2. //
  3. // Licensed under the Apache License, Version 2.0 (the "License");
  4. // you may not use this file except in compliance with the License.
  5. // You may obtain a copy of the License at
  6. //
  7. // http://www.apache.org/licenses/LICENSE-2.0
  8. //
  9. // Unless required by applicable law or agreed to in writing, software
  10. // distributed under the License is distributed on an "AS IS" BASIS,
  11. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  12. // See the License for the specific language governing permissions and
  13. // limitations under the License.
  14. package kube
  15. import (
  16. "encoding/json"
  17. "errors"
  18. "fmt"
  19. "io/ioutil"
  20. "net"
  21. "os"
  22. "time"
  23. "github.com/coreos/flannel/pkg/ip"
  24. "github.com/coreos/flannel/subnet"
  25. "golang.org/x/net/context"
  26. v1 "k8s.io/api/core/v1"
  27. metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
  28. "k8s.io/apimachinery/pkg/runtime"
  29. "k8s.io/apimachinery/pkg/types"
  30. "k8s.io/apimachinery/pkg/util/strategicpatch"
  31. "k8s.io/apimachinery/pkg/util/wait"
  32. "k8s.io/apimachinery/pkg/watch"
  33. clientset "k8s.io/client-go/kubernetes"
  34. listers "k8s.io/client-go/listers/core/v1"
  35. "k8s.io/client-go/rest"
  36. "k8s.io/client-go/tools/cache"
  37. "k8s.io/client-go/tools/clientcmd"
  38. log "k8s.io/klog"
  39. )
  40. var (
  41. ErrUnimplemented = errors.New("unimplemented")
  42. )
  43. const (
  44. resyncPeriod = 5 * time.Minute
  45. nodeControllerSyncTimeout = 10 * time.Minute
  46. )
  47. type kubeSubnetManager struct {
  48. annotations annotations
  49. client clientset.Interface
  50. nodeName string
  51. nodeStore listers.NodeLister
  52. nodeController cache.Controller
  53. subnetConf *subnet.Config
  54. events chan subnet.Event
  55. }
  56. func NewSubnetManager(ctx context.Context, apiUrl, kubeconfig, prefix, netConfPath string) (subnet.Manager, error) {
  57. var cfg *rest.Config
  58. var err error
  59. // Try to build kubernetes config from a master url or a kubeconfig filepath. If neither masterUrl
  60. // or kubeconfigPath are passed in we fall back to inClusterConfig. If inClusterConfig fails,
  61. // we fallback to the default config.
  62. cfg, err = clientcmd.BuildConfigFromFlags(apiUrl, kubeconfig)
  63. if err != nil {
  64. return nil, fmt.Errorf("fail to create kubernetes config: %v", err)
  65. }
  66. c, err := clientset.NewForConfig(cfg)
  67. if err != nil {
  68. return nil, fmt.Errorf("unable to initialize client: %v", err)
  69. }
  70. // The kube subnet mgr needs to know the k8s node name that it's running on so it can annotate it.
  71. // If we're running as a pod then the POD_NAME and POD_NAMESPACE will be populated and can be used to find the node
  72. // name. Otherwise, the environment variable NODE_NAME can be passed in.
  73. nodeName := os.Getenv("NODE_NAME")
  74. if nodeName == "" {
  75. podName := os.Getenv("POD_NAME")
  76. podNamespace := os.Getenv("POD_NAMESPACE")
  77. if podName == "" || podNamespace == "" {
  78. return nil, fmt.Errorf("env variables POD_NAME and POD_NAMESPACE must be set")
  79. }
  80. pod, err := c.CoreV1().Pods(podNamespace).Get(ctx, podName, metav1.GetOptions{})
  81. if err != nil {
  82. return nil, fmt.Errorf("error retrieving pod spec for '%s/%s': %v", podNamespace, podName, err)
  83. }
  84. nodeName = pod.Spec.NodeName
  85. if nodeName == "" {
  86. return nil, fmt.Errorf("node name not present in pod spec '%s/%s'", podNamespace, podName)
  87. }
  88. }
  89. netConf, err := ioutil.ReadFile(netConfPath)
  90. if err != nil {
  91. return nil, fmt.Errorf("failed to read net conf: %v", err)
  92. }
  93. sc, err := subnet.ParseConfig(string(netConf))
  94. if err != nil {
  95. return nil, fmt.Errorf("error parsing subnet config: %s", err)
  96. }
  97. sm, err := newKubeSubnetManager(ctx, c, sc, nodeName, prefix)
  98. if err != nil {
  99. return nil, fmt.Errorf("error creating network manager: %s", err)
  100. }
  101. go sm.Run(context.Background())
  102. log.Infof("Waiting %s for node controller to sync", nodeControllerSyncTimeout)
  103. err = wait.Poll(time.Second, nodeControllerSyncTimeout, func() (bool, error) {
  104. return sm.nodeController.HasSynced(), nil
  105. })
  106. if err != nil {
  107. return nil, fmt.Errorf("error waiting for nodeController to sync state: %v", err)
  108. }
  109. log.Infof("Node controller sync successful")
  110. return sm, nil
  111. }
  112. func newKubeSubnetManager(ctx context.Context, c clientset.Interface, sc *subnet.Config, nodeName, prefix string) (*kubeSubnetManager, error) {
  113. var err error
  114. var ksm kubeSubnetManager
  115. ksm.annotations, err = newAnnotations(prefix)
  116. if err != nil {
  117. return nil, err
  118. }
  119. ksm.client = c
  120. ksm.nodeName = nodeName
  121. ksm.subnetConf = sc
  122. ksm.events = make(chan subnet.Event, 5000)
  123. indexer, controller := cache.NewIndexerInformer(
  124. &cache.ListWatch{
  125. ListFunc: func(options metav1.ListOptions) (runtime.Object, error) {
  126. return ksm.client.CoreV1().Nodes().List(ctx, options)
  127. },
  128. WatchFunc: func(options metav1.ListOptions) (watch.Interface, error) {
  129. return ksm.client.CoreV1().Nodes().Watch(ctx, options)
  130. },
  131. },
  132. &v1.Node{},
  133. resyncPeriod,
  134. cache.ResourceEventHandlerFuncs{
  135. AddFunc: func(obj interface{}) {
  136. ksm.handleAddLeaseEvent(subnet.EventAdded, obj)
  137. },
  138. UpdateFunc: ksm.handleUpdateLeaseEvent,
  139. DeleteFunc: func(obj interface{}) {
  140. node, isNode := obj.(*v1.Node)
  141. // We can get DeletedFinalStateUnknown instead of *api.Node here and we need to handle that correctly.
  142. if !isNode {
  143. deletedState, ok := obj.(cache.DeletedFinalStateUnknown)
  144. if !ok {
  145. log.Infof("Error received unexpected object: %v", obj)
  146. return
  147. }
  148. node, ok = deletedState.Obj.(*v1.Node)
  149. if !ok {
  150. log.Infof("Error deletedFinalStateUnknown contained non-Node object: %v", deletedState.Obj)
  151. return
  152. }
  153. obj = node
  154. }
  155. ksm.handleAddLeaseEvent(subnet.EventRemoved, obj)
  156. },
  157. },
  158. cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc},
  159. )
  160. ksm.nodeController = controller
  161. ksm.nodeStore = listers.NewNodeLister(indexer)
  162. return &ksm, nil
  163. }
  164. func (ksm *kubeSubnetManager) handleAddLeaseEvent(et subnet.EventType, obj interface{}) {
  165. n := obj.(*v1.Node)
  166. if s, ok := n.Annotations[ksm.annotations.SubnetKubeManaged]; !ok || s != "true" {
  167. return
  168. }
  169. l, err := ksm.nodeToLease(*n)
  170. if err != nil {
  171. log.Infof("Error turning node %q to lease: %v", n.ObjectMeta.Name, err)
  172. return
  173. }
  174. ksm.events <- subnet.Event{et, l}
  175. }
  176. func (ksm *kubeSubnetManager) handleUpdateLeaseEvent(oldObj, newObj interface{}) {
  177. o := oldObj.(*v1.Node)
  178. n := newObj.(*v1.Node)
  179. if s, ok := n.Annotations[ksm.annotations.SubnetKubeManaged]; !ok || s != "true" {
  180. return
  181. }
  182. if o.Annotations[ksm.annotations.BackendData] == n.Annotations[ksm.annotations.BackendData] &&
  183. o.Annotations[ksm.annotations.BackendType] == n.Annotations[ksm.annotations.BackendType] &&
  184. o.Annotations[ksm.annotations.BackendPublicIP] == n.Annotations[ksm.annotations.BackendPublicIP] {
  185. return // No change to lease
  186. }
  187. l, err := ksm.nodeToLease(*n)
  188. if err != nil {
  189. log.Infof("Error turning node %q to lease: %v", n.ObjectMeta.Name, err)
  190. return
  191. }
  192. ksm.events <- subnet.Event{subnet.EventAdded, l}
  193. }
  194. func (ksm *kubeSubnetManager) GetNetworkConfig(ctx context.Context) (*subnet.Config, error) {
  195. return ksm.subnetConf, nil
  196. }
  197. func (ksm *kubeSubnetManager) AcquireLease(ctx context.Context, attrs *subnet.LeaseAttrs) (*subnet.Lease, error) {
  198. cachedNode, err := ksm.nodeStore.Get(ksm.nodeName)
  199. if err != nil {
  200. return nil, err
  201. }
  202. n := cachedNode.DeepCopy()
  203. if n.Spec.PodCIDR == "" {
  204. return nil, fmt.Errorf("node %q pod cidr not assigned", ksm.nodeName)
  205. }
  206. bd, err := attrs.BackendData.MarshalJSON()
  207. if err != nil {
  208. return nil, err
  209. }
  210. _, cidr, err := net.ParseCIDR(n.Spec.PodCIDR)
  211. if err != nil {
  212. return nil, err
  213. }
  214. if n.Annotations[ksm.annotations.BackendData] != string(bd) ||
  215. n.Annotations[ksm.annotations.BackendType] != attrs.BackendType ||
  216. n.Annotations[ksm.annotations.BackendPublicIP] != attrs.PublicIP.String() ||
  217. n.Annotations[ksm.annotations.SubnetKubeManaged] != "true" ||
  218. (n.Annotations[ksm.annotations.BackendPublicIPOverwrite] != "" && n.Annotations[ksm.annotations.BackendPublicIPOverwrite] != attrs.PublicIP.String()) {
  219. n.Annotations[ksm.annotations.BackendType] = attrs.BackendType
  220. n.Annotations[ksm.annotations.BackendData] = string(bd)
  221. if n.Annotations[ksm.annotations.BackendPublicIPOverwrite] != "" {
  222. if n.Annotations[ksm.annotations.BackendPublicIP] != n.Annotations[ksm.annotations.BackendPublicIPOverwrite] {
  223. log.Infof("Overriding public ip with '%s' from node annotation '%s'",
  224. n.Annotations[ksm.annotations.BackendPublicIPOverwrite],
  225. ksm.annotations.BackendPublicIPOverwrite)
  226. n.Annotations[ksm.annotations.BackendPublicIP] = n.Annotations[ksm.annotations.BackendPublicIPOverwrite]
  227. }
  228. } else {
  229. n.Annotations[ksm.annotations.BackendPublicIP] = attrs.PublicIP.String()
  230. }
  231. n.Annotations[ksm.annotations.SubnetKubeManaged] = "true"
  232. oldData, err := json.Marshal(cachedNode)
  233. if err != nil {
  234. return nil, err
  235. }
  236. newData, err := json.Marshal(n)
  237. if err != nil {
  238. return nil, err
  239. }
  240. patchBytes, err := strategicpatch.CreateTwoWayMergePatch(oldData, newData, v1.Node{})
  241. if err != nil {
  242. return nil, fmt.Errorf("failed to create patch for node %q: %v", ksm.nodeName, err)
  243. }
  244. _, err = ksm.client.CoreV1().Nodes().Patch(ctx, ksm.nodeName, types.StrategicMergePatchType, patchBytes, metav1.PatchOptions{}, "status")
  245. if err != nil {
  246. return nil, err
  247. }
  248. }
  249. err = ksm.setNodeNetworkUnavailableFalse(ctx)
  250. if err != nil {
  251. log.Errorf("Unable to set NetworkUnavailable to False for %q: %v", ksm.nodeName, err)
  252. }
  253. return &subnet.Lease{
  254. Subnet: ip.FromIPNet(cidr),
  255. Attrs: *attrs,
  256. Expiration: time.Now().Add(24 * time.Hour),
  257. }, nil
  258. }
  259. func (ksm *kubeSubnetManager) WatchLeases(ctx context.Context, cursor interface{}) (subnet.LeaseWatchResult, error) {
  260. select {
  261. case event := <-ksm.events:
  262. return subnet.LeaseWatchResult{
  263. Events: []subnet.Event{event},
  264. }, nil
  265. case <-ctx.Done():
  266. return subnet.LeaseWatchResult{}, context.Canceled
  267. }
  268. }
  269. func (ksm *kubeSubnetManager) Run(ctx context.Context) {
  270. log.Infof("Starting kube subnet manager")
  271. ksm.nodeController.Run(ctx.Done())
  272. }
  273. func (ksm *kubeSubnetManager) nodeToLease(n v1.Node) (l subnet.Lease, err error) {
  274. l.Attrs.PublicIP, err = ip.ParseIP4(n.Annotations[ksm.annotations.BackendPublicIP])
  275. if err != nil {
  276. return l, err
  277. }
  278. l.Attrs.BackendType = n.Annotations[ksm.annotations.BackendType]
  279. l.Attrs.BackendData = json.RawMessage(n.Annotations[ksm.annotations.BackendData])
  280. _, cidr, err := net.ParseCIDR(n.Spec.PodCIDR)
  281. if err != nil {
  282. return l, err
  283. }
  284. l.Subnet = ip.FromIPNet(cidr)
  285. return l, nil
  286. }
  287. // RenewLease: unimplemented
  288. func (ksm *kubeSubnetManager) RenewLease(ctx context.Context, lease *subnet.Lease) error {
  289. return ErrUnimplemented
  290. }
  291. func (ksm *kubeSubnetManager) WatchLease(ctx context.Context, sn ip.IP4Net, cursor interface{}) (subnet.LeaseWatchResult, error) {
  292. return subnet.LeaseWatchResult{}, ErrUnimplemented
  293. }
  294. func (ksm *kubeSubnetManager) Name() string {
  295. return fmt.Sprintf("Kubernetes Subnet Manager - %s", ksm.nodeName)
  296. }
  297. // Set Kubernetes NodeNetworkUnavailable to false when starting
  298. // https://kubernetes.io/docs/concepts/architecture/nodes/#condition
  299. func (ksm *kubeSubnetManager) setNodeNetworkUnavailableFalse(ctx context.Context) error {
  300. condition := v1.NodeCondition{
  301. Type: v1.NodeNetworkUnavailable,
  302. Status: v1.ConditionFalse,
  303. Reason: "FlannelIsUp",
  304. Message: "Flannel is running on this node",
  305. LastTransitionTime: metav1.Now(),
  306. LastHeartbeatTime: metav1.Now(),
  307. }
  308. raw, err := json.Marshal(&[]v1.NodeCondition{condition})
  309. if err != nil {
  310. return err
  311. }
  312. patch := []byte(fmt.Sprintf(`{"status":{"conditions":%s}}`, raw))
  313. _, err = ksm.client.CoreV1().Nodes().PatchStatus(ctx, ksm.nodeName, patch)
  314. return err
  315. }