k8saas: add support for Kubernetes endpoints
This commit is contained in:
parent
d29fc3d820
commit
f628842763
2
godo.go
2
godo.go
|
@ -65,6 +65,7 @@ type Client struct {
|
||||||
Certificates CertificatesService
|
Certificates CertificatesService
|
||||||
Firewalls FirewallsService
|
Firewalls FirewallsService
|
||||||
Projects ProjectsService
|
Projects ProjectsService
|
||||||
|
Kubernetes KubernetesService
|
||||||
|
|
||||||
// Optional function called after every successful request made to the DO APIs
|
// Optional function called after every successful request made to the DO APIs
|
||||||
onRequestCompleted RequestCompletionCallback
|
onRequestCompleted RequestCompletionCallback
|
||||||
|
@ -178,6 +179,7 @@ func NewClient(httpClient *http.Client) *Client {
|
||||||
c.Storage = &StorageServiceOp{client: c}
|
c.Storage = &StorageServiceOp{client: c}
|
||||||
c.StorageActions = &StorageActionsServiceOp{client: c}
|
c.StorageActions = &StorageActionsServiceOp{client: c}
|
||||||
c.Tags = &TagsServiceOp{client: c}
|
c.Tags = &TagsServiceOp{client: c}
|
||||||
|
c.Kubernetes = &KubernetesServiceOp{client: c}
|
||||||
|
|
||||||
return c
|
return c
|
||||||
}
|
}
|
||||||
|
|
|
@ -0,0 +1,381 @@
|
||||||
|
package godo
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"net/http"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
kubernetesBasePath = "/v2/kubernetes"
|
||||||
|
kubernetesClustersPath = kubernetesBasePath + "/clusters"
|
||||||
|
kubernetesOptionsPath = kubernetesBasePath + "/options"
|
||||||
|
)
|
||||||
|
|
||||||
|
// KubernetesService is an interface for interfacing with the kubernetes endpoints
|
||||||
|
// of the DigitalOcean API.
|
||||||
|
// See: https://developers.digitalocean.com/documentation/v2#kubernetes
|
||||||
|
type KubernetesService interface {
|
||||||
|
Create(context.Context, *KubernetesClusterCreateRequest) (*KubernetesCluster, *Response, error)
|
||||||
|
Get(context.Context, string) (*KubernetesCluster, *Response, error)
|
||||||
|
GetKubeConfig(context.Context, string) (*KubernetesClusterConfig, *Response, error)
|
||||||
|
List(context.Context, *ListOptions) ([]*KubernetesCluster, *Response, error)
|
||||||
|
Update(context.Context, string, *KubernetesClusterUpdateRequest) (*KubernetesCluster, *Response, error)
|
||||||
|
Delete(context.Context, string) (*Response, error)
|
||||||
|
|
||||||
|
CreateNodePool(ctx context.Context, clusterID string, req *KubernetesNodePoolCreateRequest) (*KubernetesNodePool, *Response, error)
|
||||||
|
GetNodePool(ctx context.Context, clusterID, poolID string) (*KubernetesNodePool, *Response, error)
|
||||||
|
ListNodePools(ctx context.Context, clusterID string, opts *ListOptions) ([]*KubernetesNodePool, *Response, error)
|
||||||
|
UpdateNodePool(ctx context.Context, clusterID, poolID string, req *KubernetesNodePoolUpdateRequest) (*KubernetesNodePool, *Response, error)
|
||||||
|
RecycleNodePoolNodes(ctx context.Context, clusterID, poolID string, req *KubernetesNodePoolRecycleNodesRequest) (*Response, error)
|
||||||
|
DeleteNodePool(ctx context.Context, clusterID, poolID string) (*Response, error)
|
||||||
|
|
||||||
|
GetOptions(context.Context) (*KubernetesOptions, *Response, error)
|
||||||
|
}
|
||||||
|
|
||||||
|
var _ KubernetesService = &KubernetesServiceOp{}
|
||||||
|
|
||||||
|
// KubernetesServiceOp handles communication with Kubernetes methods of the DigitalOcean API.
|
||||||
|
type KubernetesServiceOp struct {
|
||||||
|
client *Client
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesClusterCreateRequest represents a request to create a Kubernetes cluster.
|
||||||
|
type KubernetesClusterCreateRequest struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
RegionSlug string `json:"region,omitempty"`
|
||||||
|
VersionSlug string `json:"version,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
|
||||||
|
NodePools []*KubernetesNodePoolCreateRequest `json:"node_pools,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesClusterUpdateRequest represents a request to update a Kubernetes cluster.
|
||||||
|
type KubernetesClusterUpdateRequest struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNodePoolCreateRequest represents a request to create a node pool for a
|
||||||
|
// Kubernetes cluster.
|
||||||
|
type KubernetesNodePoolCreateRequest struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Size string `json:"size,omitempty"`
|
||||||
|
Count int `json:"count,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNodePoolUpdateRequest represents a request to update a node pool in a
|
||||||
|
// Kubernetes cluster.
|
||||||
|
type KubernetesNodePoolUpdateRequest struct {
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Count int `json:"count,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNodePoolRecycleNodesRequest represents a request to recycle a set of
|
||||||
|
// nodes in a node pool. This will recycle the nodes by ID.
|
||||||
|
type KubernetesNodePoolRecycleNodesRequest struct {
|
||||||
|
Nodes []string `json:"nodes,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesCluster represents a Kubernetes cluster.
|
||||||
|
type KubernetesCluster struct {
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
RegionSlug string `json:"region,omitempty"`
|
||||||
|
VersionSlug string `json:"version,omitempty"`
|
||||||
|
ClusterSubnet string `json:"cluster_subnet,omitempty"`
|
||||||
|
ServiceSubnet string `json:"service_subnet,omitempty"`
|
||||||
|
IPv4 string `json:"ipv4,omitempty"`
|
||||||
|
Endpoint string `json:"endpoint,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
|
||||||
|
NodePools []*KubernetesNodePool `json:"node_pools,omitempty"`
|
||||||
|
|
||||||
|
Status *KubernetesClusterStatus `json:"status,omitempty"`
|
||||||
|
CreatedAt time.Time `json:"created_at,omitempty"`
|
||||||
|
UpdatedAt time.Time `json:"updated_at,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesClusterStatus describes the status of a cluster.
|
||||||
|
type KubernetesClusterStatus struct {
|
||||||
|
State string `json:"state,omitempty"`
|
||||||
|
Message string `json:"message,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNodePool represents a node pool in a Kubernetes cluster.
|
||||||
|
type KubernetesNodePool struct {
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Size string `json:"size,omitempty"`
|
||||||
|
Count int `json:"count,omitempty"`
|
||||||
|
Tags []string `json:"tags,omitempty"`
|
||||||
|
|
||||||
|
Nodes []*KubernetesNode `json:"nodes,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNode represents a Node in a node pool in a Kubernetes cluster.
|
||||||
|
type KubernetesNode struct {
|
||||||
|
ID string `json:"id,omitempty"`
|
||||||
|
Name string `json:"name,omitempty"`
|
||||||
|
Status *KubernetesNodeStatus `json:"status,omitempty"`
|
||||||
|
|
||||||
|
CreatedAt time.Time `json:"created_at,omitempty"`
|
||||||
|
UpdatedAt time.Time `json:"updated_at,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesNodeStatus represents the status of a particular Node in a Kubernetes cluster.
|
||||||
|
type KubernetesNodeStatus struct {
|
||||||
|
State string `json:"state,omitempty"`
|
||||||
|
Message string `json:"message,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesOptions represents options available for creating Kubernetes clusters.
|
||||||
|
type KubernetesOptions struct {
|
||||||
|
Versions []*KubernetesVersion `json:"versions,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesVersion is a DigitalOcean Kubernetes release.
|
||||||
|
type KubernetesVersion struct {
|
||||||
|
Slug string `json:"slug,omitempty"`
|
||||||
|
KubernetesVersion string `json:"kubernetes_version,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type kubernetesClustersRoot struct {
|
||||||
|
Clusters []*KubernetesCluster `json:"kubernetes_clusters,omitempty"`
|
||||||
|
Links *Links `json:"links,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type kubernetesClusterRoot struct {
|
||||||
|
Cluster *KubernetesCluster `json:"kubernetes_cluster,omitempty"`
|
||||||
|
Links *Links `json:"links,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type kubernetesNodePoolRoot struct {
|
||||||
|
NodePool *KubernetesNodePool `json:"node_pool,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type kubernetesNodePoolsRoot struct {
|
||||||
|
NodePool []*KubernetesNodePool `json:"node_pools,omitempty"`
|
||||||
|
Links *Links `json:"links,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get retrieves the details of a Kubernetes cluster.
|
||||||
|
func (svc *KubernetesServiceOp) Get(ctx context.Context, clusterID string) (*KubernetesCluster, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s", kubernetesClustersPath, clusterID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesClusterRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.Cluster, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create creates a Kubernetes cluster.
|
||||||
|
func (svc *KubernetesServiceOp) Create(ctx context.Context, create *KubernetesClusterCreateRequest) (*KubernetesCluster, *Response, error) {
|
||||||
|
path := kubernetesClustersPath
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodPost, path, create)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesClusterRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.Cluster, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete deletes a Kubernetes cluster. There is no way to recover a cluster
|
||||||
|
// once it has been destroyed.
|
||||||
|
func (svc *KubernetesServiceOp) Delete(ctx context.Context, clusterID string) (*Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s", kubernetesClustersPath, clusterID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodDelete, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
resp, err := svc.client.Do(ctx, req, nil)
|
||||||
|
if err != nil {
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
return resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// List returns a list of the Kubernetes clusters visible with the caller's API token.
|
||||||
|
func (svc *KubernetesServiceOp) List(ctx context.Context, opts *ListOptions) ([]*KubernetesCluster, *Response, error) {
|
||||||
|
path := kubernetesClustersPath
|
||||||
|
path, err := addOptions(path, opts)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesClustersRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.Clusters, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// KubernetesClusterConfig is the content of a Kubernetes config file, which can be
|
||||||
|
// used to interact with your Kubernetes cluster using `kubectl`.
|
||||||
|
// See: https://kubernetes.io/docs/tasks/tools/install-kubectl/
|
||||||
|
type KubernetesClusterConfig struct {
|
||||||
|
KubeconfigYAML []byte
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetKubeConfig returns a Kubernetes config file for the specified cluster.
|
||||||
|
func (svc *KubernetesServiceOp) GetKubeConfig(ctx context.Context, clusterID string) (*KubernetesClusterConfig, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/kubeconfig", kubernetesClustersPath, clusterID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
configBytes := bytes.NewBuffer(nil)
|
||||||
|
resp, err := svc.client.Do(ctx, req, configBytes)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
res := &KubernetesClusterConfig{
|
||||||
|
KubeconfigYAML: configBytes.Bytes(),
|
||||||
|
}
|
||||||
|
return res, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update updates a Kubernetes cluster's properties.
|
||||||
|
func (svc *KubernetesServiceOp) Update(ctx context.Context, clusterID string, update *KubernetesClusterUpdateRequest) (*KubernetesCluster, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s", kubernetesClustersPath, clusterID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodPut, path, update)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesClusterRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.Cluster, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// CreateNodePool creates a new node pool in an existing Kubernetes cluster.
|
||||||
|
func (svc *KubernetesServiceOp) CreateNodePool(ctx context.Context, clusterID string, create *KubernetesNodePoolCreateRequest) (*KubernetesNodePool, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools", kubernetesClustersPath, clusterID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodPost, path, create)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesNodePoolRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.NodePool, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetNodePool retrieves an existing node pool in a Kubernetes cluster.
|
||||||
|
func (svc *KubernetesServiceOp) GetNodePool(ctx context.Context, clusterID, poolID string) (*KubernetesNodePool, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools/%s", kubernetesClustersPath, clusterID, poolID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesNodePoolRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.NodePool, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// ListNodePools lists all the node pools found in a Kubernetes cluster.
|
||||||
|
func (svc *KubernetesServiceOp) ListNodePools(ctx context.Context, clusterID string, opts *ListOptions) ([]*KubernetesNodePool, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools", kubernetesClustersPath, clusterID)
|
||||||
|
path, err := addOptions(path, opts)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesNodePoolsRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.NodePool, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// UpdateNodePool updates the details of an existing node pool.
|
||||||
|
func (svc *KubernetesServiceOp) UpdateNodePool(ctx context.Context, clusterID, poolID string, update *KubernetesNodePoolUpdateRequest) (*KubernetesNodePool, *Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools/%s", kubernetesClustersPath, clusterID, poolID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodPut, path, update)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesNodePoolRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.NodePool, resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// RecycleNodePoolNodes schedules nodes in a node pool for recycling.
|
||||||
|
func (svc *KubernetesServiceOp) RecycleNodePoolNodes(ctx context.Context, clusterID, poolID string, recycle *KubernetesNodePoolRecycleNodesRequest) (*Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools/%s/recycle", kubernetesClustersPath, clusterID, poolID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodPost, path, recycle)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
resp, err := svc.client.Do(ctx, req, nil)
|
||||||
|
if err != nil {
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
return resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// DeleteNodePool deletes a node pool, and subsequently all the nodes in that pool.
|
||||||
|
func (svc *KubernetesServiceOp) DeleteNodePool(ctx context.Context, clusterID, poolID string) (*Response, error) {
|
||||||
|
path := fmt.Sprintf("%s/%s/node_pools/%s", kubernetesClustersPath, clusterID, poolID)
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodDelete, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
resp, err := svc.client.Do(ctx, req, nil)
|
||||||
|
if err != nil {
|
||||||
|
return resp, err
|
||||||
|
}
|
||||||
|
return resp, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
type kubernetesOptionsRoot struct {
|
||||||
|
Options *KubernetesOptions `json:"options,omitempty"`
|
||||||
|
Links *Links `json:"links,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetOptions returns options about the Kubernetes service, such as the versions available for
|
||||||
|
// cluster creation.
|
||||||
|
func (svc *KubernetesServiceOp) GetOptions(ctx context.Context) (*KubernetesOptions, *Response, error) {
|
||||||
|
path := kubernetesOptionsPath
|
||||||
|
req, err := svc.client.NewRequest(ctx, http.MethodGet, path, nil)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, err
|
||||||
|
}
|
||||||
|
root := new(kubernetesOptionsRoot)
|
||||||
|
resp, err := svc.client.Do(ctx, req, root)
|
||||||
|
if err != nil {
|
||||||
|
return nil, resp, err
|
||||||
|
}
|
||||||
|
return root.Options, resp, nil
|
||||||
|
}
|
|
@ -0,0 +1,771 @@
|
||||||
|
package godo
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"net/http"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestKubernetesClusters_ListClusters(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := []*KubernetesCluster{
|
||||||
|
&KubernetesCluster{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
Name: "blablabla",
|
||||||
|
RegionSlug: "nyc1",
|
||||||
|
VersionSlug: "1.10.0-gen0",
|
||||||
|
ClusterSubnet: "10.244.0.0/16",
|
||||||
|
ServiceSubnet: "10.245.0.0/16",
|
||||||
|
IPv4: "",
|
||||||
|
Tags: []string(nil),
|
||||||
|
Status: &KubernetesClusterStatus{
|
||||||
|
State: "running",
|
||||||
|
},
|
||||||
|
NodePools: []*KubernetesNodePool{
|
||||||
|
{
|
||||||
|
ID: "1a17a012-cb31-4886-a787-deadbeef1191",
|
||||||
|
Name: "blablabla-1",
|
||||||
|
Size: "s-1vcpu-2gb",
|
||||||
|
Count: 2,
|
||||||
|
Nodes: []*KubernetesNode{
|
||||||
|
{
|
||||||
|
ID: "",
|
||||||
|
Name: "",
|
||||||
|
Status: &KubernetesNodeStatus{},
|
||||||
|
CreatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ID: "",
|
||||||
|
Name: "",
|
||||||
|
Status: &KubernetesNodeStatus{},
|
||||||
|
CreatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
CreatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
},
|
||||||
|
&KubernetesCluster{
|
||||||
|
ID: "deadbeef-dead-4aa5-beef-deadbeef347d",
|
||||||
|
Name: "antoine",
|
||||||
|
RegionSlug: "nyc1",
|
||||||
|
VersionSlug: "1.10.0-gen0",
|
||||||
|
ClusterSubnet: "10.244.0.0/16",
|
||||||
|
ServiceSubnet: "10.245.0.0/16",
|
||||||
|
IPv4: "1.2.3.4",
|
||||||
|
Status: &KubernetesClusterStatus{
|
||||||
|
State: "running",
|
||||||
|
},
|
||||||
|
NodePools: []*KubernetesNodePool{
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b3",
|
||||||
|
Name: "antoine-1",
|
||||||
|
Size: "s-1vcpu-2gb",
|
||||||
|
Count: 5,
|
||||||
|
Nodes: []*KubernetesNode{
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b1",
|
||||||
|
Name: "worker-393",
|
||||||
|
Status: &KubernetesNodeStatus{State: "running"},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b2",
|
||||||
|
Name: "worker-394",
|
||||||
|
Status: &KubernetesNodeStatus{State: "running"},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"kubernetes_clusters": [
|
||||||
|
{
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
"name": "blablabla",
|
||||||
|
"region": "nyc1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"cluster_subnet": "10.244.0.0/16",
|
||||||
|
"service_subnet": "10.245.0.0/16",
|
||||||
|
"ipv4": "",
|
||||||
|
"tags": null,
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "1a17a012-cb31-4886-a787-deadbeef1191",
|
||||||
|
"name": "blablabla-1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"size": "s-1vcpu-2gb",
|
||||||
|
"count": 2,
|
||||||
|
"tags": null,
|
||||||
|
"nodes": [
|
||||||
|
{
|
||||||
|
"id": "",
|
||||||
|
"name": "",
|
||||||
|
"status": {
|
||||||
|
"state": ""
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-21T08:44:38Z",
|
||||||
|
"updated_at": "2018-06-21T08:44:38Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "",
|
||||||
|
"name": "",
|
||||||
|
"status": {
|
||||||
|
"state": ""
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-21T08:44:38Z",
|
||||||
|
"updated_at": "2018-06-21T08:44:38Z"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"created_at": "2018-06-21T08:44:38Z",
|
||||||
|
"updated_at": "2018-06-21T08:44:38Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-4aa5-beef-deadbeef347d",
|
||||||
|
"name": "antoine",
|
||||||
|
"region": "nyc1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"cluster_subnet": "10.244.0.0/16",
|
||||||
|
"service_subnet": "10.245.0.0/16",
|
||||||
|
"ipv4": "1.2.3.4",
|
||||||
|
"tags": null,
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b3",
|
||||||
|
"name": "antoine-1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"size": "s-1vcpu-2gb",
|
||||||
|
"count": 5,
|
||||||
|
"tags": null,
|
||||||
|
"nodes": [
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b1",
|
||||||
|
"name": "worker-393",
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b2",
|
||||||
|
"name": "worker-394",
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.List(ctx, nil)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_Get(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
want := &KubernetesCluster{
|
||||||
|
ID: "deadbeef-dead-4aa5-beef-deadbeef347d",
|
||||||
|
Name: "antoine",
|
||||||
|
RegionSlug: "nyc1",
|
||||||
|
VersionSlug: "1.10.0-gen0",
|
||||||
|
ClusterSubnet: "10.244.0.0/16",
|
||||||
|
ServiceSubnet: "10.245.0.0/16",
|
||||||
|
IPv4: "1.2.3.4",
|
||||||
|
Status: &KubernetesClusterStatus{
|
||||||
|
State: "running",
|
||||||
|
},
|
||||||
|
NodePools: []*KubernetesNodePool{
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b3",
|
||||||
|
Name: "antoine-1",
|
||||||
|
Size: "s-1vcpu-2gb",
|
||||||
|
Count: 5,
|
||||||
|
Nodes: []*KubernetesNode{
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b1",
|
||||||
|
Name: "worker-393",
|
||||||
|
Status: &KubernetesNodeStatus{State: "running"},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ID: "deadbeef-dead-beef-dead-deadbeefb4b2",
|
||||||
|
Name: "worker-394",
|
||||||
|
Status: &KubernetesNodeStatus{State: "running"},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
CreatedAt: time.Date(2018, 6, 15, 7, 10, 23, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 15, 7, 11, 26, 0, time.UTC),
|
||||||
|
}
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"kubernetes_cluster": {
|
||||||
|
"id": "deadbeef-dead-4aa5-beef-deadbeef347d",
|
||||||
|
"name": "antoine",
|
||||||
|
"region": "nyc1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"cluster_subnet": "10.244.0.0/16",
|
||||||
|
"service_subnet": "10.245.0.0/16",
|
||||||
|
"ipv4": "1.2.3.4",
|
||||||
|
"tags": null,
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b3",
|
||||||
|
"name": "antoine-1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"size": "s-1vcpu-2gb",
|
||||||
|
"count": 5,
|
||||||
|
"tags": null,
|
||||||
|
"nodes": [
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b1",
|
||||||
|
"name": "worker-393",
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "deadbeef-dead-beef-dead-deadbeefb4b2",
|
||||||
|
"name": "worker-394",
|
||||||
|
"status": {
|
||||||
|
"state": "running"
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"created_at": "2018-06-15T07:10:23Z",
|
||||||
|
"updated_at": "2018-06-15T07:11:26Z"
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/deadbeef-dead-4aa5-beef-deadbeef347d", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
got, _, err := kubeSvc.Get(ctx, "deadbeef-dead-4aa5-beef-deadbeef347d")
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_GetKubeConfig(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
want := "some YAML"
|
||||||
|
blob := []byte(want)
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/deadbeef-dead-4aa5-beef-deadbeef347d/kubeconfig", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, want)
|
||||||
|
})
|
||||||
|
got, _, err := kubeSvc.GetKubeConfig(ctx, "deadbeef-dead-4aa5-beef-deadbeef347d")
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, blob, got.KubeconfigYAML)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_Create(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesCluster{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
Name: "antoine-test-cluster",
|
||||||
|
RegionSlug: "s2r1",
|
||||||
|
VersionSlug: "1.10.0-gen0",
|
||||||
|
ClusterSubnet: "10.244.0.0/16",
|
||||||
|
ServiceSubnet: "10.245.0.0/16",
|
||||||
|
Tags: []string{"cluster-tag-1", "cluster-tag-2"},
|
||||||
|
NodePools: []*KubernetesNodePool{
|
||||||
|
&KubernetesNodePool{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
Size: "s-1vcpu-1gb",
|
||||||
|
Count: 2,
|
||||||
|
Name: "pool-a",
|
||||||
|
Tags: []string{"tag-1"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
createRequest := &KubernetesClusterCreateRequest{
|
||||||
|
Name: want.Name,
|
||||||
|
RegionSlug: want.RegionSlug,
|
||||||
|
VersionSlug: want.VersionSlug,
|
||||||
|
Tags: want.Tags,
|
||||||
|
NodePools: []*KubernetesNodePoolCreateRequest{
|
||||||
|
&KubernetesNodePoolCreateRequest{
|
||||||
|
Size: want.NodePools[0].Size,
|
||||||
|
Count: want.NodePools[0].Count,
|
||||||
|
Name: want.NodePools[0].Name,
|
||||||
|
Tags: want.NodePools[0].Tags,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"kubernetes_cluster": {
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
"name": "antoine-test-cluster",
|
||||||
|
"region": "s2r1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"cluster_subnet": "10.244.0.0/16",
|
||||||
|
"service_subnet": "10.245.0.0/16",
|
||||||
|
"tags": [
|
||||||
|
"cluster-tag-1",
|
||||||
|
"cluster-tag-2"
|
||||||
|
],
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
"size": "s-1vcpu-1gb",
|
||||||
|
"count": 2,
|
||||||
|
"name": "pool-a",
|
||||||
|
"tags": [
|
||||||
|
"tag-1"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
v := new(KubernetesClusterCreateRequest)
|
||||||
|
err := json.NewDecoder(r.Body).Decode(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
testMethod(t, r, http.MethodPost)
|
||||||
|
require.Equal(t, v, createRequest)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.Create(ctx, createRequest)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_Update(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesCluster{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
Name: "antoine-test-cluster",
|
||||||
|
RegionSlug: "s2r1",
|
||||||
|
VersionSlug: "1.10.0-gen0",
|
||||||
|
ClusterSubnet: "10.244.0.0/16",
|
||||||
|
ServiceSubnet: "10.245.0.0/16",
|
||||||
|
Tags: []string{"cluster-tag-1", "cluster-tag-2"},
|
||||||
|
NodePools: []*KubernetesNodePool{
|
||||||
|
&KubernetesNodePool{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
Size: "s-1vcpu-1gb",
|
||||||
|
Count: 2,
|
||||||
|
Name: "pool-a",
|
||||||
|
Tags: []string{"tag-1"},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
updateRequest := &KubernetesClusterUpdateRequest{
|
||||||
|
Name: want.Name,
|
||||||
|
Tags: want.Tags,
|
||||||
|
}
|
||||||
|
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"kubernetes_cluster": {
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8f",
|
||||||
|
"name": "antoine-test-cluster",
|
||||||
|
"region": "s2r1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"cluster_subnet": "10.244.0.0/16",
|
||||||
|
"service_subnet": "10.245.0.0/16",
|
||||||
|
"tags": [
|
||||||
|
"cluster-tag-1",
|
||||||
|
"cluster-tag-2"
|
||||||
|
],
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
"size": "s-1vcpu-1gb",
|
||||||
|
"count": 2,
|
||||||
|
"name": "pool-a",
|
||||||
|
"tags": [
|
||||||
|
"tag-1"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
v := new(KubernetesClusterUpdateRequest)
|
||||||
|
err := json.NewDecoder(r.Body).Decode(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
testMethod(t, r, http.MethodPut)
|
||||||
|
require.Equal(t, v, updateRequest)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.Update(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", updateRequest)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_Destroy(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/deadbeef-dead-4aa5-beef-deadbeef347d", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodDelete)
|
||||||
|
})
|
||||||
|
|
||||||
|
_, err := kubeSvc.Delete(ctx, "deadbeef-dead-4aa5-beef-deadbeef347d")
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_CreateNodePool(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesNodePool{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
Size: "s-1vcpu-1gb",
|
||||||
|
Count: 2,
|
||||||
|
Name: "pool-a",
|
||||||
|
Tags: []string{"tag-1"},
|
||||||
|
}
|
||||||
|
createRequest := &KubernetesNodePoolCreateRequest{
|
||||||
|
Size: want.Size,
|
||||||
|
Count: want.Count,
|
||||||
|
Name: want.Name,
|
||||||
|
Tags: want.Tags,
|
||||||
|
}
|
||||||
|
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"node_pool": {
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
"size": "s-1vcpu-1gb",
|
||||||
|
"count": 2,
|
||||||
|
"name": "pool-a",
|
||||||
|
"tags": [
|
||||||
|
"tag-1"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f/node_pools", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
v := new(KubernetesNodePoolCreateRequest)
|
||||||
|
err := json.NewDecoder(r.Body).Decode(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
testMethod(t, r, http.MethodPost)
|
||||||
|
require.Equal(t, v, createRequest)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.CreateNodePool(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", createRequest)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_GetNodePool(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesNodePool{
|
||||||
|
ID: "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
Size: "s-1vcpu-1gb",
|
||||||
|
Count: 2,
|
||||||
|
Name: "pool-a",
|
||||||
|
Tags: []string{"tag-1"},
|
||||||
|
}
|
||||||
|
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"node_pool": {
|
||||||
|
"id": "8d91899c-0739-4a1a-acc5-deadbeefbb8a",
|
||||||
|
"size": "s-1vcpu-1gb",
|
||||||
|
"count": 2,
|
||||||
|
"name": "pool-a",
|
||||||
|
"tags": [
|
||||||
|
"tag-1"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f/node_pools/8d91899c-0739-4a1a-acc5-deadbeefbb8a", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.GetNodePool(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", "8d91899c-0739-4a1a-acc5-deadbeefbb8a")
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_ListNodePools(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := []*KubernetesNodePool{
|
||||||
|
{
|
||||||
|
ID: "1a17a012-cb31-4886-a787-deadbeef1191",
|
||||||
|
Name: "blablabla-1",
|
||||||
|
Size: "s-1vcpu-2gb",
|
||||||
|
Count: 2,
|
||||||
|
Nodes: []*KubernetesNode{
|
||||||
|
{
|
||||||
|
ID: "",
|
||||||
|
Name: "",
|
||||||
|
Status: &KubernetesNodeStatus{},
|
||||||
|
CreatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
ID: "",
|
||||||
|
Name: "",
|
||||||
|
Status: &KubernetesNodeStatus{},
|
||||||
|
CreatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
UpdatedAt: time.Date(2018, 6, 21, 8, 44, 38, 0, time.UTC),
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"node_pools": [
|
||||||
|
{
|
||||||
|
"id": "1a17a012-cb31-4886-a787-deadbeef1191",
|
||||||
|
"name": "blablabla-1",
|
||||||
|
"version": "1.10.0-gen0",
|
||||||
|
"size": "s-1vcpu-2gb",
|
||||||
|
"count": 2,
|
||||||
|
"tags": null,
|
||||||
|
"nodes": [
|
||||||
|
{
|
||||||
|
"id": "",
|
||||||
|
"name": "",
|
||||||
|
"status": {
|
||||||
|
"state": ""
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-21T08:44:38Z",
|
||||||
|
"updated_at": "2018-06-21T08:44:38Z"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "",
|
||||||
|
"name": "",
|
||||||
|
"status": {
|
||||||
|
"state": ""
|
||||||
|
},
|
||||||
|
"created_at": "2018-06-21T08:44:38Z",
|
||||||
|
"updated_at": "2018-06-21T08:44:38Z"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f/node_pools", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.ListNodePools(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", nil)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_UpdateNodePool(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesNodePool{
|
||||||
|
ID: "8d91899c-nodepool-4a1a-acc5-deadbeefbb8a",
|
||||||
|
Name: "a better name",
|
||||||
|
Size: "s-1vcpu-1gb",
|
||||||
|
Count: 4,
|
||||||
|
Tags: []string{"tag-1", "tag-2"},
|
||||||
|
}
|
||||||
|
updateRequest := &KubernetesNodePoolUpdateRequest{
|
||||||
|
Name: "a better name",
|
||||||
|
Count: 4,
|
||||||
|
Tags: []string{"tag-1", "tag-2"},
|
||||||
|
}
|
||||||
|
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"node_pool": {
|
||||||
|
"id": "8d91899c-nodepool-4a1a-acc5-deadbeefbb8a",
|
||||||
|
"size": "s-1vcpu-1gb",
|
||||||
|
"count": 4,
|
||||||
|
"name": "a better name",
|
||||||
|
"tags": [
|
||||||
|
"tag-1", "tag-2"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f/node_pools/8d91899c-nodepool-4a1a-acc5-deadbeefbb8a", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
v := new(KubernetesNodePoolUpdateRequest)
|
||||||
|
err := json.NewDecoder(r.Body).Decode(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
testMethod(t, r, http.MethodPut)
|
||||||
|
require.Equal(t, v, updateRequest)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.UpdateNodePool(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", "8d91899c-nodepool-4a1a-acc5-deadbeefbb8a", updateRequest)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_DeleteNodePool(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/deadbeef-dead-4aa5-beef-deadbeef347d/node_pools/8d91899c-nodepool-4a1a-acc5-deadbeefbb8a", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodDelete)
|
||||||
|
})
|
||||||
|
|
||||||
|
_, err := kubeSvc.DeleteNodePool(ctx, "deadbeef-dead-4aa5-beef-deadbeef347d", "8d91899c-nodepool-4a1a-acc5-deadbeefbb8a")
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesClusters_RecycleNodePoolNodes(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
recycleRequest := &KubernetesNodePoolRecycleNodesRequest{
|
||||||
|
Nodes: []string{"node1", "node2"},
|
||||||
|
}
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/clusters/8d91899c-0739-4a1a-acc5-deadbeefbb8f/node_pools/8d91899c-nodepool-4a1a-acc5-deadbeefbb8a/recycle", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
v := new(KubernetesNodePoolRecycleNodesRequest)
|
||||||
|
err := json.NewDecoder(r.Body).Decode(v)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
testMethod(t, r, http.MethodPost)
|
||||||
|
require.Equal(t, v, recycleRequest)
|
||||||
|
})
|
||||||
|
|
||||||
|
_, err := kubeSvc.RecycleNodePoolNodes(ctx, "8d91899c-0739-4a1a-acc5-deadbeefbb8f", "8d91899c-nodepool-4a1a-acc5-deadbeefbb8a", recycleRequest)
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestKubernetesVersions_List(t *testing.T) {
|
||||||
|
setup()
|
||||||
|
defer teardown()
|
||||||
|
|
||||||
|
kubeSvc := client.Kubernetes
|
||||||
|
|
||||||
|
want := &KubernetesOptions{
|
||||||
|
Versions: []*KubernetesVersion{
|
||||||
|
{Slug: "1.10.0-gen0", KubernetesVersion: "1.10.0"},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jBlob := `
|
||||||
|
{
|
||||||
|
"options": {
|
||||||
|
"versions": [
|
||||||
|
{
|
||||||
|
"slug": "1.10.0-gen0",
|
||||||
|
"kubernetes_version": "1.10.0"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}`
|
||||||
|
|
||||||
|
mux.HandleFunc("/v2/kubernetes/options", func(w http.ResponseWriter, r *http.Request) {
|
||||||
|
testMethod(t, r, http.MethodGet)
|
||||||
|
fmt.Fprint(w, jBlob)
|
||||||
|
})
|
||||||
|
|
||||||
|
got, _, err := kubeSvc.GetOptions(ctx)
|
||||||
|
require.NoError(t, err)
|
||||||
|
require.Equal(t, want, got)
|
||||||
|
}
|
Loading…
Reference in New Issue