2021-02-03 19:32:29 +00:00
// Copyright 2020-2021 the Pinniped contributors. All Rights Reserved.
2021-01-22 18:00:27 +00:00
// SPDX-License-Identifier: Apache-2.0
package integration
import (
2021-03-06 00:14:45 +00:00
"bytes"
2021-01-22 18:00:27 +00:00
"context"
2021-04-20 15:19:58 +00:00
"crypto/ecdsa"
"crypto/elliptic"
"crypto/rand"
2021-03-10 00:58:44 +00:00
"crypto/x509"
2021-04-20 15:19:58 +00:00
"crypto/x509/pkix"
2021-03-03 20:53:23 +00:00
"encoding/base64"
2021-03-10 00:58:44 +00:00
"encoding/json"
2021-03-10 18:30:06 +00:00
"encoding/pem"
2021-01-22 18:00:27 +00:00
"fmt"
2021-03-06 00:14:45 +00:00
"io/ioutil"
2021-01-22 18:00:27 +00:00
"net/http"
"net/url"
2021-03-06 00:14:45 +00:00
"os"
"os/exec"
"path/filepath"
"strings"
2021-03-11 20:10:16 +00:00
"sync"
2021-01-22 18:00:27 +00:00
"testing"
"time"
2021-03-10 23:49:09 +00:00
"github.com/gorilla/websocket"
2021-03-11 18:04:36 +00:00
"github.com/stretchr/testify/assert"
2021-01-22 18:00:27 +00:00
"github.com/stretchr/testify/require"
2021-03-11 23:49:24 +00:00
"golang.org/x/net/http2"
2021-04-09 21:52:53 +00:00
authenticationv1 "k8s.io/api/authentication/v1"
authorizationv1 "k8s.io/api/authorization/v1"
2021-04-20 15:19:58 +00:00
certificatesv1 "k8s.io/api/certificates/v1"
certificatesv1beta1 "k8s.io/api/certificates/v1beta1"
2021-01-22 18:00:27 +00:00
corev1 "k8s.io/api/core/v1"
2021-02-23 01:23:11 +00:00
rbacv1 "k8s.io/api/rbac/v1"
2021-02-23 18:38:02 +00:00
k8serrors "k8s.io/apimachinery/pkg/api/errors"
2021-01-22 18:00:27 +00:00
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
2021-02-23 01:23:11 +00:00
"k8s.io/apimachinery/pkg/labels"
2021-02-23 18:38:02 +00:00
"k8s.io/apimachinery/pkg/types"
2021-03-10 00:58:44 +00:00
"k8s.io/apimachinery/pkg/watch"
2021-04-09 21:52:53 +00:00
"k8s.io/apiserver/pkg/authentication/authenticator"
"k8s.io/apiserver/pkg/authentication/request/bearertoken"
"k8s.io/apiserver/pkg/authentication/serviceaccount"
2021-02-23 01:23:11 +00:00
k8sinformers "k8s.io/client-go/informers"
2021-01-22 18:00:27 +00:00
"k8s.io/client-go/kubernetes"
"k8s.io/client-go/rest"
2021-04-09 21:52:53 +00:00
"k8s.io/client-go/transport"
2021-04-20 15:19:58 +00:00
"k8s.io/client-go/util/cert"
"k8s.io/client-go/util/certificate/csr"
2021-04-09 21:52:53 +00:00
"k8s.io/client-go/util/keyutil"
2021-02-12 01:22:47 +00:00
"sigs.k8s.io/yaml"
2021-01-22 18:00:27 +00:00
2021-03-15 18:42:57 +00:00
conciergev1alpha "go.pinniped.dev/generated/latest/apis/concierge/config/v1alpha1"
2021-03-11 00:57:15 +00:00
identityv1alpha1 "go.pinniped.dev/generated/latest/apis/concierge/identity/v1alpha1"
2021-03-10 18:30:06 +00:00
loginv1alpha1 "go.pinniped.dev/generated/latest/apis/concierge/login/v1alpha1"
2021-03-27 02:32:33 +00:00
pinnipedconciergeclientset "go.pinniped.dev/generated/latest/client/concierge/clientset/versioned"
2021-02-12 01:22:47 +00:00
"go.pinniped.dev/internal/concierge/impersonator"
2021-04-09 21:52:53 +00:00
"go.pinniped.dev/internal/httputil/roundtripper"
2021-03-11 00:57:15 +00:00
"go.pinniped.dev/internal/kubeclient"
2021-03-06 00:14:45 +00:00
"go.pinniped.dev/internal/testutil"
2021-01-22 18:00:27 +00:00
"go.pinniped.dev/test/library"
)
2021-03-11 20:10:16 +00:00
// syncBuffer wraps bytes.Buffer with a mutex so we don't have races in our test code.
type syncBuffer struct {
buf bytes . Buffer
mu sync . Mutex
}
func ( sb * syncBuffer ) String ( ) string {
sb . mu . Lock ( )
defer sb . mu . Unlock ( )
return sb . buf . String ( )
}
func ( sb * syncBuffer ) Read ( b [ ] byte ) ( int , error ) {
sb . mu . Lock ( )
defer sb . mu . Unlock ( )
return sb . buf . Read ( b )
}
func ( sb * syncBuffer ) Write ( b [ ] byte ) ( int , error ) {
sb . mu . Lock ( )
defer sb . mu . Unlock ( )
return sb . buf . Write ( b )
}
2021-02-25 22:40:02 +00:00
// Note that this test supports being run on all of our integration test cluster types:
2021-03-26 16:28:42 +00:00
// - TKGS acceptance (long-running) cluster: auto mode will choose disabled, supports LBs, does not have squid.
// - GKE acceptance (long-running) cluster: auto will choose enabled, support LBs, does not have squid.
// - kind: auto mode will choose disabled, does not support LBs, has squid.
// - GKE ephemeral clusters: auto mode will choose enabled, supports LBs, has squid.
// - AKS ephemeral clusters: auto mode will choose enabled, supports LBs, has squid.
// - EKS ephemeral clusters: auto mode will choose enabled, supports LBs, has squid.
2021-03-10 18:30:06 +00:00
func TestImpersonationProxy ( t * testing . T ) { //nolint:gocyclo // yeah, it's complex.
2021-01-22 18:00:27 +00:00
env := library . IntegrationEnv ( t )
2021-03-26 16:49:49 +00:00
impersonatorShouldHaveStartedAutomaticallyByDefault := ! env . HasCapability ( library . ClusterSigningKeyIsAvailable )
2021-03-26 16:28:42 +00:00
clusterSupportsLoadBalancers := env . HasCapability ( library . HasExternalLoadBalancerProvider )
2021-03-03 20:53:23 +00:00
ctx , cancel := context . WithTimeout ( context . Background ( ) , 20 * time . Minute )
2021-01-22 18:00:27 +00:00
defer cancel ( )
// Create a client using the admin kubeconfig.
2021-02-12 01:22:47 +00:00
adminClient := library . NewKubernetesClientset ( t )
2021-03-03 20:53:23 +00:00
adminConciergeClient := library . NewConciergeClientset ( t )
2021-01-22 18:00:27 +00:00
2021-03-11 00:57:15 +00:00
// Create a WebhookAuthenticator and prepare a TokenCredentialRequestSpec using the authenticator for use later.
credentialRequestSpecWithWorkingCredentials := loginv1alpha1 . TokenCredentialRequestSpec {
Token : env . TestUser . Token ,
Authenticator : library . CreateTestWebhookAuthenticator ( ctx , t ) ,
}
2021-01-22 18:00:27 +00:00
2021-02-25 22:40:02 +00:00
// The address of the ClusterIP service that points at the impersonation proxy's port (used when there is no load balancer).
2021-02-25 18:27:19 +00:00
proxyServiceEndpoint := fmt . Sprintf ( "%s-proxy.%s.svc.cluster.local" , env . ConciergeAppName , env . ConciergeNamespace )
2021-01-22 18:00:27 +00:00
2021-04-09 21:52:53 +00:00
var (
mostRecentTokenCredentialRequestResponse * loginv1alpha1 . TokenCredentialRequest
mostRecentTokenCredentialRequestResponseLock sync . Mutex
)
2021-03-27 02:32:33 +00:00
refreshCredential := func ( t * testing . T , impersonationProxyURL string , impersonationProxyCACertPEM [ ] byte ) * loginv1alpha1 . ClusterCredential {
2021-04-09 21:52:53 +00:00
mostRecentTokenCredentialRequestResponseLock . Lock ( )
defer mostRecentTokenCredentialRequestResponseLock . Unlock ( )
2021-03-11 00:57:15 +00:00
if mostRecentTokenCredentialRequestResponse == nil || credentialAlmostExpired ( t , mostRecentTokenCredentialRequestResponse ) {
2021-03-10 18:30:06 +00:00
var err error
// Make a TokenCredentialRequest. This can either return a cert signed by the Kube API server's CA (e.g. on kind)
// or a cert signed by the impersonator's signing CA (e.g. on GKE). Either should be accepted by the impersonation
// proxy server as a valid authentication.
//
// However, we issue short-lived certs, so this cert will only be valid for a few minutes.
// Cache it until it is almost expired and then refresh it whenever it is close to expired.
2021-03-27 02:32:33 +00:00
//
// Also, use an anonymous client which goes through the impersonation proxy to make the request because that's
// what would normally happen when a user is using a kubeconfig where the server is the impersonation proxy,
// so it more closely simulates the normal use case, and also because we want this to work on AKS clusters
// which do not allow anonymous requests.
2021-04-09 21:52:53 +00:00
client := newAnonymousImpersonationProxyClient ( t , impersonationProxyURL , impersonationProxyCACertPEM , nil ) . PinnipedConcierge
2021-03-30 18:47:25 +00:00
require . Eventually ( t , func ( ) bool {
mostRecentTokenCredentialRequestResponse , err = createTokenCredentialRequest ( credentialRequestSpecWithWorkingCredentials , client )
if err != nil {
t . Logf ( "failed to make TokenCredentialRequest: %s" , library . Sdump ( err ) )
return false
}
return true
} , 5 * time . Minute , 5 * time . Second )
2021-03-10 18:30:06 +00:00
2021-03-11 00:57:15 +00:00
require . Nil ( t , mostRecentTokenCredentialRequestResponse . Status . Message ,
"expected no error message but got: %s" , library . Sdump ( mostRecentTokenCredentialRequestResponse . Status . Message ) )
require . NotEmpty ( t , mostRecentTokenCredentialRequestResponse . Status . Credential . ClientCertificateData )
require . NotEmpty ( t , mostRecentTokenCredentialRequestResponse . Status . Credential . ClientKeyData )
2021-03-10 18:30:06 +00:00
// At the moment the credential request should not have returned a token. In the future, if we make it return
// tokens, we should revisit this test's rest config below.
2021-03-11 00:57:15 +00:00
require . Empty ( t , mostRecentTokenCredentialRequestResponse . Status . Credential . Token )
2021-03-02 01:53:26 +00:00
}
2021-03-11 00:57:15 +00:00
return mostRecentTokenCredentialRequestResponse . Status . Credential
2021-03-02 01:53:26 +00:00
}
2021-03-02 17:31:24 +00:00
oldConfigMap , err := adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyConfigMapName ( env ) , metav1 . GetOptions { } )
2021-02-25 22:40:02 +00:00
if ! k8serrors . IsNotFound ( err ) {
require . NoError ( t , err ) // other errors aside from NotFound are unexpected
2021-02-18 23:58:27 +00:00
t . Logf ( "stashing a pre-existing configmap %s" , oldConfigMap . Name )
2021-03-02 17:31:24 +00:00
require . NoError ( t , adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Delete ( ctx , impersonationProxyConfigMapName ( env ) , metav1 . DeleteOptions { } ) )
2021-02-17 18:32:29 +00:00
}
2021-03-04 00:23:07 +00:00
// At the end of the test, clean up the ConfigMap.
t . Cleanup ( func ( ) {
2021-04-09 21:52:53 +00:00
ctx , cancel := context . WithTimeout ( context . Background ( ) , 2 * time . Minute )
2021-03-04 00:23:07 +00:00
defer cancel ( )
// Delete any version that was created by this test.
t . Logf ( "cleaning up configmap at end of test %s" , impersonationProxyConfigMapName ( env ) )
2021-03-15 16:37:19 +00:00
err := adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Delete ( ctx , impersonationProxyConfigMapName ( env ) , metav1 . DeleteOptions { } )
2021-03-04 00:23:07 +00:00
if ! k8serrors . IsNotFound ( err ) {
require . NoError ( t , err ) // only not found errors are acceptable
}
// Only recreate it if it already existed at the start of this test.
if len ( oldConfigMap . Data ) != 0 {
t . Log ( oldConfigMap )
oldConfigMap . UID = "" // cant have a UID yet
oldConfigMap . ResourceVersion = ""
t . Logf ( "restoring a pre-existing configmap %s" , oldConfigMap . Name )
_ , err = adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Create ( ctx , oldConfigMap , metav1 . CreateOptions { } )
require . NoError ( t , err )
}
2021-03-19 16:54:37 +00:00
// If we are running on an environment that has a load balancer, expect that the
// CredentialIssuer will be updated eventually with a successful impersonation proxy frontend.
// We do this to ensure that future tests that use the impersonation proxy (e.g.,
// TestE2EFullIntegration) will start with a known-good state.
2021-03-26 16:28:42 +00:00
if clusterSupportsLoadBalancers {
2021-03-19 16:54:37 +00:00
performImpersonatorDiscovery ( ctx , t , env , adminConciergeClient )
}
2021-03-04 00:23:07 +00:00
} )
2021-02-12 01:22:47 +00:00
2021-03-26 16:28:42 +00:00
// Done with set-up and ready to get started with the test. There are several states that we could be in at
// this point depending on the capabilities of the cluster under test. We handle each possible case here.
switch {
case impersonatorShouldHaveStartedAutomaticallyByDefault && clusterSupportsLoadBalancers :
// Auto mode should have decided that the impersonator will run and should have started a load balancer,
// and we will be able to use the load balancer to access the impersonator. (e.g. GKE, AKS, EKS)
2021-03-03 20:53:23 +00:00
// Check that load balancer has been automatically created by the impersonator's "auto" mode.
2021-02-25 22:40:02 +00:00
library . RequireEventuallyWithoutError ( t , func ( ) ( bool , error ) {
2021-03-02 17:31:24 +00:00
return hasImpersonationProxyLoadBalancerService ( ctx , env , adminClient )
2021-03-02 23:49:01 +00:00
} , 30 * time . Second , 500 * time . Millisecond )
2021-03-26 16:28:42 +00:00
case impersonatorShouldHaveStartedAutomaticallyByDefault && ! clusterSupportsLoadBalancers :
t . Fatal ( "None of the clusters types that we currently test against should automatically" +
"enable the impersonation proxy without also supporting load balancers. If we add such a" +
"cluster type in the future then we should enhance this test." )
case ! impersonatorShouldHaveStartedAutomaticallyByDefault && clusterSupportsLoadBalancers :
// Auto mode should have decided that the impersonator will be disabled. We need to manually enable it.
// The cluster supports load balancers so we should enable it and let the impersonator create a load balancer
// automatically. (e.g. TKGS)
// The CredentialIssuer's strategies array should have been updated to include an unsuccessful impersonation
// strategy saying that it was automatically disabled.
requireDisabledStrategy ( ctx , t , env , adminConciergeClient )
// Create configuration to make the impersonation proxy turn on with no endpoint (i.e. automatically create a load balancer).
configMap := impersonationProxyConfigMapForConfig ( t , env , impersonator . Config { Mode : impersonator . ModeEnabled } )
t . Logf ( "creating configmap %s" , configMap . Name )
_ , err = adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Create ( ctx , & configMap , metav1 . CreateOptions { } )
require . NoError ( t , err )
default :
// Auto mode should have decided that the impersonator will be disabled. We need to manually enable it.
// However, the cluster does not support load balancers so we should enable it without a load balancer
// and use squid to make requests. (e.g. kind)
2021-03-26 21:28:33 +00:00
if env . Proxy == "" {
t . Skip ( "test cluster does not support load balancers but also doesn't have a squid proxy... " +
2021-02-25 22:40:02 +00:00
"this is not a supported configuration for test clusters" )
2021-03-26 21:28:33 +00:00
}
2021-02-25 22:40:02 +00:00
2021-03-03 20:53:23 +00:00
// Check that no load balancer has been created by the impersonator's "auto" mode.
2021-02-25 22:40:02 +00:00
library . RequireNeverWithoutError ( t , func ( ) ( bool , error ) {
2021-03-02 17:31:24 +00:00
return hasImpersonationProxyLoadBalancerService ( ctx , env , adminClient )
2021-02-12 01:22:47 +00:00
} , 10 * time . Second , 500 * time . Millisecond )
2021-02-25 22:40:02 +00:00
// Check that we can't use the impersonation proxy to execute kubectl commands yet.
2021-04-09 21:52:53 +00:00
_ , err = impersonationProxyViaSquidKubeClientWithoutCredential ( t , proxyServiceEndpoint ) . CoreV1 ( ) . Namespaces ( ) . List ( ctx , metav1 . ListOptions { } )
2021-03-18 20:34:30 +00:00
isErr , message := isServiceUnavailableViaSquidError ( err , proxyServiceEndpoint )
require . Truef ( t , isErr , "wanted error %q to be service unavailable via squid error, but: %s" , err , message )
2021-02-12 01:22:47 +00:00
2021-03-03 20:53:23 +00:00
// Create configuration to make the impersonation proxy turn on with a hard coded endpoint (without a load balancer).
2021-03-26 16:28:42 +00:00
configMap := impersonationProxyConfigMapForConfig ( t , env , impersonator . Config {
2021-02-12 01:22:47 +00:00
Mode : impersonator . ModeEnabled ,
2021-02-25 18:27:19 +00:00
Endpoint : proxyServiceEndpoint ,
2021-02-12 01:22:47 +00:00
} )
2021-02-18 23:58:27 +00:00
t . Logf ( "creating configmap %s" , configMap . Name )
2021-02-12 01:22:47 +00:00
_ , err = adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Create ( ctx , & configMap , metav1 . CreateOptions { } )
require . NoError ( t , err )
}
2021-03-03 20:53:23 +00:00
// At this point the impersonator should be starting/running. When it is ready, the CredentialIssuer's
// strategies array should be updated to include a successful impersonation strategy which can be used
// to discover the impersonator's URL and CA certificate. Until it has finished starting, it may not be included
// in the strategies array or it may be included in an error state. It can be in an error state for
// awhile when it is waiting for the load balancer to be assigned an ip/hostname.
impersonationProxyURL , impersonationProxyCACertPEM := performImpersonatorDiscovery ( ctx , t , env , adminConciergeClient )
2021-03-26 16:28:42 +00:00
if ! clusterSupportsLoadBalancers {
2021-03-03 20:53:23 +00:00
// In this case, we specified the endpoint in the configmap, so check that it was reported correctly in the CredentialIssuer.
require . Equal ( t , "https://" + proxyServiceEndpoint , impersonationProxyURL )
2021-03-10 18:30:06 +00:00
}
// Because our credentials expire so quickly, we'll always use a new client, to give us a chance to refresh our
// credentials before they expire. Create a closure to capture the arguments to newImpersonationProxyClient
// so we don't have to keep repeating them.
// This client performs TLS checks, so it also provides test coverage that the impersonation proxy server is generating TLS certs correctly.
2021-03-27 02:32:33 +00:00
impersonationProxyKubeClient := func ( t * testing . T ) kubernetes . Interface {
2021-04-09 21:52:53 +00:00
return newImpersonationProxyClient ( t , impersonationProxyURL , impersonationProxyCACertPEM , nil , refreshCredential ) . Kubernetes
2021-02-25 22:40:02 +00:00
}
2021-03-18 17:44:37 +00:00
2021-03-15 21:34:09 +00:00
t . Run ( "positive tests" , func ( t * testing . T ) {
2021-03-16 23:57:28 +00:00
// Create an RBAC rule to allow this user to read/write everything.
library . CreateTestClusterRoleBinding ( t ,
rbacv1 . Subject { Kind : rbacv1 . UserKind , APIGroup : rbacv1 . GroupName , Name : env . TestUser . ExpectedUsername } ,
rbacv1 . RoleRef { Kind : "ClusterRole" , APIGroup : rbacv1 . GroupName , Name : "edit" } ,
)
2021-03-18 17:44:37 +00:00
// Wait for the above RBAC rule to take effect.
2021-04-09 21:52:53 +00:00
library . WaitForUserToHaveAccess ( t , env . TestUser . ExpectedUsername , [ ] string { } , & authorizationv1 . ResourceAttributes {
2021-03-18 17:44:37 +00:00
Verb : "get" , Group : "" , Version : "v1" , Resource : "namespaces" ,
} )
2021-03-16 23:57:28 +00:00
// Get pods in concierge namespace and pick one.
// this is for tests that require performing actions against a running pod. We use the concierge pod because we already have it handy.
// We want to make sure it's a concierge pod (not cert agent), because we need to be able to "exec echo" and port-forward a running port.
pods , err := adminClient . CoreV1 ( ) . Pods ( env . ConciergeNamespace ) . List ( ctx , metav1 . ListOptions { } )
require . NoError ( t , err )
require . Greater ( t , len ( pods . Items ) , 0 )
var conciergePod * corev1 . Pod
for _ , pod := range pods . Items {
pod := pod
if ! strings . Contains ( pod . Name , "kube-cert-agent" ) {
conciergePod = & pod
}
}
require . NotNil ( t , conciergePod , "could not find a concierge pod" )
2021-03-15 21:34:09 +00:00
// Test that the user can perform basic actions through the client with their username and group membership
// influencing RBAC checks correctly.
2021-01-22 18:00:27 +00:00
t . Run (
2021-03-15 21:34:09 +00:00
"access as user" ,
2021-03-27 02:32:33 +00:00
library . AccessAsUserTest ( ctx , env . TestUser . ExpectedUsername , impersonationProxyKubeClient ( t ) ) ,
2021-01-22 18:00:27 +00:00
)
2021-03-15 21:34:09 +00:00
for _ , group := range env . TestUser . ExpectedGroups {
group := group
t . Run (
"access as group " + group ,
2021-03-27 02:32:33 +00:00
library . AccessAsGroupTest ( ctx , group , impersonationProxyKubeClient ( t ) ) ,
2021-03-15 21:34:09 +00:00
)
}
2021-02-12 01:22:47 +00:00
2021-03-29 16:30:20 +00:00
t . Run ( "kubectl port-forward and keeping the connection open for over a minute (non-idle)" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-25 23:57:37 +00:00
kubeconfigPath , envVarsWithProxy , _ := getImpersonationKubeconfig ( t , env , impersonationProxyURL , impersonationProxyCACertPEM , credentialRequestSpecWithWorkingCredentials . Authenticator )
2021-03-15 19:28:53 +00:00
2021-03-19 16:36:05 +00:00
// Run the kubectl port-forward command.
2021-03-15 21:34:09 +00:00
timeout , cancelFunc := context . WithTimeout ( ctx , 2 * time . Minute )
defer cancelFunc ( )
2021-04-09 21:52:53 +00:00
portForwardCmd , _ , portForwardStderr := kubectlCommand ( timeout , t , kubeconfigPath , envVarsWithProxy , "port-forward" , "--namespace" , env . ConciergeNamespace , conciergePod . Name , "10443:8443" )
2021-03-15 21:34:09 +00:00
portForwardCmd . Env = envVarsWithProxy
2021-03-19 16:36:05 +00:00
// Start, but don't wait for the command to finish.
2021-03-18 15:24:02 +00:00
err := portForwardCmd . Start ( )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err , ` "kubectl port-forward" failed ` )
go func ( ) {
assert . EqualErrorf ( t , portForwardCmd . Wait ( ) , "signal: killed" , ` wanted "kubectl port-forward" to get signaled because context was cancelled (stderr: %q) ` , portForwardStderr . String ( ) )
} ( )
2021-03-15 23:31:54 +00:00
2021-03-29 16:30:20 +00:00
// The server should recognize this this
2021-03-19 16:36:05 +00:00
// is going to be a long-running command and keep the connection open as long as the client stays connected.
2021-03-15 21:34:09 +00:00
2021-03-29 16:30:20 +00:00
// curl the endpoint as many times as we can within 70 seconds.
// this will ensure that we don't run into idle timeouts.
var curlStdOut , curlStdErr bytes . Buffer
timeout , cancelFunc = context . WithTimeout ( ctx , 75 * time . Second )
defer cancelFunc ( )
startTime := time . Now ( )
for time . Now ( ) . Before ( startTime . Add ( 70 * time . Second ) ) {
2021-04-09 21:52:53 +00:00
curlCmd := exec . CommandContext ( timeout , "curl" , "-k" , "-sS" , "https://127.0.0.1:10443" ) // -sS turns off the progressbar but still prints errors
2021-03-15 21:34:09 +00:00
curlCmd . Stdout = & curlStdOut
curlCmd . Stderr = & curlStdErr
2021-03-29 16:30:20 +00:00
curlErr := curlCmd . Run ( )
if curlErr != nil {
t . Log ( "curl error: " + curlErr . Error ( ) )
2021-03-15 21:34:09 +00:00
t . Log ( "curlStdErr: " + curlStdErr . String ( ) )
t . Log ( "stdout: " + curlStdOut . String ( ) )
}
2021-04-05 22:14:24 +00:00
t . Log ( "Running curl through the kubectl port-forward port for 70 seconds. Elapsed time:" , time . Since ( startTime ) )
2021-03-29 16:30:20 +00:00
time . Sleep ( 1 * time . Second )
}
// curl the endpoint once more, once 70 seconds has elapsed, to make sure the connection is still open.
timeout , cancelFunc = context . WithTimeout ( ctx , 30 * time . Second )
defer cancelFunc ( )
2021-04-09 21:52:53 +00:00
curlCmd := exec . CommandContext ( timeout , "curl" , "-k" , "-sS" , "https://127.0.0.1:10443" ) // -sS turns off the progressbar but still prints errors
2021-03-29 16:30:20 +00:00
curlCmd . Stdout = & curlStdOut
curlCmd . Stderr = & curlStdErr
curlErr := curlCmd . Run ( )
if curlErr != nil {
t . Log ( "curl error: " + curlErr . Error ( ) )
t . Log ( "curlStdErr: " + curlStdErr . String ( ) )
t . Log ( "stdout: " + curlStdOut . String ( ) )
}
// We expect this to 403, but all we care is that it gets through.
require . NoError ( t , curlErr )
2021-04-09 21:52:53 +00:00
require . Contains ( t , curlStdOut . String ( ) , ` "forbidden: User \"system:anonymous\" cannot get path \"/\"" ` )
2021-03-29 16:30:20 +00:00
} )
t . Run ( "kubectl port-forward and keeping the connection open for over a minute (idle)" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-29 16:30:20 +00:00
kubeconfigPath , envVarsWithProxy , _ := getImpersonationKubeconfig ( t , env , impersonationProxyURL , impersonationProxyCACertPEM , credentialRequestSpecWithWorkingCredentials . Authenticator )
// Run the kubectl port-forward command.
timeout , cancelFunc := context . WithTimeout ( ctx , 2 * time . Minute )
defer cancelFunc ( )
2021-04-09 21:52:53 +00:00
portForwardCmd , _ , portForwardStderr := kubectlCommand ( timeout , t , kubeconfigPath , envVarsWithProxy , "port-forward" , "--namespace" , env . ConciergeNamespace , conciergePod . Name , "10444:8443" )
2021-03-29 16:30:20 +00:00
portForwardCmd . Env = envVarsWithProxy
// Start, but don't wait for the command to finish.
err := portForwardCmd . Start ( )
require . NoError ( t , err , ` "kubectl port-forward" failed ` )
go func ( ) {
assert . EqualErrorf ( t , portForwardCmd . Wait ( ) , "signal: killed" , ` wanted "kubectl port-forward" to get signaled because context was cancelled (stderr: %q) ` , portForwardStderr . String ( ) )
} ( )
// Wait to see if we time out. The default timeout is 60 seconds, but the server should recognize this this
// is going to be a long-running command and keep the connection open as long as the client stays connected.
time . Sleep ( 70 * time . Second )
timeout , cancelFunc = context . WithTimeout ( ctx , 2 * time . Minute )
defer cancelFunc ( )
2021-04-09 21:52:53 +00:00
curlCmd := exec . CommandContext ( timeout , "curl" , "-k" , "-sS" , "https://127.0.0.1:10444" ) // -sS turns off the progressbar but still prints errors
2021-03-29 16:30:20 +00:00
var curlStdOut , curlStdErr bytes . Buffer
curlCmd . Stdout = & curlStdOut
curlCmd . Stderr = & curlStdErr
err = curlCmd . Run ( )
if err != nil {
t . Log ( "curl error: " + err . Error ( ) )
t . Log ( "curlStdErr: " + curlStdErr . String ( ) )
t . Log ( "stdout: " + curlStdOut . String ( ) )
}
// We expect this to 403, but all we care is that it gets through.
require . NoError ( t , err )
2021-04-09 21:52:53 +00:00
require . Contains ( t , curlStdOut . String ( ) , ` "forbidden: User \"system:anonymous\" cannot get path \"/\"" ` )
2021-02-23 01:23:11 +00:00
} )
2021-02-25 22:40:02 +00:00
2021-03-15 21:34:09 +00:00
t . Run ( "using and watching all the basic verbs" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-15 21:34:09 +00:00
// Create a namespace, because it will be easier to exercise "deletecollection" if we have a namespace.
namespaceName := createTestNamespace ( t , adminClient )
// Create and start informer to exercise the "watch" verb for us.
informerFactory := k8sinformers . NewSharedInformerFactoryWithOptions (
2021-03-27 02:32:33 +00:00
impersonationProxyKubeClient ( t ) ,
2021-03-15 21:34:09 +00:00
0 ,
k8sinformers . WithNamespace ( namespaceName ) )
informer := informerFactory . Core ( ) . V1 ( ) . ConfigMaps ( )
informer . Informer ( ) // makes sure that the informer will cache
stopChannel := make ( chan struct { } )
informerFactory . Start ( stopChannel )
t . Cleanup ( func ( ) {
// Shut down the informer.
close ( stopChannel )
} )
informerFactory . WaitForCacheSync ( ctx . Done ( ) )
// Use labels on our created ConfigMaps to avoid accidentally listing other ConfigMaps that might
// exist in the namespace. In Kube 1.20+ there is a default ConfigMap in every namespace.
configMapLabels := labels . Set {
"pinniped.dev/testConfigMap" : library . RandHex ( t , 8 ) ,
}
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Test "create" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
_ , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Create ( ctx ,
2021-03-15 21:34:09 +00:00
& corev1 . ConfigMap { ObjectMeta : metav1 . ObjectMeta { Name : "configmap-1" , Labels : configMapLabels } } ,
metav1 . CreateOptions { } ,
)
require . NoError ( t , err )
2021-03-27 02:32:33 +00:00
_ , err = impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Create ( ctx ,
2021-03-15 21:34:09 +00:00
& corev1 . ConfigMap { ObjectMeta : metav1 . ObjectMeta { Name : "configmap-2" , Labels : configMapLabels } } ,
metav1 . CreateOptions { } ,
)
require . NoError ( t , err )
2021-03-27 02:32:33 +00:00
_ , err = impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Create ( ctx ,
2021-03-15 21:34:09 +00:00
& corev1 . ConfigMap { ObjectMeta : metav1 . ObjectMeta { Name : "configmap-3" , Labels : configMapLabels } } ,
metav1 . CreateOptions { } ,
)
require . NoError ( t , err )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Make sure that all of the created ConfigMaps show up in the informer's cache to
// demonstrate that the informer's "watch" verb is working through the impersonation proxy.
require . Eventually ( t , func ( ) bool {
_ , err1 := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-1" )
_ , err2 := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-2" )
_ , err3 := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-3" )
return err1 == nil && err2 == nil && err3 == nil
} , 10 * time . Second , 50 * time . Millisecond )
// Test "get" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
configMap3 , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Get ( ctx , "configmap-3" , metav1 . GetOptions { } )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Test "list" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
listResult , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . List ( ctx , metav1 . ListOptions {
2021-03-15 21:34:09 +00:00
LabelSelector : configMapLabels . String ( ) ,
} )
require . NoError ( t , err )
require . Len ( t , listResult . Items , 3 )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Test "update" verb through the impersonation proxy.
configMap3 . Data = map [ string ] string { "foo" : "bar" }
2021-03-27 02:32:33 +00:00
updateResult , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Update ( ctx , configMap3 , metav1 . UpdateOptions { } )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err )
require . Equal ( t , "bar" , updateResult . Data [ "foo" ] )
// Make sure that the updated ConfigMap shows up in the informer's cache.
require . Eventually ( t , func ( ) bool {
configMap , err := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-3" )
return err == nil && configMap . Data [ "foo" ] == "bar"
} , 10 * time . Second , 50 * time . Millisecond )
// Test "patch" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
patchResult , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Patch ( ctx ,
2021-03-15 21:34:09 +00:00
"configmap-3" ,
types . MergePatchType ,
[ ] byte ( ` { "data": { "baz":"42"}} ` ) ,
metav1 . PatchOptions { } ,
)
require . NoError ( t , err )
require . Equal ( t , "bar" , patchResult . Data [ "foo" ] )
require . Equal ( t , "42" , patchResult . Data [ "baz" ] )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Make sure that the patched ConfigMap shows up in the informer's cache.
require . Eventually ( t , func ( ) bool {
configMap , err := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-3" )
return err == nil && configMap . Data [ "foo" ] == "bar" && configMap . Data [ "baz" ] == "42"
} , 10 * time . Second , 50 * time . Millisecond )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Test "delete" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
err = impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Delete ( ctx , "configmap-3" , metav1 . DeleteOptions { } )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Make sure that the deleted ConfigMap shows up in the informer's cache.
require . Eventually ( t , func ( ) bool {
_ , getErr := informer . Lister ( ) . ConfigMaps ( namespaceName ) . Get ( "configmap-3" )
list , listErr := informer . Lister ( ) . ConfigMaps ( namespaceName ) . List ( configMapLabels . AsSelector ( ) )
return k8serrors . IsNotFound ( getErr ) && listErr == nil && len ( list ) == 2
} , 10 * time . Second , 50 * time . Millisecond )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Test "deletecollection" verb through the impersonation proxy.
2021-03-27 02:32:33 +00:00
err = impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . DeleteCollection ( ctx , metav1 . DeleteOptions { } , metav1 . ListOptions { } )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err )
2021-02-23 18:38:02 +00:00
2021-03-15 21:34:09 +00:00
// Make sure that the deleted ConfigMaps shows up in the informer's cache.
require . Eventually ( t , func ( ) bool {
list , listErr := informer . Lister ( ) . ConfigMaps ( namespaceName ) . List ( configMapLabels . AsSelector ( ) )
return listErr == nil && len ( list ) == 0
} , 10 * time . Second , 50 * time . Millisecond )
2021-02-19 00:27:03 +00:00
2021-03-15 21:34:09 +00:00
// There should be no ConfigMaps left.
2021-03-27 02:32:33 +00:00
listResult , err = impersonationProxyKubeClient ( t ) . CoreV1 ( ) . ConfigMaps ( namespaceName ) . List ( ctx , metav1 . ListOptions {
2021-03-15 21:34:09 +00:00
LabelSelector : configMapLabels . String ( ) ,
} )
require . NoError ( t , err )
require . Len ( t , listResult . Items , 0 )
2021-03-02 01:53:26 +00:00
} )
2021-04-09 21:52:53 +00:00
t . Run ( "nested impersonation as a regular user is allowed if they have enough RBAC permissions" , func ( t * testing . T ) {
t . Parallel ( )
2021-03-15 21:34:09 +00:00
// Make a client which will send requests through the impersonation proxy and will also add
// impersonate headers to the request.
2021-04-09 21:52:53 +00:00
nestedImpersonationClient := newImpersonationProxyClient ( t , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig { UserName : "other-user-to-impersonate" } , refreshCredential )
2021-03-15 21:34:09 +00:00
// Check that we can get some resource through the impersonation proxy without any impersonation headers on the request.
// We could use any resource for this, but we happen to know that this one should exist.
2021-03-27 02:32:33 +00:00
_ , err := impersonationProxyKubeClient ( t ) . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
2021-03-15 21:34:09 +00:00
require . NoError ( t , err )
2021-03-02 01:53:26 +00:00
2021-03-15 21:34:09 +00:00
// Now we'll see what happens when we add an impersonation header to the request. This should generate a
// request similar to the one above, except that it will also have an impersonation header.
2021-04-09 21:52:53 +00:00
_ , err = nestedImpersonationClient . Kubernetes . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
// this user is not allowed to impersonate other users
require . True ( t , k8serrors . IsForbidden ( err ) , err )
2021-03-15 21:34:09 +00:00
require . EqualError ( t , err , fmt . Sprintf (
` users "other-user-to-impersonate" is forbidden: ` +
2021-04-09 21:52:53 +00:00
` User "%s" cannot impersonate resource "users" in API group "" at the cluster scope ` ,
env . TestUser . ExpectedUsername ) )
// impersonate the GC service account instead which can read anything (the binding to edit allows this)
nestedImpersonationClientAsSA := newImpersonationProxyClient ( t , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig { UserName : "system:serviceaccount:kube-system:generic-garbage-collector" } , refreshCredential )
_ , err = nestedImpersonationClientAsSA . Kubernetes . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
require . NoError ( t , err )
expectedGroups := make ( [ ] string , 0 , len ( env . TestUser . ExpectedGroups ) + 1 ) // make sure we do not mutate env.TestUser.ExpectedGroups
expectedGroups = append ( expectedGroups , env . TestUser . ExpectedGroups ... )
expectedGroups = append ( expectedGroups , "system:authenticated" )
expectedOriginalUserInfo := authenticationv1 . UserInfo {
Username : env . TestUser . ExpectedUsername ,
Groups : expectedGroups ,
}
expectedOriginalUserInfoJSON , err := json . Marshal ( expectedOriginalUserInfo )
require . NoError ( t , err )
// check that we impersonated the correct user and that the original user is retained in the extra
whoAmI , err := nestedImpersonationClientAsSA . PinnipedConcierge . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
require . Equal ( t ,
expectedWhoAmIRequestResponse (
"system:serviceaccount:kube-system:generic-garbage-collector" ,
[ ] string { "system:serviceaccounts" , "system:serviceaccounts:kube-system" , "system:authenticated" } ,
map [ string ] identityv1alpha1 . ExtraValue {
"original-user-info.impersonation-proxy.concierge.pinniped.dev" : { string ( expectedOriginalUserInfoJSON ) } ,
} ,
) ,
whoAmI ,
)
_ , err = newImpersonationProxyClient ( t , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig {
UserName : "system:serviceaccount:kube-system:generic-garbage-collector" ,
Extra : map [ string ] [ ] string {
"some-fancy-key" : { "with a dangerous value" } ,
} ,
} ,
refreshCredential ) . PinnipedConcierge . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
// this user should not be able to impersonate extra
require . True ( t , k8serrors . IsForbidden ( err ) , err )
require . EqualError ( t , err , fmt . Sprintf (
` userextras.authentication.k8s.io "with a dangerous value" is forbidden: ` +
` User "%s" cannot impersonate resource "userextras/some-fancy-key" in API group "authentication.k8s.io" at the cluster scope ` ,
2021-03-15 21:34:09 +00:00
env . TestUser . ExpectedUsername ) )
} )
2021-03-02 01:53:26 +00:00
2021-04-09 21:52:53 +00:00
t . Run ( "nested impersonation as a cluster admin user is allowed" , func ( t * testing . T ) {
t . Parallel ( )
2021-03-15 21:34:09 +00:00
// Copy the admin credentials from the admin kubeconfig.
adminClientRestConfig := library . NewClientConfig ( t )
2021-04-09 21:52:53 +00:00
clusterAdminCredentials := getCredForConfig ( t , adminClientRestConfig )
2021-03-11 20:52:39 +00:00
2021-04-09 21:52:53 +00:00
// figure out who the admin user is
whoAmIAdmin , err := newImpersonationProxyClientWithCredentials ( t ,
clusterAdminCredentials , impersonationProxyURL , impersonationProxyCACertPEM , nil ) .
PinnipedConcierge . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
2021-03-11 20:52:39 +00:00
2021-05-10 17:22:51 +00:00
// The WhoAmI API is lossy:
// - It drops UID
// - It lowercases all extra keys
// the admin user on EKS has both a UID set and an extra key with uppercase characters
// Thus we fallback to the CSR API to grab the UID and Extra to handle this scenario
uid , extra := getUIDAndExtraViaCSR ( ctx , t , whoAmIAdmin . Status . KubernetesUserInfo . User . UID ,
newImpersonationProxyClientWithCredentials ( t ,
clusterAdminCredentials , impersonationProxyURL , impersonationProxyCACertPEM , nil ) .
Kubernetes ,
)
expectedExtra := make ( map [ string ] authenticationv1 . ExtraValue , len ( extra ) )
for k , v := range extra {
2021-04-09 21:52:53 +00:00
expectedExtra [ k ] = authenticationv1 . ExtraValue ( v )
}
expectedOriginalUserInfo := authenticationv1 . UserInfo {
Username : whoAmIAdmin . Status . KubernetesUserInfo . User . Username ,
2021-05-10 17:22:51 +00:00
UID : uid ,
Groups : whoAmIAdmin . Status . KubernetesUserInfo . User . Groups ,
Extra : expectedExtra ,
2021-03-15 21:34:09 +00:00
}
2021-04-09 21:52:53 +00:00
expectedOriginalUserInfoJSON , err := json . Marshal ( expectedOriginalUserInfo )
require . NoError ( t , err )
2021-03-11 20:52:39 +00:00
2021-03-15 21:34:09 +00:00
// Make a client using the admin credentials which will send requests through the impersonation proxy
// and will also add impersonate headers to the request.
2021-04-09 21:52:53 +00:00
nestedImpersonationClient := newImpersonationProxyClientWithCredentials ( t ,
clusterAdminCredentials , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig {
UserName : "other-user-to-impersonate" ,
Groups : [ ] string { "other-group-1" , "other-group-2" } ,
Extra : map [ string ] [ ] string {
"this-key" : { "to this value" } ,
} ,
} ,
)
_ , err = nestedImpersonationClient . Kubernetes . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
// the impersonated user lacks the RBAC to perform this call
require . True ( t , k8serrors . IsForbidden ( err ) , err )
require . EqualError ( t , err , fmt . Sprintf (
` secrets "%s" is forbidden: User "other-user-to-impersonate" cannot get resource "secrets" in API group "" in the namespace "%s" ` ,
impersonationProxyTLSSecretName ( env ) , env . ConciergeNamespace ,
) )
// check that we impersonated the correct user and that the original user is retained in the extra
whoAmI , err := nestedImpersonationClient . PinnipedConcierge . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
require . Equal ( t ,
expectedWhoAmIRequestResponse (
"other-user-to-impersonate" ,
[ ] string { "other-group-1" , "other-group-2" , "system:authenticated" } ,
map [ string ] identityv1alpha1 . ExtraValue {
"this-key" : { "to this value" } ,
"original-user-info.impersonation-proxy.concierge.pinniped.dev" : { string ( expectedOriginalUserInfoJSON ) } ,
} ,
) ,
whoAmI ,
)
} )
t . Run ( "nested impersonation as a cluster admin fails on reserved key" , func ( t * testing . T ) {
t . Parallel ( )
adminClientRestConfig := library . NewClientConfig ( t )
clusterAdminCredentials := getCredForConfig ( t , adminClientRestConfig )
nestedImpersonationClient := newImpersonationProxyClientWithCredentials ( t ,
clusterAdminCredentials , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig {
UserName : "other-user-to-impersonate" ,
Groups : [ ] string { "other-group-1" , "other-group-2" } ,
Extra : map [ string ] [ ] string {
"this-good-key" : { "to this good value" } ,
"something.impersonation-proxy.concierge.pinniped.dev" : { "super sneaky value" } ,
} ,
} ,
)
_ , err := nestedImpersonationClient . Kubernetes . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
require . EqualError ( t , err , "Internal error occurred: unimplemented functionality - unable to act as current user" )
require . True ( t , k8serrors . IsInternalError ( err ) , err )
require . Equal ( t , & k8serrors . StatusError {
ErrStatus : metav1 . Status {
Status : metav1 . StatusFailure ,
Code : http . StatusInternalServerError ,
Reason : metav1 . StatusReasonInternalError ,
Details : & metav1 . StatusDetails {
Causes : [ ] metav1 . StatusCause {
{
Message : "unimplemented functionality - unable to act as current user" ,
} ,
} ,
} ,
Message : "Internal error occurred: unimplemented functionality - unable to act as current user" ,
} ,
} , err )
} )
// this works because impersonation cannot set UID and thus the final user info the proxy sees has no UID
t . Run ( "nested impersonation as a service account is allowed if it has enough RBAC permissions" , func ( t * testing . T ) {
t . Parallel ( )
namespaceName := createTestNamespace ( t , adminClient )
saName , saToken , saUID := createServiceAccountToken ( ctx , t , adminClient , namespaceName )
nestedImpersonationClient := newImpersonationProxyClientWithCredentials ( t ,
& loginv1alpha1 . ClusterCredential { Token : saToken } , impersonationProxyURL , impersonationProxyCACertPEM ,
& rest . ImpersonationConfig { UserName : "system:serviceaccount:kube-system:root-ca-cert-publisher" } ) . PinnipedConcierge
_ , err := nestedImpersonationClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
// this SA is not yet allowed to impersonate SAs
require . True ( t , k8serrors . IsForbidden ( err ) , err )
require . EqualError ( t , err , fmt . Sprintf (
` serviceaccounts "root-ca-cert-publisher" is forbidden: ` +
` User "%s" cannot impersonate resource "serviceaccounts" in API group "" in the namespace "kube-system" ` ,
serviceaccount . MakeUsername ( namespaceName , saName ) ) )
// webhook authorizer deny cache TTL is 10 seconds so we need to wait long enough for it to drain
time . Sleep ( 15 * time . Second )
// allow the test SA to impersonate any SA
library . CreateTestClusterRoleBinding ( t ,
rbacv1 . Subject { Kind : rbacv1 . ServiceAccountKind , Name : saName , Namespace : namespaceName } ,
rbacv1 . RoleRef { Kind : "ClusterRole" , APIGroup : rbacv1 . GroupName , Name : "edit" } ,
)
library . WaitForUserToHaveAccess ( t , serviceaccount . MakeUsername ( namespaceName , saName ) , [ ] string { } , & authorizationv1 . ResourceAttributes {
Verb : "impersonate" , Group : "" , Version : "v1" , Resource : "serviceaccounts" ,
} )
whoAmI , err := nestedImpersonationClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
require . Equal ( t ,
expectedWhoAmIRequestResponse (
"system:serviceaccount:kube-system:root-ca-cert-publisher" ,
[ ] string { "system:serviceaccounts" , "system:serviceaccounts:kube-system" , "system:authenticated" } ,
map [ string ] identityv1alpha1 . ExtraValue {
"original-user-info.impersonation-proxy.concierge.pinniped.dev" : {
fmt . Sprintf ( ` { "username":"%s","uid":"%s","groups":["system:serviceaccounts","system:serviceaccounts:%s","system:authenticated"]} ` ,
serviceaccount . MakeUsername ( namespaceName , saName ) , saUID , namespaceName ) ,
} ,
} ,
) ,
whoAmI ,
2021-03-15 21:34:09 +00:00
)
} )
2021-03-11 00:57:15 +00:00
2021-03-15 21:34:09 +00:00
t . Run ( "WhoAmIRequests and different kinds of authentication through the impersonation proxy" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-15 21:34:09 +00:00
// Test using the TokenCredentialRequest for authentication.
2021-03-27 02:32:33 +00:00
impersonationProxyPinnipedConciergeClient := newImpersonationProxyClient ( t ,
2021-04-09 21:52:53 +00:00
impersonationProxyURL , impersonationProxyCACertPEM , nil , refreshCredential ,
2021-03-15 21:34:09 +00:00
) . PinnipedConcierge
whoAmI , err := impersonationProxyPinnipedConciergeClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
2021-04-09 21:52:53 +00:00
expectedGroups := make ( [ ] string , 0 , len ( env . TestUser . ExpectedGroups ) + 1 ) // make sure we do not mutate env.TestUser.ExpectedGroups
expectedGroups = append ( expectedGroups , env . TestUser . ExpectedGroups ... )
expectedGroups = append ( expectedGroups , "system:authenticated" )
2021-03-15 21:34:09 +00:00
require . Equal ( t ,
expectedWhoAmIRequestResponse (
env . TestUser . ExpectedUsername ,
2021-04-09 21:52:53 +00:00
expectedGroups ,
nil ,
2021-03-15 21:34:09 +00:00
) ,
whoAmI ,
)
// Test an unauthenticated request which does not include any credentials.
impersonationProxyAnonymousPinnipedConciergeClient := newAnonymousImpersonationProxyClient (
2021-04-09 21:52:53 +00:00
t , impersonationProxyURL , impersonationProxyCACertPEM , nil ,
2021-03-15 21:34:09 +00:00
) . PinnipedConcierge
whoAmI , err = impersonationProxyAnonymousPinnipedConciergeClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
require . Equal ( t ,
expectedWhoAmIRequestResponse (
"system:anonymous" ,
[ ] string { "system:unauthenticated" } ,
2021-04-09 21:52:53 +00:00
nil ,
2021-03-15 21:34:09 +00:00
) ,
whoAmI ,
)
2021-04-20 15:19:58 +00:00
// Test using a service account token.
2021-03-15 21:34:09 +00:00
namespaceName := createTestNamespace ( t , adminClient )
2021-04-20 15:19:58 +00:00
saName , saToken , _ := createServiceAccountToken ( ctx , t , adminClient , namespaceName )
2021-04-09 21:52:53 +00:00
impersonationProxyServiceAccountPinnipedConciergeClient := newImpersonationProxyClientWithCredentials ( t ,
& loginv1alpha1 . ClusterCredential { Token : saToken } ,
impersonationProxyURL , impersonationProxyCACertPEM , nil ) . PinnipedConcierge
2021-04-20 15:19:58 +00:00
whoAmI , err = impersonationProxyServiceAccountPinnipedConciergeClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
2021-03-15 21:34:09 +00:00
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
2021-04-20 15:19:58 +00:00
require . NoError ( t , err )
require . Equal ( t ,
expectedWhoAmIRequestResponse (
serviceaccount . MakeUsername ( namespaceName , saName ) ,
[ ] string { "system:serviceaccounts" , "system:serviceaccounts:" + namespaceName , "system:authenticated" } ,
nil ,
) ,
whoAmI ,
)
} )
t . Run ( "WhoAmIRequests and SA token request" , func ( t * testing . T ) {
namespaceName := createTestNamespace ( t , adminClient )
kubeClient := adminClient . CoreV1 ( )
saName , _ , saUID := createServiceAccountToken ( ctx , t , adminClient , namespaceName )
_ , tokenRequestProbeErr := kubeClient . ServiceAccounts ( namespaceName ) . CreateToken ( ctx , saName , & authenticationv1 . TokenRequest { } , metav1 . CreateOptions { } )
if k8serrors . IsNotFound ( tokenRequestProbeErr ) && tokenRequestProbeErr . Error ( ) == "the server could not find the requested resource" {
return // stop test early since the token request API is not enabled on this cluster - other errors are caught below
}
pod , err := kubeClient . Pods ( namespaceName ) . Create ( ctx , & corev1 . Pod {
ObjectMeta : metav1 . ObjectMeta {
GenerateName : "test-impersonation-proxy-" ,
} ,
Spec : corev1 . PodSpec {
Containers : [ ] corev1 . Container {
{
Name : "ignored-but-required" ,
Image : "does-not-matter" ,
2021-03-14 01:25:23 +00:00
} ,
} ,
2021-04-20 15:19:58 +00:00
ServiceAccountName : saName ,
2021-03-14 01:25:23 +00:00
} ,
2021-04-20 15:19:58 +00:00
} , metav1 . CreateOptions { } )
require . NoError ( t , err )
tokenRequestBadAudience , err := kubeClient . ServiceAccounts ( namespaceName ) . CreateToken ( ctx , saName , & authenticationv1 . TokenRequest {
Spec : authenticationv1 . TokenRequestSpec {
Audiences : [ ] string { "should-fail-because-wrong-audience" } , // anything that is not an API server audience
BoundObjectRef : & authenticationv1 . BoundObjectReference {
Kind : "Pod" ,
APIVersion : "" ,
Name : pod . Name ,
UID : pod . UID ,
} ,
} ,
} , metav1 . CreateOptions { } )
require . NoError ( t , err )
impersonationProxySABadAudPinnipedConciergeClient := newImpersonationProxyClientWithCredentials ( t ,
& loginv1alpha1 . ClusterCredential { Token : tokenRequestBadAudience . Status . Token } ,
impersonationProxyURL , impersonationProxyCACertPEM , nil ) . PinnipedConcierge
_ , badAudErr := impersonationProxySABadAudPinnipedConciergeClient . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . True ( t , k8serrors . IsUnauthorized ( badAudErr ) , library . Sdump ( badAudErr ) )
tokenRequest , err := kubeClient . ServiceAccounts ( namespaceName ) . CreateToken ( ctx , saName , & authenticationv1 . TokenRequest {
Spec : authenticationv1 . TokenRequestSpec {
Audiences : [ ] string { } ,
BoundObjectRef : & authenticationv1 . BoundObjectReference {
Kind : "Pod" ,
APIVersion : "" ,
Name : pod . Name ,
UID : pod . UID ,
} ,
} ,
} , metav1 . CreateOptions { } )
require . NoError ( t , err )
impersonationProxySAClient := newImpersonationProxyClientWithCredentials ( t ,
& loginv1alpha1 . ClusterCredential { Token : tokenRequest . Status . Token } ,
impersonationProxyURL , impersonationProxyCACertPEM , nil )
whoAmITokenReq , err := impersonationProxySAClient . PinnipedConcierge . IdentityV1alpha1 ( ) . WhoAmIRequests ( ) .
Create ( ctx , & identityv1alpha1 . WhoAmIRequest { } , metav1 . CreateOptions { } )
require . NoError ( t , err )
// new service account tokens include the pod info in the extra fields
require . Equal ( t ,
expectedWhoAmIRequestResponse (
serviceaccount . MakeUsername ( namespaceName , saName ) ,
[ ] string { "system:serviceaccounts" , "system:serviceaccounts:" + namespaceName , "system:authenticated" } ,
map [ string ] identityv1alpha1 . ExtraValue {
"authentication.kubernetes.io/pod-name" : { pod . Name } ,
"authentication.kubernetes.io/pod-uid" : { string ( pod . UID ) } ,
} ,
) ,
whoAmITokenReq ,
)
// allow the test SA to create CSRs
library . CreateTestClusterRoleBinding ( t ,
rbacv1 . Subject { Kind : rbacv1 . ServiceAccountKind , Name : saName , Namespace : namespaceName } ,
rbacv1 . RoleRef { Kind : "ClusterRole" , APIGroup : rbacv1 . GroupName , Name : "system:node-bootstrapper" } ,
)
library . WaitForUserToHaveAccess ( t , serviceaccount . MakeUsername ( namespaceName , saName ) , [ ] string { } , & authorizationv1 . ResourceAttributes {
Verb : "create" , Group : certificatesv1 . GroupName , Version : "*" , Resource : "certificatesigningrequests" ,
} )
privateKey , err := ecdsa . GenerateKey ( elliptic . P256 ( ) , rand . Reader )
require . NoError ( t , err )
csrPEM , err := cert . MakeCSR ( privateKey , & pkix . Name {
CommonName : "panda-man" ,
Organization : [ ] string { "living-the-dream" , "need-more-sleep" } ,
} , nil , nil )
require . NoError ( t , err )
csrName , _ , err := csr . RequestCertificate (
impersonationProxySAClient . Kubernetes ,
csrPEM ,
"" ,
certificatesv1 . KubeAPIServerClientSignerName ,
[ ] certificatesv1 . KeyUsage { certificatesv1 . UsageClientAuth } ,
privateKey ,
)
require . NoError ( t , err )
saCSR , err := impersonationProxySAClient . Kubernetes . CertificatesV1beta1 ( ) . CertificateSigningRequests ( ) . Get ( ctx , csrName , metav1 . GetOptions { } )
require . NoError ( t , err )
err = adminClient . CertificatesV1beta1 ( ) . CertificateSigningRequests ( ) . Delete ( ctx , csrName , metav1 . DeleteOptions { } )
require . NoError ( t , err )
// make sure the user info that the CSR captured matches the SA, including the UID
require . Equal ( t , serviceaccount . MakeUsername ( namespaceName , saName ) , saCSR . Spec . Username )
require . Equal ( t , string ( saUID ) , saCSR . Spec . UID )
require . Equal ( t , [ ] string { "system:serviceaccounts" , "system:serviceaccounts:" + namespaceName , "system:authenticated" } , saCSR . Spec . Groups )
require . Equal ( t , map [ string ] certificatesv1beta1 . ExtraValue {
"authentication.kubernetes.io/pod-name" : { pod . Name } ,
"authentication.kubernetes.io/pod-uid" : { string ( pod . UID ) } ,
} , saCSR . Spec . Extra )
2021-03-06 00:14:45 +00:00
} )
2021-03-15 21:34:09 +00:00
t . Run ( "kubectl as a client" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-25 23:57:37 +00:00
kubeconfigPath , envVarsWithProxy , tempDir := getImpersonationKubeconfig ( t , env , impersonationProxyURL , impersonationProxyCACertPEM , credentialRequestSpecWithWorkingCredentials . Authenticator )
2021-03-15 21:34:09 +00:00
// Try "kubectl exec" through the impersonation proxy.
echoString := "hello world"
remoteEchoFile := fmt . Sprintf ( "/tmp/test-impersonation-proxy-echo-file-%d.txt" , time . Now ( ) . Unix ( ) )
stdout , err := runKubectl ( t , kubeconfigPath , envVarsWithProxy , "exec" , "--namespace" , env . ConciergeNamespace , conciergePod . Name , "--" , "bash" , "-c" , fmt . Sprintf ( ` echo "%s" | tee %s ` , echoString , remoteEchoFile ) )
require . NoError ( t , err , ` "kubectl exec" failed ` )
require . Equal ( t , echoString + "\n" , stdout )
// run the kubectl cp command
localEchoFile := filepath . Join ( tempDir , filepath . Base ( remoteEchoFile ) )
_ , err = runKubectl ( t , kubeconfigPath , envVarsWithProxy , "cp" , fmt . Sprintf ( "%s/%s:%s" , env . ConciergeNamespace , conciergePod . Name , remoteEchoFile ) , localEchoFile )
require . NoError ( t , err , ` "kubectl cp" failed ` )
localEchoFileData , err := ioutil . ReadFile ( localEchoFile )
require . NoError ( t , err )
require . Equal ( t , echoString + "\n" , string ( localEchoFileData ) )
defer func ( ) {
_ , _ = runKubectl ( t , kubeconfigPath , envVarsWithProxy , "exec" , "--namespace" , env . ConciergeNamespace , conciergePod . Name , "--" , "rm" , remoteEchoFile ) // cleanup remote echo file
} ( )
2021-03-11 20:10:16 +00:00
2021-03-15 21:34:09 +00:00
// run the kubectl logs command
logLinesCount := 10
stdout , err = runKubectl ( t , kubeconfigPath , envVarsWithProxy , "logs" , "--namespace" , env . ConciergeNamespace , conciergePod . Name , fmt . Sprintf ( "--tail=%d" , logLinesCount ) )
require . NoError ( t , err , ` "kubectl logs" failed ` )
require . Equalf ( t , logLinesCount , strings . Count ( stdout , "\n" ) , "wanted %d newlines in kubectl logs output:\n%s" , logLinesCount , stdout )
// run the kubectl attach command
namespaceName := createTestNamespace ( t , adminClient )
attachPod := library . CreatePod ( ctx , t , "impersonation-proxy-attach" , namespaceName , corev1 . PodSpec {
Containers : [ ] corev1 . Container {
{
Name : "impersonation-proxy-attach" ,
Image : conciergePod . Spec . Containers [ 0 ] . Image ,
Command : [ ] string { "bash" } ,
Args : [ ] string { "-c" , ` while true; do read VAR; echo "VAR: $VAR"; done ` } ,
Stdin : true ,
} ,
} ,
} )
timeout , cancelFunc := context . WithTimeout ( ctx , 2 * time . Minute )
defer cancelFunc ( )
2021-03-19 14:42:11 +00:00
attachCmd , attachStdout , attachStderr := kubectlCommand ( timeout , t , kubeconfigPath , envVarsWithProxy , "attach" , "--stdin=true" , "--namespace" , namespaceName , attachPod . Name , "-v=10" )
2021-03-15 21:34:09 +00:00
attachCmd . Env = envVarsWithProxy
attachStdin , err := attachCmd . StdinPipe ( )
require . NoError ( t , err )
2021-03-11 20:10:16 +00:00
2021-03-15 21:34:09 +00:00
// start but don't wait for the attach command
err = attachCmd . Start ( )
require . NoError ( t , err )
2021-03-19 13:59:24 +00:00
attachExitCh := make ( chan struct { } )
go func ( ) {
assert . NoError ( t , attachCmd . Wait ( ) )
close ( attachExitCh )
} ( )
2021-03-11 20:10:16 +00:00
2021-03-15 21:34:09 +00:00
// write to stdin on the attach process
_ , err = attachStdin . Write ( [ ] byte ( echoString + "\n" ) )
require . NoError ( t , err )
2021-03-11 20:10:16 +00:00
2021-03-15 21:34:09 +00:00
// see that we can read stdout and it spits out stdin output back to us
wantAttachStdout := fmt . Sprintf ( "VAR: %s\n" , echoString )
2021-03-18 21:53:14 +00:00
require . Eventuallyf ( t , func ( ) bool { return attachStdout . String ( ) == wantAttachStdout } , time . Second * 60 , time . Millisecond * 250 , ` got "kubectl attach" stdout: %q, wanted: %q (stderr: %q) ` , attachStdout . String ( ) , wantAttachStdout , attachStderr . String ( ) )
2021-03-06 00:14:45 +00:00
2021-03-15 21:34:09 +00:00
// close stdin and attach process should exit
err = attachStdin . Close ( )
require . NoError ( t , err )
2021-03-19 13:59:24 +00:00
requireClose ( t , attachExitCh , time . Second * 20 )
2021-03-10 23:49:09 +00:00
} )
2021-03-15 21:34:09 +00:00
t . Run ( "websocket client" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-15 21:34:09 +00:00
namespaceName := createTestNamespace ( t , adminClient )
2021-03-16 23:57:28 +00:00
2021-03-27 02:32:33 +00:00
impersonationRestConfig := impersonationProxyRestConfig (
refreshCredential ( t , impersonationProxyURL , impersonationProxyCACertPEM ) ,
2021-04-09 21:52:53 +00:00
impersonationProxyURL , impersonationProxyCACertPEM , nil ,
2021-03-27 02:32:33 +00:00
)
2021-03-15 21:34:09 +00:00
tlsConfig , err := rest . TLSConfigFor ( impersonationRestConfig )
require . NoError ( t , err )
2021-03-10 23:49:09 +00:00
2021-03-15 21:34:09 +00:00
wantConfigMapLabelKey , wantConfigMapLabelValue := "some-label-key" , "some-label-value"
dest , _ := url . Parse ( impersonationProxyURL )
dest . Scheme = "wss"
dest . Path = "/api/v1/namespaces/" + namespaceName + "/configmaps"
dest . RawQuery = url . Values {
"watch" : { "1" } ,
"labelSelector" : { fmt . Sprintf ( "%s=%s" , wantConfigMapLabelKey , wantConfigMapLabelValue ) } ,
"resourceVersion" : { "0" } ,
} . Encode ( )
dialer := websocket . Dialer {
TLSClientConfig : tlsConfig ,
2021-03-10 23:49:09 +00:00
}
2021-03-26 16:28:42 +00:00
if ! clusterSupportsLoadBalancers {
2021-03-15 21:34:09 +00:00
dialer . Proxy = func ( req * http . Request ) ( * url . URL , error ) {
proxyURL , err := url . Parse ( env . Proxy )
require . NoError ( t , err )
t . Logf ( "passing request for %s through proxy %s" , req . URL , proxyURL . String ( ) )
return proxyURL , nil
}
}
c , r , err := dialer . Dial ( dest . String ( ) , http . Header { "Origin" : { dest . String ( ) } } )
if r != nil {
defer func ( ) {
require . NoError ( t , r . Body . Close ( ) )
} ( )
}
if err != nil && r != nil {
body , _ := ioutil . ReadAll ( r . Body )
t . Logf ( "websocket dial failed: %d:%s" , r . StatusCode , body )
}
require . NoError ( t , err )
2021-03-10 00:58:44 +00:00
2021-03-15 21:34:09 +00:00
// perform a create through the admin client
wantConfigMap := & corev1 . ConfigMap {
ObjectMeta : metav1 . ObjectMeta { Name : "configmap-1" , Labels : map [ string ] string { wantConfigMapLabelKey : wantConfigMapLabelValue } } ,
}
wantConfigMap , err = adminClient . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Create ( ctx ,
wantConfigMap ,
metav1 . CreateOptions { } ,
)
require . NoError ( t , err )
t . Cleanup ( func ( ) {
require . NoError ( t , adminClient . CoreV1 ( ) . ConfigMaps ( namespaceName ) .
DeleteCollection ( context . Background ( ) , metav1 . DeleteOptions { } , metav1 . ListOptions { } ) )
} )
2021-03-11 23:49:24 +00:00
2021-03-15 21:34:09 +00:00
// see if the websocket client received an event for the create
_ , message , err := c . ReadMessage ( )
if err != nil {
t . Fatalf ( "Unexpected error: %v" , err )
}
var got watchJSON
err = json . Unmarshal ( message , & got )
require . NoError ( t , err )
if got . Type != watch . Added {
t . Errorf ( "Unexpected type: %v" , got . Type )
}
var actualConfigMap corev1 . ConfigMap
require . NoError ( t , json . Unmarshal ( got . Object , & actualConfigMap ) )
actualConfigMap . TypeMeta = metav1 . TypeMeta { } // This isn't filled out in the wantConfigMap we got back from create.
require . Equal ( t , * wantConfigMap , actualConfigMap )
2021-03-11 23:49:24 +00:00
} )
2021-03-15 21:34:09 +00:00
t . Run ( "http2 client" , func ( t * testing . T ) {
2021-04-09 21:52:53 +00:00
t . Parallel ( )
2021-03-15 21:34:09 +00:00
namespaceName := createTestNamespace ( t , adminClient )
2021-03-16 23:57:28 +00:00
2021-03-15 21:34:09 +00:00
wantConfigMapLabelKey , wantConfigMapLabelValue := "some-label-key" , "some-label-value"
wantConfigMap := & corev1 . ConfigMap {
ObjectMeta : metav1 . ObjectMeta { Name : "configmap-1" , Labels : map [ string ] string { wantConfigMapLabelKey : wantConfigMapLabelValue } } ,
}
wantConfigMap , err = adminClient . CoreV1 ( ) . ConfigMaps ( namespaceName ) . Create ( ctx ,
wantConfigMap ,
metav1 . CreateOptions { } ,
)
require . NoError ( t , err )
t . Cleanup ( func ( ) {
_ = adminClient . CoreV1 ( ) . ConfigMaps ( namespaceName ) . DeleteCollection ( ctx , metav1 . DeleteOptions { } , metav1 . ListOptions { } )
} )
2021-03-11 23:49:24 +00:00
2021-03-15 21:34:09 +00:00
// create rest client
2021-03-27 02:32:33 +00:00
restConfig := impersonationProxyRestConfig (
refreshCredential ( t , impersonationProxyURL , impersonationProxyCACertPEM ) ,
2021-04-09 21:52:53 +00:00
impersonationProxyURL , impersonationProxyCACertPEM , nil ,
2021-03-27 02:32:33 +00:00
)
2021-03-11 23:49:24 +00:00
2021-03-15 21:34:09 +00:00
tlsConfig , err := rest . TLSConfigFor ( restConfig )
require . NoError ( t , err )
httpTransport := http . Transport {
TLSClientConfig : tlsConfig ,
2021-03-11 23:49:24 +00:00
}
2021-03-26 16:28:42 +00:00
if ! clusterSupportsLoadBalancers {
2021-03-15 21:34:09 +00:00
httpTransport . Proxy = func ( req * http . Request ) ( * url . URL , error ) {
proxyURL , err := url . Parse ( env . Proxy )
require . NoError ( t , err )
t . Logf ( "passing request for %s through proxy %s" , req . URL , proxyURL . String ( ) )
return proxyURL , nil
}
}
err = http2 . ConfigureTransport ( & httpTransport )
require . NoError ( t , err )
2021-03-11 23:49:24 +00:00
2021-03-15 21:34:09 +00:00
httpClient := http . Client {
Transport : & httpTransport ,
}
2021-03-11 23:49:24 +00:00
2021-03-15 21:34:09 +00:00
dest , _ := url . Parse ( impersonationProxyURL )
dest . Path = "/api/v1/namespaces/" + namespaceName + "/configmaps/configmap-1"
getConfigmapRequest , err := http . NewRequestWithContext ( ctx , http . MethodGet , dest . String ( ) , nil )
require . NoError ( t , err )
response , err := httpClient . Do ( getConfigmapRequest )
require . NoError ( t , err )
body , _ := ioutil . ReadAll ( response . Body )
t . Logf ( "http2 status code: %d, proto: %s, message: %s" , response . StatusCode , response . Proto , body )
require . Equal ( t , "HTTP/2.0" , response . Proto )
require . Equal ( t , http . StatusOK , response . StatusCode )
defer func ( ) {
require . NoError ( t , response . Body . Close ( ) )
} ( )
var actualConfigMap corev1 . ConfigMap
require . NoError ( t , json . Unmarshal ( body , & actualConfigMap ) )
actualConfigMap . TypeMeta = metav1 . TypeMeta { } // This isn't filled out in the wantConfigMap we got back from create.
require . Equal ( t , * wantConfigMap , actualConfigMap )
// watch configmaps
dest . Path = "/api/v1/namespaces/" + namespaceName + "/configmaps"
dest . RawQuery = url . Values {
"watch" : { "1" } ,
"labelSelector" : { fmt . Sprintf ( "%s=%s" , wantConfigMapLabelKey , wantConfigMapLabelValue ) } ,
"resourceVersion" : { "0" } ,
} . Encode ( )
watchConfigmapsRequest , err := http . NewRequestWithContext ( ctx , http . MethodGet , dest . String ( ) , nil )
require . NoError ( t , err )
response , err = httpClient . Do ( watchConfigmapsRequest )
require . NoError ( t , err )
require . Equal ( t , "HTTP/2.0" , response . Proto )
require . Equal ( t , http . StatusOK , response . StatusCode )
defer func ( ) {
require . NoError ( t , response . Body . Close ( ) )
} ( )
// decode
decoder := json . NewDecoder ( response . Body )
var got watchJSON
err = decoder . Decode ( & got )
require . NoError ( t , err )
if got . Type != watch . Added {
t . Errorf ( "Unexpected type: %v" , got . Type )
}
err = json . Unmarshal ( got . Object , & actualConfigMap )
require . NoError ( t , err )
require . Equal ( t , "configmap-1" , actualConfigMap . Name )
actualConfigMap . TypeMeta = metav1 . TypeMeta { } // This isn't filled out in the wantConfigMap we got back from create.
require . Equal ( t , * wantConfigMap , actualConfigMap )
} )
2021-03-10 00:58:44 +00:00
} )
2021-03-10 21:08:15 +00:00
t . Run ( "manually disabling the impersonation proxy feature" , func ( t * testing . T ) {
// Update configuration to force the proxy to disabled mode
2021-03-26 16:28:42 +00:00
configMap := impersonationProxyConfigMapForConfig ( t , env , impersonator . Config { Mode : impersonator . ModeDisabled } )
if clusterSupportsLoadBalancers {
2021-03-10 21:08:15 +00:00
t . Logf ( "creating configmap %s" , configMap . Name )
2021-03-18 15:24:02 +00:00
_ , err := adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Create ( ctx , & configMap , metav1 . CreateOptions { } )
2021-03-10 21:08:15 +00:00
require . NoError ( t , err )
} else {
t . Logf ( "updating configmap %s" , configMap . Name )
2021-03-18 15:24:02 +00:00
_ , err := adminClient . CoreV1 ( ) . ConfigMaps ( env . ConciergeNamespace ) . Update ( ctx , & configMap , metav1 . UpdateOptions { } )
2021-03-10 21:08:15 +00:00
require . NoError ( t , err )
}
2021-02-12 01:22:47 +00:00
2021-03-26 16:28:42 +00:00
if clusterSupportsLoadBalancers {
2021-03-10 21:08:15 +00:00
// The load balancer should have been deleted when we disabled the impersonation proxy.
// Note that this can take kind of a long time on real cloud providers (e.g. ~22 seconds on EKS).
library . RequireEventuallyWithoutError ( t , func ( ) ( bool , error ) {
hasService , err := hasImpersonationProxyLoadBalancerService ( ctx , env , adminClient )
return ! hasService , err
} , 2 * time . Minute , 500 * time . Millisecond )
}
2021-02-25 18:27:19 +00:00
2021-03-10 21:08:15 +00:00
// Check that the impersonation proxy port has shut down.
// Ideally we could always check that the impersonation proxy's port has shut down, but on clusters where we
// do not run the squid proxy we have no easy way to see beyond the load balancer to see inside the cluster,
// so we'll skip this check on clusters which have load balancers but don't run the squid proxy.
// The other cluster types that do run the squid proxy will give us sufficient coverage here.
if env . Proxy != "" {
require . Eventually ( t , func ( ) bool {
// It's okay if this returns RBAC errors because this user has no role bindings.
// What we want to see is that the proxy eventually shuts down entirely.
2021-04-09 21:52:53 +00:00
_ , err := impersonationProxyViaSquidKubeClientWithoutCredential ( t , proxyServiceEndpoint ) . CoreV1 ( ) . Namespaces ( ) . List ( ctx , metav1 . ListOptions { } )
2021-03-18 20:34:30 +00:00
isErr , _ := isServiceUnavailableViaSquidError ( err , proxyServiceEndpoint )
return isErr
2021-03-10 21:08:15 +00:00
} , 20 * time . Second , 500 * time . Millisecond )
}
2021-02-25 22:40:02 +00:00
2021-03-10 21:08:15 +00:00
// Check that the generated TLS cert Secret was deleted by the controller because it's supposed to clean this up
// when we disable the impersonator.
require . Eventually ( t , func ( ) bool {
2021-03-18 15:24:02 +00:00
_ , err := adminClient . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyTLSSecretName ( env ) , metav1 . GetOptions { } )
2021-03-10 21:08:15 +00:00
return k8serrors . IsNotFound ( err )
} , 10 * time . Second , 250 * time . Millisecond )
// Check that the generated CA cert Secret was not deleted by the controller because it's supposed to keep this
// around in case we decide to later re-enable the impersonator. We want to avoid generating new CA certs when
// possible because they make their way into kubeconfigs on client machines.
2021-03-18 15:24:02 +00:00
_ , err := adminClient . CoreV1 ( ) . Secrets ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyCASecretName ( env ) , metav1 . GetOptions { } )
2021-03-10 21:08:15 +00:00
require . NoError ( t , err )
2021-03-03 20:53:23 +00:00
2021-03-10 21:08:15 +00:00
// At this point the impersonator should be stopped. The CredentialIssuer's strategies array should be updated to
// include an unsuccessful impersonation strategy saying that it was manually configured to be disabled.
2021-03-26 16:28:42 +00:00
requireDisabledStrategy ( ctx , t , env , adminConciergeClient )
2021-03-10 21:08:15 +00:00
2021-03-27 02:32:33 +00:00
if ! env . HasCapability ( library . ClusterSigningKeyIsAvailable ) && env . HasCapability ( library . AnonymousAuthenticationSupported ) {
2021-03-10 21:08:15 +00:00
// This cluster does not support the cluster signing key strategy, so now that we've manually disabled the
// impersonation strategy, we should be left with no working strategies.
// Given that there are no working strategies, a TokenCredentialRequest which would otherwise work should now
// fail, because there is no point handing out credentials that are not going to work for any strategy.
2021-03-27 02:32:33 +00:00
// Note that library.CreateTokenCredentialRequest makes an unauthenticated request, so we can't meaningfully
// perform this part of the test on a cluster which does not allow anonymous authentication.
2021-03-11 00:57:15 +00:00
tokenCredentialRequestResponse , err := library . CreateTokenCredentialRequest ( ctx , t , credentialRequestSpecWithWorkingCredentials )
2021-03-10 21:08:15 +00:00
require . NoError ( t , err )
2021-03-10 22:50:46 +00:00
2021-03-10 21:08:15 +00:00
require . NotNil ( t , tokenCredentialRequestResponse . Status . Message , "expected an error message but got nil" )
require . Equal ( t , "authentication failed" , * tokenCredentialRequestResponse . Status . Message )
require . Nil ( t , tokenCredentialRequestResponse . Status . Credential )
}
} )
2021-03-03 20:53:23 +00:00
}
2021-03-11 18:01:17 +00:00
func createTestNamespace ( t * testing . T , adminClient kubernetes . Interface ) string {
t . Helper ( )
ctx , cancel := context . WithTimeout ( context . Background ( ) , time . Minute )
defer cancel ( )
namespace , err := adminClient . CoreV1 ( ) . Namespaces ( ) . Create ( ctx , & corev1 . Namespace {
ObjectMeta : metav1 . ObjectMeta { GenerateName : "impersonation-integration-test-" } ,
} , metav1 . CreateOptions { } )
require . NoError ( t , err )
t . Cleanup ( func ( ) {
ctx , cancel := context . WithTimeout ( context . Background ( ) , 2 * time . Minute )
defer cancel ( )
t . Logf ( "cleaning up test namespace %s" , namespace . Name )
require . NoError ( t , adminClient . CoreV1 ( ) . Namespaces ( ) . Delete ( ctx , namespace . Name , metav1 . DeleteOptions { } ) )
} )
return namespace . Name
}
2021-04-09 21:52:53 +00:00
func createServiceAccountToken ( ctx context . Context , t * testing . T , adminClient kubernetes . Interface , namespaceName string ) ( name , token string , uid types . UID ) {
2021-03-11 18:01:17 +00:00
t . Helper ( )
serviceAccount , err := adminClient . CoreV1 ( ) . ServiceAccounts ( namespaceName ) . Create ( ctx ,
& corev1 . ServiceAccount { ObjectMeta : metav1 . ObjectMeta { GenerateName : "int-test-service-account-" } } , metav1 . CreateOptions { } )
require . NoError ( t , err )
t . Cleanup ( func ( ) {
require . NoError ( t , adminClient . CoreV1 ( ) . ServiceAccounts ( namespaceName ) .
Delete ( context . Background ( ) , serviceAccount . Name , metav1 . DeleteOptions { } ) )
} )
secret , err := adminClient . CoreV1 ( ) . Secrets ( namespaceName ) . Create ( ctx , & corev1 . Secret {
ObjectMeta : metav1 . ObjectMeta {
GenerateName : "int-test-service-account-token-" ,
Annotations : map [ string ] string {
corev1 . ServiceAccountNameKey : serviceAccount . Name ,
} ,
} ,
Type : corev1 . SecretTypeServiceAccountToken ,
} , metav1 . CreateOptions { } )
require . NoError ( t , err )
t . Cleanup ( func ( ) {
require . NoError ( t , adminClient . CoreV1 ( ) . Secrets ( namespaceName ) .
Delete ( context . Background ( ) , secret . Name , metav1 . DeleteOptions { } ) )
} )
library . RequireEventuallyWithoutError ( t , func ( ) ( bool , error ) {
secret , err = adminClient . CoreV1 ( ) . Secrets ( namespaceName ) . Get ( ctx , secret . Name , metav1 . GetOptions { } )
if err != nil {
return false , err
}
return len ( secret . Data [ corev1 . ServiceAccountTokenKey ] ) > 0 , nil
} , time . Minute , time . Second )
2021-04-09 21:52:53 +00:00
return serviceAccount . Name , string ( secret . Data [ corev1 . ServiceAccountTokenKey ] ) , serviceAccount . UID
2021-03-11 18:01:17 +00:00
}
2021-04-09 21:52:53 +00:00
func expectedWhoAmIRequestResponse ( username string , groups [ ] string , extra map [ string ] identityv1alpha1 . ExtraValue ) * identityv1alpha1 . WhoAmIRequest {
2021-03-11 00:57:15 +00:00
return & identityv1alpha1 . WhoAmIRequest {
Status : identityv1alpha1 . WhoAmIRequestStatus {
KubernetesUserInfo : identityv1alpha1 . KubernetesUserInfo {
User : identityv1alpha1 . UserInfo {
Username : username ,
UID : "" , // no way to impersonate UID: https://github.com/kubernetes/kubernetes/issues/93699
Groups : groups ,
2021-04-09 21:52:53 +00:00
Extra : extra ,
2021-03-11 00:57:15 +00:00
} ,
} ,
} ,
}
}
2021-03-27 02:32:33 +00:00
func performImpersonatorDiscovery ( ctx context . Context , t * testing . T , env * library . TestEnv , adminConciergeClient pinnipedconciergeclientset . Interface ) ( string , [ ] byte ) {
2021-03-03 20:53:23 +00:00
t . Helper ( )
2021-04-09 21:52:53 +00:00
2021-03-03 20:53:23 +00:00
var impersonationProxyURL string
var impersonationProxyCACertPEM [ ] byte
t . Log ( "Waiting for CredentialIssuer strategy to be successful" )
library . RequireEventuallyWithoutError ( t , func ( ) ( bool , error ) {
credentialIssuer , err := adminConciergeClient . ConfigV1alpha1 ( ) . CredentialIssuers ( ) . Get ( ctx , credentialIssuerName ( env ) , metav1 . GetOptions { } )
if err != nil || credentialIssuer . Status . Strategies == nil {
t . Log ( "Did not find any CredentialIssuer with any strategies" )
return false , nil // didn't find it, but keep trying
}
for _ , strategy := range credentialIssuer . Status . Strategies {
// There will be other strategy types in the list, so ignore those.
2021-03-15 18:42:57 +00:00
if strategy . Type == conciergev1alpha . ImpersonationProxyStrategyType && strategy . Status == conciergev1alpha . SuccessStrategyStatus { //nolint:nestif
2021-03-03 20:53:23 +00:00
if strategy . Frontend == nil {
return false , fmt . Errorf ( "did not find a Frontend" ) // unexpected, fail the test
}
if strategy . Frontend . ImpersonationProxyInfo == nil {
return false , fmt . Errorf ( "did not find an ImpersonationProxyInfo" ) // unexpected, fail the test
}
impersonationProxyURL = strategy . Frontend . ImpersonationProxyInfo . Endpoint
impersonationProxyCACertPEM , err = base64 . StdEncoding . DecodeString ( strategy . Frontend . ImpersonationProxyInfo . CertificateAuthorityData )
if err != nil {
return false , err // unexpected, fail the test
}
return true , nil // found it, continue the test!
2021-03-15 18:42:57 +00:00
} else if strategy . Type == conciergev1alpha . ImpersonationProxyStrategyType {
2021-03-03 20:53:23 +00:00
t . Logf ( "Waiting for successful impersonation proxy strategy on %s: found status %s with reason %s and message: %s" ,
credentialIssuerName ( env ) , strategy . Status , strategy . Reason , strategy . Message )
2021-03-15 18:42:57 +00:00
if strategy . Reason == conciergev1alpha . ErrorDuringSetupStrategyReason {
2021-03-03 20:53:23 +00:00
// The server encountered an unexpected error while starting the impersonator, so fail the test fast.
return false , fmt . Errorf ( "found impersonation strategy in %s state with message: %s" , strategy . Reason , strategy . Message )
}
}
}
t . Log ( "Did not find any impersonation proxy strategy on CredentialIssuer" )
return false , nil // didn't find it, but keep trying
} , 10 * time . Minute , 10 * time . Second )
t . Log ( "Found successful CredentialIssuer strategy" )
return impersonationProxyURL , impersonationProxyCACertPEM
}
2021-03-27 02:32:33 +00:00
func requireDisabledStrategy ( ctx context . Context , t * testing . T , env * library . TestEnv , adminConciergeClient pinnipedconciergeclientset . Interface ) {
2021-03-03 20:53:23 +00:00
t . Helper ( )
library . RequireEventuallyWithoutError ( t , func ( ) ( bool , error ) {
credentialIssuer , err := adminConciergeClient . ConfigV1alpha1 ( ) . CredentialIssuers ( ) . Get ( ctx , credentialIssuerName ( env ) , metav1 . GetOptions { } )
if err != nil || credentialIssuer . Status . Strategies == nil {
t . Log ( "Did not find any CredentialIssuer with any strategies" )
return false , nil // didn't find it, but keep trying
}
for _ , strategy := range credentialIssuer . Status . Strategies {
// There will be other strategy types in the list, so ignore those.
2021-03-15 18:42:57 +00:00
if strategy . Type == conciergev1alpha . ImpersonationProxyStrategyType &&
strategy . Status == conciergev1alpha . ErrorStrategyStatus &&
strategy . Reason == conciergev1alpha . DisabledStrategyReason { //nolint:nestif
2021-03-03 20:53:23 +00:00
return true , nil // found it, continue the test!
2021-03-15 18:42:57 +00:00
} else if strategy . Type == conciergev1alpha . ImpersonationProxyStrategyType {
2021-03-03 20:53:23 +00:00
t . Logf ( "Waiting for disabled impersonation proxy strategy on %s: found status %s with reason %s and message: %s" ,
credentialIssuerName ( env ) , strategy . Status , strategy . Reason , strategy . Message )
2021-03-15 18:42:57 +00:00
if strategy . Reason == conciergev1alpha . ErrorDuringSetupStrategyReason {
2021-03-03 20:53:23 +00:00
// The server encountered an unexpected error while stopping the impersonator, so fail the test fast.
return false , fmt . Errorf ( "found impersonation strategy in %s state with message: %s" , strategy . Reason , strategy . Message )
}
}
}
t . Log ( "Did not find any impersonation proxy strategy on CredentialIssuer" )
return false , nil // didn't find it, but keep trying
} , 1 * time . Minute , 500 * time . Millisecond )
2021-02-12 01:22:47 +00:00
}
2021-03-11 00:57:15 +00:00
func credentialAlmostExpired ( t * testing . T , credential * loginv1alpha1 . TokenCredentialRequest ) bool {
t . Helper ( )
2021-04-09 21:52:53 +00:00
2021-03-11 00:57:15 +00:00
pemBlock , _ := pem . Decode ( [ ] byte ( credential . Status . Credential . ClientCertificateData ) )
parsedCredential , err := x509 . ParseCertificate ( pemBlock . Bytes )
require . NoError ( t , err )
timeRemaining := time . Until ( parsedCredential . NotAfter )
if timeRemaining < 2 * time . Minute {
t . Logf ( "The TokenCredentialRequest cred is almost expired and needs to be refreshed. Expires in %s." , timeRemaining )
return true
}
t . Logf ( "The TokenCredentialRequest cred is good for some more time (%s) so using it." , timeRemaining )
return false
}
2021-04-09 21:52:53 +00:00
func impersonationProxyRestConfig ( credential * loginv1alpha1 . ClusterCredential , host string , caData [ ] byte , nestedImpersonationConfig * rest . ImpersonationConfig ) * rest . Config {
2021-03-11 00:57:15 +00:00
config := rest . Config {
Host : host ,
TLSClientConfig : rest . TLSClientConfig {
Insecure : caData == nil ,
CAData : caData ,
CertData : [ ] byte ( credential . ClientCertificateData ) ,
KeyData : [ ] byte ( credential . ClientKeyData ) ,
} ,
// kubectl would set both the client cert and the token, so we'll do that too.
// The Kube API server will ignore the token if the client cert successfully authenticates.
// Only if the client cert is not present or fails to authenticate will it use the token.
// Historically, it works that way because some web browsers will always send your
// corporate-assigned client cert even if it is not valid, and it doesn't want to treat
// that as a failure if you also sent a perfectly good token.
// We would like the impersonation proxy to imitate that behavior, so we test it here.
BearerToken : credential . Token ,
}
2021-04-09 21:52:53 +00:00
if nestedImpersonationConfig != nil {
config . Impersonate = * nestedImpersonationConfig
2021-03-11 00:57:15 +00:00
}
return & config
}
func kubeconfigProxyFunc ( t * testing . T , squidProxyURL string ) func ( req * http . Request ) ( * url . URL , error ) {
return func ( req * http . Request ) ( * url . URL , error ) {
t . Helper ( )
2021-04-09 21:52:53 +00:00
2021-03-11 00:57:15 +00:00
parsedSquidProxyURL , err := url . Parse ( squidProxyURL )
require . NoError ( t , err )
t . Logf ( "passing request for %s through proxy %s" , req . URL , parsedSquidProxyURL . String ( ) )
return parsedSquidProxyURL , nil
}
}
2021-03-26 16:28:42 +00:00
func impersonationProxyConfigMapForConfig ( t * testing . T , env * library . TestEnv , config impersonator . Config ) corev1 . ConfigMap {
2021-03-03 20:53:23 +00:00
t . Helper ( )
2021-04-09 21:52:53 +00:00
2021-02-12 01:22:47 +00:00
configString , err := yaml . Marshal ( config )
require . NoError ( t , err )
configMap := corev1 . ConfigMap {
2021-02-18 23:58:27 +00:00
ObjectMeta : metav1 . ObjectMeta {
2021-03-02 17:31:24 +00:00
Name : impersonationProxyConfigMapName ( env ) ,
2021-02-18 23:58:27 +00:00
} ,
2021-02-12 01:22:47 +00:00
Data : map [ string ] string {
"config.yaml" : string ( configString ) ,
} }
return configMap
}
2021-03-02 17:31:24 +00:00
func hasImpersonationProxyLoadBalancerService ( ctx context . Context , env * library . TestEnv , client kubernetes . Interface ) ( bool , error ) {
service , err := client . CoreV1 ( ) . Services ( env . ConciergeNamespace ) . Get ( ctx , impersonationProxyLoadBalancerName ( env ) , metav1 . GetOptions { } )
2021-02-25 22:40:02 +00:00
if k8serrors . IsNotFound ( err ) {
return false , nil
}
if err != nil {
return false , err
}
return service . Spec . Type == corev1 . ServiceTypeLoadBalancer , nil
}
2021-02-12 01:22:47 +00:00
2021-03-02 17:31:24 +00:00
func impersonationProxyConfigMapName ( env * library . TestEnv ) string {
return env . ConciergeAppName + "-impersonation-proxy-config"
}
func impersonationProxyTLSSecretName ( env * library . TestEnv ) string {
return env . ConciergeAppName + "-impersonation-proxy-tls-serving-certificate"
}
func impersonationProxyCASecretName ( env * library . TestEnv ) string {
return env . ConciergeAppName + "-impersonation-proxy-ca-certificate"
}
func impersonationProxyLoadBalancerName ( env * library . TestEnv ) string {
return env . ConciergeAppName + "-impersonation-proxy-load-balancer"
}
2021-03-03 20:53:23 +00:00
func credentialIssuerName ( env * library . TestEnv ) string {
return env . ConciergeAppName + "-config"
}
2021-03-10 00:58:44 +00:00
2021-03-25 23:57:37 +00:00
func getImpersonationKubeconfig ( t * testing . T , env * library . TestEnv , impersonationProxyURL string , impersonationProxyCACertPEM [ ] byte , authenticator corev1 . TypedLocalObjectReference ) ( string , [ ] string , string ) {
2021-03-15 19:28:53 +00:00
t . Helper ( )
pinnipedExe := library . PinnipedCLIPath ( t )
tempDir := testutil . TempDir ( t )
var envVarsWithProxy [ ] string
if ! env . HasCapability ( library . HasExternalLoadBalancerProvider ) {
// Only if you don't have a load balancer, use the squid proxy when it's available.
envVarsWithProxy = append ( os . Environ ( ) , env . ProxyEnv ( ) ... )
}
// Get the kubeconfig.
getKubeConfigCmd := [ ] string { "get" , "kubeconfig" ,
"--concierge-api-group-suffix" , env . APIGroupSuffix ,
"--oidc-skip-browser" ,
"--static-token" , env . TestUser . Token ,
2021-03-25 23:57:37 +00:00
"--concierge-authenticator-name" , authenticator . Name ,
2021-03-26 00:19:47 +00:00
"--concierge-authenticator-type" , "webhook" ,
2021-03-15 19:28:53 +00:00
// Force the use of impersonation proxy strategy, but let it auto-discover the endpoint and CA.
"--concierge-mode" , "ImpersonationProxy" }
t . Log ( "Running:" , pinnipedExe , getKubeConfigCmd )
kubeconfigYAML , getKubeConfigStderr := runPinnipedCLI ( t , envVarsWithProxy , pinnipedExe , getKubeConfigCmd ... )
// "pinniped get kubectl" prints some status messages to stderr
t . Log ( getKubeConfigStderr )
// Make sure that the "pinniped get kubeconfig" auto-discovered the impersonation proxy and we're going to
// make our kubectl requests through the impersonation proxy. Avoid using require.Contains because the error
// message would contain credentials.
require . True ( t ,
strings . Contains ( kubeconfigYAML , "server: " + impersonationProxyURL + "\n" ) ,
"the generated kubeconfig did not include the expected impersonation server address: %s" ,
impersonationProxyURL ,
)
require . True ( t ,
strings . Contains ( kubeconfigYAML , "- --concierge-ca-bundle-data=" + base64 . StdEncoding . EncodeToString ( impersonationProxyCACertPEM ) + "\n" ) ,
"the generated kubeconfig did not include the base64 encoded version of this expected impersonation CA cert: %s" ,
impersonationProxyCACertPEM ,
)
// Write the kubeconfig to a temp file.
kubeconfigPath := filepath . Join ( tempDir , "kubeconfig.yaml" )
require . NoError ( t , ioutil . WriteFile ( kubeconfigPath , [ ] byte ( kubeconfigYAML ) , 0600 ) )
return kubeconfigPath , envVarsWithProxy , tempDir
}
// func to create kubectl commands with a kubeconfig.
func kubectlCommand ( timeout context . Context , t * testing . T , kubeconfigPath string , envVarsWithProxy [ ] string , args ... string ) ( * exec . Cmd , * syncBuffer , * syncBuffer ) {
2021-04-09 21:52:53 +00:00
t . Helper ( )
2021-03-15 19:28:53 +00:00
allArgs := append ( [ ] string { "--kubeconfig" , kubeconfigPath } , args ... )
//nolint:gosec // we are not performing malicious argument injection against ourselves
kubectlCmd := exec . CommandContext ( timeout , "kubectl" , allArgs ... )
var stdout , stderr syncBuffer
kubectlCmd . Stdout = & stdout
kubectlCmd . Stderr = & stderr
kubectlCmd . Env = envVarsWithProxy
t . Log ( "starting kubectl subprocess: kubectl" , strings . Join ( allArgs , " " ) )
return kubectlCmd , & stdout , & stderr
}
// Func to run kubeconfig commands.
func runKubectl ( t * testing . T , kubeconfigPath string , envVarsWithProxy [ ] string , args ... string ) ( string , error ) {
timeout , cancelFunc := context . WithTimeout ( context . Background ( ) , 2 * time . Minute )
defer cancelFunc ( )
kubectlCmd , stdout , stderr := kubectlCommand ( timeout , t , kubeconfigPath , envVarsWithProxy , args ... )
err := kubectlCmd . Run ( )
t . Logf ( "kubectl stdout output: %s" , stdout . String ( ) )
t . Logf ( "kubectl stderr output: %s" , stderr . String ( ) )
return stdout . String ( ) , err
}
2021-03-10 18:30:06 +00:00
// watchJSON defines the expected JSON wire equivalent of watch.Event.
2021-03-10 00:58:44 +00:00
type watchJSON struct {
Type watch . EventType ` json:"type,omitempty" `
Object json . RawMessage ` json:"object,omitempty" `
}
2021-03-18 20:34:30 +00:00
// requireServiceUnavailableViaSquidError returns whether the provided err is the error that is
// returned by squid when the impersonation proxy port inside the cluster is not listening.
func isServiceUnavailableViaSquidError ( err error , proxyServiceEndpoint string ) ( bool , string ) {
if err == nil {
return false , "error is nil"
}
for _ , wantContains := range [ ] string {
fmt . Sprintf ( ` Get "https://%s/api/v1/namespaces" ` , proxyServiceEndpoint ) ,
": Service Unavailable" ,
} {
if ! strings . Contains ( err . Error ( ) , wantContains ) {
return false , fmt . Sprintf ( "error does not contain %q" , wantContains )
}
}
return true , ""
}
2021-03-19 13:59:24 +00:00
func requireClose ( t * testing . T , c chan struct { } , timeout time . Duration ) {
t . Helper ( )
2021-04-09 21:52:53 +00:00
2021-03-19 13:59:24 +00:00
timer := time . NewTimer ( timeout )
select {
case <- c :
if ! timer . Stop ( ) {
<- timer . C
}
case <- timer . C :
require . FailNow ( t , "failed to receive from channel within " + timeout . String ( ) )
}
}
2021-03-27 02:32:33 +00:00
func createTokenCredentialRequest (
spec loginv1alpha1 . TokenCredentialRequestSpec ,
client pinnipedconciergeclientset . Interface ,
) ( * loginv1alpha1 . TokenCredentialRequest , error ) {
ctx , cancel := context . WithTimeout ( context . Background ( ) , time . Minute )
defer cancel ( )
return client . LoginV1alpha1 ( ) . TokenCredentialRequests ( ) . Create ( ctx ,
& loginv1alpha1 . TokenCredentialRequest { Spec : spec } , metav1 . CreateOptions { } ,
)
}
2021-04-09 21:52:53 +00:00
func newImpersonationProxyClientWithCredentials ( t * testing . T , credentials * loginv1alpha1 . ClusterCredential , impersonationProxyURL string , impersonationProxyCACertPEM [ ] byte , nestedImpersonationConfig * rest . ImpersonationConfig ) * kubeclient . Client {
t . Helper ( )
env := library . IntegrationEnv ( t )
clusterSupportsLoadBalancers := env . HasCapability ( library . HasExternalLoadBalancerProvider )
kubeconfig := impersonationProxyRestConfig ( credentials , impersonationProxyURL , impersonationProxyCACertPEM , nestedImpersonationConfig )
if ! clusterSupportsLoadBalancers {
// Only if there is no possibility to send traffic through a load balancer, then send the traffic through the Squid proxy.
// Prefer to go through a load balancer because that's how the impersonator is intended to be used in the real world.
kubeconfig . Proxy = kubeconfigProxyFunc ( t , env . Proxy )
}
return library . NewKubeclient ( t , kubeconfig )
}
func newAnonymousImpersonationProxyClient ( t * testing . T , impersonationProxyURL string , impersonationProxyCACertPEM [ ] byte , nestedImpersonationConfig * rest . ImpersonationConfig ) * kubeclient . Client {
t . Helper ( )
emptyCredentials := & loginv1alpha1 . ClusterCredential { }
return newImpersonationProxyClientWithCredentials ( t , emptyCredentials , impersonationProxyURL , impersonationProxyCACertPEM , nestedImpersonationConfig )
}
func impersonationProxyViaSquidKubeClientWithoutCredential ( t * testing . T , proxyServiceEndpoint string ) kubernetes . Interface {
t . Helper ( )
env := library . IntegrationEnv ( t )
proxyURL := "https://" + proxyServiceEndpoint
kubeconfig := impersonationProxyRestConfig ( & loginv1alpha1 . ClusterCredential { } , proxyURL , nil , nil )
kubeconfig . Proxy = kubeconfigProxyFunc ( t , env . Proxy )
return library . NewKubeclient ( t , kubeconfig ) . Kubernetes
}
func newImpersonationProxyClient (
t * testing . T ,
impersonationProxyURL string ,
impersonationProxyCACertPEM [ ] byte ,
nestedImpersonationConfig * rest . ImpersonationConfig ,
refreshCredentialFunc func ( t * testing . T , impersonationProxyURL string , impersonationProxyCACertPEM [ ] byte ) * loginv1alpha1 . ClusterCredential ,
) * kubeclient . Client {
t . Helper ( )
refreshedCredentials := refreshCredentialFunc ( t , impersonationProxyURL , impersonationProxyCACertPEM ) . DeepCopy ( )
refreshedCredentials . Token = "not a valid token" // demonstrates that client certs take precedence over tokens by setting both on the requests
return newImpersonationProxyClientWithCredentials ( t , refreshedCredentials , impersonationProxyURL , impersonationProxyCACertPEM , nestedImpersonationConfig )
}
// getCredForConfig is mostly just a hacky workaround for impersonationProxyRestConfig needing creds directly.
func getCredForConfig ( t * testing . T , config * rest . Config ) * loginv1alpha1 . ClusterCredential {
t . Helper ( )
out := & loginv1alpha1 . ClusterCredential { }
config = rest . CopyConfig ( config )
config . Wrap ( func ( rt http . RoundTripper ) http . RoundTripper {
return roundtripper . Func ( func ( req * http . Request ) ( * http . Response , error ) {
resp , err := rt . RoundTrip ( req )
r := req
if resp != nil && resp . Request != nil {
r = resp . Request
}
_ , _ , _ = bearertoken . New ( authenticator . TokenFunc ( func ( _ context . Context , token string ) ( * authenticator . Response , bool , error ) {
out . Token = token
return nil , false , nil
} ) ) . AuthenticateRequest ( r )
return resp , err
} )
} )
transportConfig , err := config . TransportConfig ( )
require . NoError ( t , err )
rt , err := transport . New ( transportConfig )
require . NoError ( t , err )
ctx , cancel := context . WithTimeout ( context . Background ( ) , 5 * time . Second )
defer cancel ( )
req , err := http . NewRequestWithContext ( ctx , "GET" , "https://localhost" , nil )
require . NoError ( t , err )
resp , _ := rt . RoundTrip ( req )
if resp != nil && resp . Body != nil {
_ = resp . Body . Close ( )
}
tlsConfig , err := transport . TLSConfigFor ( transportConfig )
require . NoError ( t , err )
if tlsConfig != nil && tlsConfig . GetClientCertificate != nil {
cert , err := tlsConfig . GetClientCertificate ( nil )
require . NoError ( t , err )
2021-05-07 21:22:08 +00:00
if len ( cert . Certificate ) > 0 {
require . Len ( t , cert . Certificate , 1 )
publicKey := pem . EncodeToMemory ( & pem . Block {
Type : "CERTIFICATE" ,
Bytes : cert . Certificate [ 0 ] ,
} )
out . ClientCertificateData = string ( publicKey )
2021-04-09 21:52:53 +00:00
2021-05-07 21:22:08 +00:00
privateKey , err := keyutil . MarshalPrivateKeyToPEM ( cert . PrivateKey )
require . NoError ( t , err )
out . ClientKeyData = string ( privateKey )
}
2021-04-09 21:52:53 +00:00
}
if * out == ( loginv1alpha1 . ClusterCredential { } ) {
t . Fatal ( "failed to get creds for config" )
}
return out
}
2021-05-10 04:50:59 +00:00
2021-05-10 17:22:51 +00:00
func getUIDAndExtraViaCSR ( ctx context . Context , t * testing . T , uid string , client kubernetes . Interface ) ( string , map [ string ] certificatesv1beta1 . ExtraValue ) {
2021-05-10 04:50:59 +00:00
t . Helper ( )
privateKey , err := ecdsa . GenerateKey ( elliptic . P256 ( ) , rand . Reader )
require . NoError ( t , err )
csrPEM , err := cert . MakeCSR ( privateKey , & pkix . Name {
CommonName : "panda-man" ,
Organization : [ ] string { "living-the-dream" , "need-more-sleep" } ,
} , nil , nil )
require . NoError ( t , err )
csrName , _ , err := csr . RequestCertificate (
client ,
csrPEM ,
"" ,
certificatesv1 . KubeAPIServerClientSignerName ,
[ ] certificatesv1 . KeyUsage { certificatesv1 . UsageClientAuth } ,
privateKey ,
)
require . NoError ( t , err )
csReq , err := client . CertificatesV1beta1 ( ) . CertificateSigningRequests ( ) . Get ( ctx , csrName , metav1 . GetOptions { } )
require . NoError ( t , err )
err = client . CertificatesV1beta1 ( ) . CertificateSigningRequests ( ) . Delete ( ctx , csrName , metav1 . DeleteOptions { } )
require . NoError ( t , err )
2021-05-10 17:22:51 +00:00
outUID := uid // in the future this may not be empty on some clusters
if len ( outUID ) == 0 {
outUID = csReq . Spec . UID
}
return outUID , csReq . Spec . Extra
2021-05-10 04:50:59 +00:00
}