Status
Current state: Under Discussion
Discussion thread: here
JIRA: KAFKA-13202
Please keep the discussion on the mailing list rather than commenting on the wiki (wiki discussions get unwieldy fast).
Motivation
The convergence on the use of OAuth/OIDC for authorization and authentication of Internet-based services is well underway. Use of a standard set of technologies allows organizations to select from among a range of OAuth/OIDC-compatible providers instead of defining, developing, and managing identity, security, and policy infrastructure on their own. Organizations can communicate with these providers using standard protocols (defined in RFCs) using code written against mature open source libraries in popular programming languages. There is already a rich (and growing!) ecosystem of tools to integrate with the providers. Although it is not a trivial task of selecting a provider and implementing an OAuth-compliant suite of applications and services in an organization, these standards provide flexibility within that organization.
When it comes to Kafka, the work done via KIP-255 (OAuth Authentication via SASL/OAUTHBEARER) introduced a framework that allowed for integration with OAuth-compliant providers. With this framework in place, Kafka clients could now pass a JWT access token to a broker when initializing the connection as a means of authentication. This means Kafka can start to leverage these standards for authorization and authentication for which many current and prospective customers are asking.
However, the KIP-255 quickly notes that:
OAuth 2 is a flexible framework with multiple ways of doing the same thing.
Because of the high degree of flexibility required by different organizations as well as differences in implementations by various OAuth providers, “supporting OAuth” isn't as straightforward as one might hope. While flexibility is one of OAuth's strengths it also presents one of its greatest challenges for adopters. Because of OAuth's flexibility, the organizational requirements, and the minor differences in provider implementation, the exact means and logic by which a JWT access token is retrieved and validated may vary on a case-by-case basis. In fact, a generalized, out-of-the-box OAuth implementation was specifically not included in KIP-255 because:
Anything beyond simple unsecured use cases requires significant functionality that is available via multiple open source libraries, and there is no need to duplicate that functionality here. It is also not desirable for the Kafka project to define a specific open source JWT/JWS/JWE library dependency; better to allow installations to use the library that they feel most comfortable with.
The implementation of KIP-255 provided a concrete example implementation that allowed clients to provide an unsecured JWT access token to the broker when initializing the connection only for–as the KIP states–”development scenarios.”
The KIP sums up this intentional missing functionality by directing organizations wishing to adopt OAuth within Kafka:
Production use cases will require writing an implementation of AuthenticateCallbackHandler
that can handle an instance of OAuthBearerTokenCallback
So the exact implementation is left up to each organization to implement. To fill the gap in the Apache Kafka project, a handful of open source Java implementations have been provided by the community. These implementations can be included in the class path of a OAuth-enabled Kafka client and configured appropriately to achieve the goal.
This project is to provide a concrete implementation of the interfaces defined in KIP-255 to allow Kafka to connect to an Open ID identity provider for authentication and token retrieval. While KIP-255 provides an unsecured JWT example for development, this will fill in the gap and provide a production-grade implementation.
Public Interfaces
No changes to the public interface are anticipated; it will leverage the existing AuthenticateCallbackHandler
API.
Proposed Changes
The OAuth/OIDC work will allow out-of-the-box configuration by any Apache Kafka users to connect to an external identity provider service (e.g. Okta, Auth0, Azure, etc.). The code will implement the standard OAuth clientcredentials
grant type.
The proposed change is largely composed of a pair of AuthenticateCallbackHandler
implementations: one to login on the client and one to validate on the broker.
As show in the above diagram, the login callback is executed on the client and the validate callback is executed on the broker.
Login Callback Handler (Client)
The login callback handler is invoked as part of the standard Kafka client login process. It is in this callback handler that the necessary HTTPS calls will be made to the OAuth provider using the provided configuration.
Token Retrieval Logic
Here is an example call to retrieve a JWT access token using curl
and jq
:
CLIENT_ID=abc123
CLIENT_SECRET='S3cr3t!'
URL=https://myidp.example.com/oauth2/default/v1/token
SCOPE=sales-pipeline
base_64_string=$(echo -n "$CLIENT_ID:$CLIENT_SECRET" | base64)
access_token=$(curl \
--silent \
--request POST \
--url $URL \
--header "accept: application/json" \
--header "authorization: Basic $base_64_string" \
--header "cache-control: no-cache" \
--header "content-type: application/x-www-form-urlencoded" \
--data "grant_type=client_credentials&scope=$SCOPE" | jq -r .access_token)
The OIDC specification requires the access token to be in the JWT format. The client will perform parsing and some basic structural validation of the JWT access token contents.
Client Configuration
The name of the implementation class will be org.apache.kafka.common.security.oauthbearer.secured.OAuthBearerLoginCallbackHandler
and it will accept instances of org.apache.kafka.common.security.oauthbearer.OAuthBearerTokenCallback
and org.apache.kafka.common.security.auth.SaslExtensionsCallback
. The fully-qualified class name will be provided to the client's sasl.login.callback.handler.class
configuration.
Because the HTTP call made to the OAuth/OIDC provider may time out or transiently fail, there will be a retry mechanism that waits between attempts. The number of attempts that are made (including the first attempt) are configured via the loginAttempts
configuration setting. The retry will use an exponential backoff approach; the first attempt to connect to the HTTP endpoint will be made immediately. If that first attempt fails, a second attempt will first wait a configurable number of milliseconds–loginRetryWaitMs–
before trying again. If that second attempt fails, the wait time (loginRetryWaitMs
) will be doubled before a third attempt. This pattern repeats up to loginAttempts
. However, there is also a configurable maximum wait time between attempts–loginRetryMaxWaitMs
–such that loginRetryWaitMs
<= loginRetryMaxWaitMs
, regardless of the number of attempts: min(currentRetryWaitMs, loginRetryMaxWaitMs)
.
There are several configuration options for this callback handler. Sensitive configuration options and SASL extensions appear under the JAAS configuration (sasl.jaas.config
) while the rest are top-level configuration.
The JAAS configuration options are:
clientId
: supports OAuthclientcredentials
grant typeclientSecret
: supports OAuth'sclientcredentials
grantscope
: optional scope to reference in the call to the OAuth server
The top-level configuration options for the client login callback handler are:
sasl.login.token.endpoint.uri
: OAuth issuer token endpoint URIsasl.login.scope.claim.name
: optional override name of the scope claim; defaults toscope
sasl.login.sub.claim.name
: optional override name of the sub claim; defaults tosub
sasl.login.connect.timeout.ms
: optional value in milliseconds for HTTPS connect timeout; defaults to10000
sasl.login.read.timeout.ms
: optional value in milliseconds for HTTPS read timeout; defaults to10000
sasl.login.attempts
: optional number of attempts to make to connect to the OAuth/OIDC identity provider; defaults to3
sasl.login.retry.wait.ms
: optional value in milliseconds for the amount of time to wait between HTTPS call attempts; defaults to 250sasl.login.retry.max.wait.ms
: optional value in milliseconds for the maximum wait between HTTPS call attempts (as described above); defaults to10000
Here's an example of the configuration as a part of a Java properties file:
sasl.login.callback.handler.class=...OAuthBearerLoginCallbackHandlersasl.login.token.endpoint.uri=https://myidp.example.com/oauth2/default/v1/token
sasl.jaas.config=...OAuthBearerLoginModule required \
clientId="abc123" \
clientSecret="S3cr3t!" \
scope="sales-pipeline" \
extension_supportFeatureX="true" \
extension_organizationId="sales-emea" ;
In the above example, only the OAuth provider’s sasl.login.token.endpoint.uri
has been specified. The values for clientId
and clientSecret
as provided by the OAuth provider for an “API” or “machine-to-machine” account are required. The optional scope
value will allow the inclusion of a scope
parameter when requesting the token.
Notice that there are two SASL extension configuration values in this example too: extension_supportFeatureX
and extension_organizationId
. These will be ignored during the OAuth token retrieval step, but will be passed to the broker through the existing SASL extension mechanism from KIP-342.
Once the login has occurred for this client, the returned access token can be reused by other connections from this client. While these additional connections will not issue the token retrieval HTTP call on the client, the broker will still validate the token once for each time it is sent by each client connection.
Per KIP-368, the OAuth token re-authentication logic from the existing implementation is automatically "inherited" by this implementation, so no additional work is needed to support that feature.
Validator Callback Handler (Broker)
The validation callback handler is invoked as part of the SASL Kafka server authentication process when a JWT is received. This callback handler will parse and validate the JWT using its contents along with provided configuration.
Validation Logic
The basic overview of the JWT validation that will be performed is:
Parse the JWT into separate header, payload, and signature sections
Base-64 decode the header and payload
Extract the necessary claims for validation
Match the key ID (
kid
) specified in the JWT header to a JWK ID from the JWK Set- Ensure the encoding algorithm (
alg
from the header) isn'tnone
and matches the expected algorithm for the JWK ID Encode the header and payload sections of the original encoded JWT access token using the public key from the JWK and ensure it matches the signature section of the JWT
Extract the
scope
,iat
,exp
, andsub
claims as these are needed by theOAuthBearerToken
object to be passed to theOAuthBearerValidatorCallback
.Optional claim validation that ensures that issuer, audience, or other claims match a given value
The extensions validation will be executed the same as in org.apache.kafka.common.security.oauthbearer.internals.unsecured.OAuthBearerUnsecuredValidatorCallbackHandler
today.
A new key ID (kid
) could appear in the header of an incoming JWT access token. Code that can retrieve the JWKS from the OAuth provider on demand will be implemented. The common case will be that the key ID is one that has been accessed recently, so it shouldn’t need to reach out to the JWKS endpoint often. The code will need to have a means to expunge old JWKs that are no longer needed.
Broker Configuration
The name of the implementation class will be org.apache.kafka.common.security.oauthbearer.secured.OAuthBearerValidatorCallbackHandler
and it will accept instances of org.apache.kafka.common.security.oauthbearer.OAuthBearerValidatorCallback
and org.apache.kafka.common.security.oauthbearer.OAuthBearerExtensionsValidatorCallback
. The fully-qualified class name will be provided to the broker's listener.name.<listener name>.oauthbearer.sasl.server.callback.handler.class
configuration.
It may be that the names of the claims used by the OAuth provider differ from what is expected. For example, the security principal for which the token is valid is usually contained in the sub
(subject) JWT claim. There may be cases where the value of that claim may not be valid or usable, and instead the value will need to be extracted from, for example, the email
claim.
There are several configuration options for this callback handler. Since there are no sensitive configuration options, they are all in the top-level configuration:
listener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri
: OAuth issuer's JWK Set endpoint URI from which to retrieve the set of JWKs managed by the provider; this can be afile://
-based URL that points to a broker file system-accessible file-based copy of the JWKS data. This allows the JWKS data to be updated on the file system and refreshed on the broker when the file is updated, thus avoiding any HTTP(S) communication with the OAuth/OIDC providerlistener.name.<listener name>.oauthbearer.sasl.jwks.refresh.interval.ms
: optional value in milliseconds for how often to refresh the JWKS from the URL pointed to bylistener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri
. Only used when using an HTTP(S)-based URI forlistener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri
. Defaults to3600000
(1 hour)listener.name.<listener name>.oauthbearer.sasl.sub.claim.name
: name of the scope from which to extract the subject claim from the JWT; defaults tosub
listener.name.<listener name>.oauthbearer.sasl.scope.claim.name
: name of the scope from which to extract the scope claim from the JWT; defaults toscope
listener.name.<listener name>.oauthbearer.sasl.clock.skew.seconds
: optional value in seconds for the clock skew between the OAuth/OIDC provider and the broker. Only used when using an HTTP(S)-based URI forlistener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri
. Defaults to30
listener.name.<listener name>.oauthbearer.sasl.expected.audience
: The (optional) comma-delimited setting for the broker to use to verify that the JWT was issued for one of the expected audiences. The JWT will be inspected for the standard OAuthaud
claim and if this configuration option is set, the broker will match the value from JWT'saud
claim to see if there is an exact match. If there is no match, the broker will reject the JWT and authentication will fail.listener.name.<listener name>.oauthbearer.sasl.expected.issuer
: Optional setting for the broker to use to verify that the JWT was created by the expected issuer. The JWT will be inspected for the standard OAuthiss
claim and if this configuration option is set, the broker will match the value from JWT'siss
claim to see if there is an exact match. If there is no match, the broker will reject the JWT and authentication will fail.
Here's an example of the configuration as a part of a Java properties file:
listener.name.<listener name>.oauthbearer.sasl.login.callback.handler.class=o.a.k...OAuthBearerValidatorCallbackHandlerlistener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri=https://myidp.example.com/oauth2/default/v1/keys
listener.name.<listener name>.oauthbearer.sasl.jaas.config=o.a.k...OAuthBearerLoginModule required \
scopeClaimName="scp" ;
In the above configuration the broker points to the appropriate OAuth provider listener.name.<listener name>.oauthbearer.sasl.jwks.endpoint.uri
to retrieve a the set of JWKs for validation. In this example, a non-default value for scopeClaimName
has been provided because the provider uses scp
for the name of the scope claim in the JWT it produces.
JWKS Management Logic
The JWKS will be kept up-to-date in two main ways:
Providing a JWKS URL. In this mode, the JWKS data will be retrieved from the OAuth provider via the configured URL on broker startup. All then-current keys will be cached on the broker (per the ‘max age’; the jose4j library has a means to keep these-up-to-date when they age out) for incoming requests. If an authentication request is received for a JWT that includes a
kid
that isn’t yet in the cache, the JWKS endpoint will be queried again on demand. However, we prefer polling via a background thread to hopefully pre-populate the cache with any forthcoming keys before any JWT requests that include them are received.Providing a JWKS file. On startup, the broker will load the JWKS file from the configured location. Any updates to the file will require a restart of the broker. The means by which the JWKS file is updated is left to the cluster administrator. In the event that an unknown JWT key is encountered, this implementation will simply issue an error and validation will fail.
If the the URL or file that is specified cannot be read, the broker will fail to start up. In the case of an HTTP(S)-based URI, the configured sasl.login.attempts
, sasl.login.retry.wait.ms
, and sasl.login.retry.max.wait.ms
values will be used to make attempts to connect to the remote OAuth provider.
It is also important that the JWKS is retrieved before the broker's ports are opened. Otherwise clients that connect to the broker before the JWKS is retrieved will experience spurious authentication failures (e.g. during broker restarts).
Broker-to-broker Support
The use of OAuth credentials for broker-to-broker communication will continue to be supported. As with the existing implementation, users can specify the protocols and implementations to use for broker-based communication. This would require providing the appropriate configuration for both client login and broker validation.
Testing
In addition to unit and integration tests, there will be a standalone tool in the tools directory/module: org.apache.kafka.tools.OAuthCompatibilityTest
. This test can be run via the existing bin/kafka-run-class.sh
script thusly:
KAFKA_OPTS=""
export KAFKA_OPTS="$KAFKA_OPTS -DclientId=$client_id"
export KAFKA_OPTS="$KAFKA_OPTS -DclientSecret=$client_secret"
export KAFKA_OPTS="$KAFKA_OPTS -Dscope=api://$client_id/.default"
export KAFKA_OPTS="$KAFKA_OPTS -DtokenEndpointUri=$token_endpoint_uri"
export KAFKA_OPTS="$KAFKA_OPTS -DjwksEndpointUri=$jwks_uri"
export KAFKA_OPTS="$KAFKA_OPTS -DclockSkew=30"
export KAFKA_OPTS="$KAFKA_OPTS -DexpectedAudience=$expected_audience"
./bin/kafka-run-class.sh org.apache.kafka.tools.OAuthCompatibilityTest
As seen in the example invocation above, the various Kafka configuration options for the client and server properties are passed in via the existing KAFKA_OPTS
environment variable. The test will connect to remote systems as needed to authenticate, retrieve tokens, retrieve JWKS, and perform validation (nothing is "mocked").
This tool serves three basic purposes:
- Allows validation that a given OAuth/OIDC provider is compliant with our implementation of the specification
- Provides users a means to reproduce and diagnose issues for support needs
- Provides a means to perform system tests against live providers for tracking regressions
The output of the test is easily consumable:
PASSED 1/5: client configuration
PASSED 2/5: client JWT retrieval
PASSED 3/5: client JWT validation
PASSED 4/5: broker configuration
PASSED 5/5: broker JWT validation
Additionally debugging can be selectively enabled using the standard tools-log4j.properties
file if errors are detected by the tool. All configuration options supported by the client and server are also supported by the tool; it attempts to be as faithful as possible to the runtime logic to minimize discrepancies between the tool and the client and server.
Compatibility, Deprecation, and Migration Plan
Users can continue to use the OAuthBearerUnsecuredLoginCallbackHandler
and/or other AuthenticateCallbackHandler
implementations. Users will need to update both clients and brokers in order to use the new functionality.
Rejected Alternatives
Removing OAuthBearerUnsecuredLoginCallbackHandler
Although this change will provide an out-of-the-box implementation of an AuthenticateCallbackHandler
that can be used to communicate with OAuth/OIDC, the exist unsecured implementation is still usable for development and testing. Given that its non-secure status is in its name and the documentation, it shouldn’t need to be removed or deprecated at this time.
Standalone Plugin
Technically the new implementation could be developed and shipped as a plugin separate from the main Apache Kafka project. Community adoption would be improved by an in-tree solution. Including this inside Apache Kafka doesn’t preclude alternative AuthenticateCallbackHandler
implementations from use by clients, if desired.