mirror of
https://github.com/zitadel/zitadel.git
synced 2025-08-11 21:37:32 +00:00
perf(milestones): refactor (#8788)
Some checks are pending
ZITADEL CI/CD / core (push) Waiting to run
ZITADEL CI/CD / console (push) Waiting to run
ZITADEL CI/CD / version (push) Waiting to run
ZITADEL CI/CD / compile (push) Blocked by required conditions
ZITADEL CI/CD / core-unit-test (push) Blocked by required conditions
ZITADEL CI/CD / core-integration-test (push) Blocked by required conditions
ZITADEL CI/CD / lint (push) Blocked by required conditions
ZITADEL CI/CD / container (push) Blocked by required conditions
ZITADEL CI/CD / e2e (push) Blocked by required conditions
ZITADEL CI/CD / release (push) Blocked by required conditions
Code Scanning / CodeQL-Build (go) (push) Waiting to run
Code Scanning / CodeQL-Build (javascript) (push) Waiting to run
Some checks are pending
ZITADEL CI/CD / core (push) Waiting to run
ZITADEL CI/CD / console (push) Waiting to run
ZITADEL CI/CD / version (push) Waiting to run
ZITADEL CI/CD / compile (push) Blocked by required conditions
ZITADEL CI/CD / core-unit-test (push) Blocked by required conditions
ZITADEL CI/CD / core-integration-test (push) Blocked by required conditions
ZITADEL CI/CD / lint (push) Blocked by required conditions
ZITADEL CI/CD / container (push) Blocked by required conditions
ZITADEL CI/CD / e2e (push) Blocked by required conditions
ZITADEL CI/CD / release (push) Blocked by required conditions
Code Scanning / CodeQL-Build (go) (push) Waiting to run
Code Scanning / CodeQL-Build (javascript) (push) Waiting to run
# Which Problems Are Solved Milestones used existing events from a number of aggregates. OIDC session is one of them. We noticed in load-tests that the reduction of the oidc_session.added event into the milestone projection is a costly business with payload based conditionals. A milestone is reached once, but even then we remain subscribed to the OIDC events. This requires the projections.current_states to be updated continuously. # How the Problems Are Solved The milestone creation is refactored to use dedicated events instead. The command side decides when a milestone is reached and creates the reached event once for each milestone when required. # Additional Changes In order to prevent reached milestones being created twice, a migration script is provided. When the old `projections.milestones` table exist, the state is read from there and `v2` milestone aggregate events are created, with the original reached and pushed dates. # Additional Context - Closes https://github.com/zitadel/zitadel/issues/8800
This commit is contained in:
@@ -16,6 +16,7 @@ import (
|
||||
|
||||
"github.com/zitadel/zitadel/internal/integration"
|
||||
"github.com/zitadel/zitadel/internal/integration/sink"
|
||||
"github.com/zitadel/zitadel/internal/repository/milestone"
|
||||
"github.com/zitadel/zitadel/pkg/grpc/app"
|
||||
"github.com/zitadel/zitadel/pkg/grpc/management"
|
||||
"github.com/zitadel/zitadel/pkg/grpc/object"
|
||||
@@ -32,12 +33,12 @@ func TestServer_TelemetryPushMilestones(t *testing.T) {
|
||||
|
||||
instance := integration.NewInstance(CTX)
|
||||
iamOwnerCtx := instance.WithAuthorization(CTX, integration.UserTypeIAMOwner)
|
||||
t.Log("testing against instance with primary domain", instance.Domain)
|
||||
awaitMilestone(t, sub, instance.Domain, "InstanceCreated")
|
||||
t.Log("testing against instance", instance.ID())
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.InstanceCreated)
|
||||
|
||||
projectAdded, err := instance.Client.Mgmt.AddProject(iamOwnerCtx, &management.AddProjectRequest{Name: "integration"})
|
||||
require.NoError(t, err)
|
||||
awaitMilestone(t, sub, instance.Domain, "ProjectCreated")
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.ProjectCreated)
|
||||
|
||||
redirectURI := "http://localhost:8888"
|
||||
application, err := instance.Client.Mgmt.AddOIDCApp(iamOwnerCtx, &management.AddOIDCAppRequest{
|
||||
@@ -52,14 +53,14 @@ func TestServer_TelemetryPushMilestones(t *testing.T) {
|
||||
AccessTokenType: app.OIDCTokenType_OIDC_TOKEN_TYPE_JWT,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
awaitMilestone(t, sub, instance.Domain, "ApplicationCreated")
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.ApplicationCreated)
|
||||
|
||||
// create the session to be used for the authN of the clients
|
||||
sessionID, sessionToken, _, _ := instance.CreatePasswordSession(t, iamOwnerCtx, instance.AdminUserID, "Password1!")
|
||||
|
||||
console := consoleOIDCConfig(t, instance)
|
||||
loginToClient(t, instance, console.GetClientId(), console.GetRedirectUris()[0], sessionID, sessionToken)
|
||||
awaitMilestone(t, sub, instance.Domain, "AuthenticationSucceededOnInstance")
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.AuthenticationSucceededOnInstance)
|
||||
|
||||
// make sure the client has been projected
|
||||
require.EventuallyWithT(t, func(collectT *assert.CollectT) {
|
||||
@@ -70,11 +71,11 @@ func TestServer_TelemetryPushMilestones(t *testing.T) {
|
||||
assert.NoError(collectT, err)
|
||||
}, time.Minute, time.Second, "app not found")
|
||||
loginToClient(t, instance, application.GetClientId(), redirectURI, sessionID, sessionToken)
|
||||
awaitMilestone(t, sub, instance.Domain, "AuthenticationSucceededOnApplication")
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.AuthenticationSucceededOnApplication)
|
||||
|
||||
_, err = integration.SystemClient().RemoveInstance(CTX, &system.RemoveInstanceRequest{InstanceId: instance.ID()})
|
||||
require.NoError(t, err)
|
||||
awaitMilestone(t, sub, instance.Domain, "InstanceDeleted")
|
||||
awaitMilestone(t, sub, instance.ID(), milestone.InstanceDeleted)
|
||||
}
|
||||
|
||||
func loginToClient(t *testing.T, instance *integration.Instance, clientID, redirectURI, sessionID, sessionToken string) {
|
||||
@@ -134,7 +135,7 @@ func consoleOIDCConfig(t *testing.T, instance *integration.Instance) *app.OIDCCo
|
||||
return apps.GetResult()[0].GetOidcConfig()
|
||||
}
|
||||
|
||||
func awaitMilestone(t *testing.T, sub *sink.Subscription, primaryDomain, expectMilestoneType string) {
|
||||
func awaitMilestone(t *testing.T, sub *sink.Subscription, instanceID string, expectMilestoneType milestone.Type) {
|
||||
for {
|
||||
select {
|
||||
case req := <-sub.Recv():
|
||||
@@ -144,17 +145,17 @@ func awaitMilestone(t *testing.T, sub *sink.Subscription, primaryDomain, expectM
|
||||
}
|
||||
t.Log("received milestone", plain.String())
|
||||
milestone := struct {
|
||||
Type string `json:"type"`
|
||||
PrimaryDomain string `json:"primaryDomain"`
|
||||
InstanceID string `json:"instanceId"`
|
||||
Type milestone.Type `json:"type"`
|
||||
}{}
|
||||
if err := json.Unmarshal(req.Body, &milestone); err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
if milestone.Type == expectMilestoneType && milestone.PrimaryDomain == primaryDomain {
|
||||
if milestone.Type == expectMilestoneType && milestone.InstanceID == instanceID {
|
||||
return
|
||||
}
|
||||
case <-time.After(2 * time.Minute): // why does it take so long to get a milestone !?
|
||||
t.Fatalf("timed out waiting for milestone %s in domain %s", expectMilestoneType, primaryDomain)
|
||||
case <-time.After(20 * time.Second):
|
||||
t.Fatalf("timed out waiting for milestone %s for instance %s", expectMilestoneType, instanceID)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
Reference in New Issue
Block a user