[ { "Release": "4.22", "TestName": "verify the cluster readiness and stability", "TotalTestCount": 6825, "TotalFailureCount": 525, "TotalFlakeCount": 0, "FailureRate": 0.07692307692307693, "AvgDurationMs": 132.5978542510121, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be rejected for evicted pods (no endpoints exist)", "TotalTestCount": 4185, "TotalFailureCount": 253, "TotalFlakeCount": 0, "FailureRate": 0.06045400238948626, "AvgDurationMs": 116221.22122194515, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:ImageStreamImportMode] ClusterVersion API desired architecture should be valid when architecture is set in release payload metadata [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 165, "TotalFlakeCount": 4, "FailureRate": 0.04041146216017634, "AvgDurationMs": 1075.8817366946773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify all nodes should be ready", "TotalTestCount": 6823, "TotalFailureCount": 126, "TotalFlakeCount": 0, "FailureRate": 0.018466950021984466, "AvgDurationMs": 83.45340080971656, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify node count should match or exceed machine count", "TotalTestCount": 6822, "TotalFailureCount": 123, "TotalFlakeCount": 0, "FailureRate": 0.01802990325417766, "AvgDurationMs": 0.00006072874493927125, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions openshift-apiserver", "TotalTestCount": 6807, "TotalFailureCount": 121, "TotalFlakeCount": 0, "FailureRate": 0.01777581900984281, "AvgDurationMs": 0.00008097165991902833, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions authentication", "TotalTestCount": 6807, "TotalFailureCount": 120, "TotalFlakeCount": 0, "FailureRate": 0.01762891141472014, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify all machines should be in Running state", "TotalTestCount": 6823, "TotalFailureCount": 115, "TotalFlakeCount": 0, "FailureRate": 0.016854755972446137, "AvgDurationMs": 64.20542510121459, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestScopedProjectAccess should succeed [apigroup:user.openshift.io][apigroup:project.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 107, "TotalFlakeCount": 85, "FailureRate": 0.02620622091599314, "AvgDurationMs": 211033.83518207277, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMWebhookProviderOpenshiftServiceCA] OLMv1 operator with webhooks should have a working validating webhook", "TotalTestCount": 3804, "TotalFailureCount": 107, "TotalFlakeCount": 0, "FailureRate": 0.028128286014721345, "AvgDurationMs": 179680.8286005831, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer][Late] kubelet terminates kube-apiserver gracefully extended [Suite:openshift/conformance/parallel]", "TotalTestCount": 5233, "TotalFailureCount": 106, "TotalFlakeCount": 1, "FailureRate": 0.02025606726543092, "AvgDurationMs": 2026.9987714987715, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][FeatureGate:VSphereDriverConfiguration][Serial][apigroup:operator.openshift.io] vSphere CSI Driver Configuration snapshot options in clusterCSIDriver should allow all limits to be set at once [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 82, "TotalFlakeCount": 0, "FailureRate": 0.1156558533145275, "AvgDurationMs": 7079.239710144927, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-ci] [Early] prow job name should match feature set [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 78, "TotalFlakeCount": 0, "FailureRate": 0.014871306005719733, "AvgDurationMs": 2839.6586977886973, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][FeatureGate:VSphereDriverConfiguration][Serial][apigroup:operator.openshift.io] vSphere CSI Driver Configuration snapshot options in clusterCSIDriver should allow setting VVOL limit [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 77, "TotalFlakeCount": 0, "FailureRate": 0.10860366713681241, "AvgDurationMs": 5503.300917874396, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with revision history limits should never persist more old deployments than acceptable after being observed by the controller [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 76, "TotalFlakeCount": 39, "FailureRate": 0.0198744769874477, "AvgDurationMs": 134078.51638235303, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with test deployments should run a deployment to completion and then scale to zero [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 76, "TotalFlakeCount": 59, "FailureRate": 0.0198744769874477, "AvgDurationMs": 138181.58291176474, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:Image] oc tag should change image reference for internal images [apigroup:build.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 75, "TotalFlakeCount": 26, "FailureRate": 0.01961297071129707, "AvgDurationMs": 69267.4078529412, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM] OLMv1 operator installation should block cluster upgrades if an incompatible operator is installed", "TotalTestCount": 3804, "TotalFailureCount": 75, "TotalFlakeCount": 0, "FailureRate": 0.01971608832807571, "AvgDurationMs": 70603.16148688046, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs when tagging images should successfully tag the deployed image [apigroup:apps.openshift.io][apigroup:authorization.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 74, "TotalFlakeCount": 2, "FailureRate": 0.019351464435146442, "AvgDurationMs": 20204.975352941176, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][FeatureGate:VSphereDriverConfiguration][Serial][apigroup:operator.openshift.io] vSphere CSI Driver Configuration snapshot options in clusterCSIDriver should allow setting VSAN limit [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 72, "TotalFlakeCount": 0, "FailureRate": 0.10155148095909731, "AvgDurationMs": 4933.958985507248, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][FeatureGate:VSphereDriverConfiguration][Serial][apigroup:operator.openshift.io] vSphere CSI Driver Configuration snapshot options in clusterCSIDriver should allow setting global snapshot limit [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 71, "TotalFlakeCount": 0, "FailureRate": 0.1001410437235543, "AvgDurationMs": 16348.50724637681, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with valid namespace labels on when fsgroup should default to OnRootMismatch if pod has none", "TotalTestCount": 4083, "TotalFailureCount": 70, "TotalFlakeCount": 0, "FailureRate": 0.017144256674014206, "AvgDurationMs": 78066.52882352941, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with valid namespace labels on when fsgroup should not override fsgroup change policy if pod already has one", "TotalTestCount": 4083, "TotalFailureCount": 68, "TotalFlakeCount": 0, "FailureRate": 0.01665442076904237, "AvgDurationMs": 77834.30028011206, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with valid namespace labels on when selinux should not override selinux change policy if pod already has one", "TotalTestCount": 4083, "TotalFailureCount": 68, "TotalFlakeCount": 0, "FailureRate": 0.01665442076904237, "AvgDurationMs": 80051.88610644257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestProjectWatch should succeed [apigroup:project.openshift.io][apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 67, "TotalFlakeCount": 30, "FailureRate": 0.016409502816556454, "AvgDurationMs": 103388.21352941178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Managed cluster should have no crashlooping recycler pods over four minutes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 67, "TotalFlakeCount": 10, "FailureRate": 0.015835499881824627, "AvgDurationMs": 238451.84062344136, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs adoption will orphan all RCs and adopt them back when recreated [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 66, "TotalFlakeCount": 9, "FailureRate": 0.017259414225941423, "AvgDurationMs": 61475.09155882353, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM] OLMv1 operator installation should install a cluster extension", "TotalTestCount": 3804, "TotalFailureCount": 66, "TotalFlakeCount": 0, "FailureRate": 0.017350157728706624, "AvgDurationMs": 69143.6637317784, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure HTTPRoute object is created [Suite:openshift/conformance/parallel]", "TotalTestCount": 4081, "TotalFailureCount": 64, "TotalFlakeCount": 16, "FailureRate": 0.0156824307767704, "AvgDurationMs": 76774.17254901958, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][sig-node][Feature:Builds][apigroup:build.openshift.io] zstd:chunked Image should successfully run date command [Suite:openshift/conformance/parallel]", "TotalTestCount": 3833, "TotalFailureCount": 62, "TotalFlakeCount": 8, "FailureRate": 0.016175319593008088, "AvgDurationMs": 236897.16093567252, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance readiness test should report ready soon after all annotated objects are ready [apigroup:template.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3833, "TotalFailureCount": 62, "TotalFlakeCount": 20, "FailureRate": 0.016175319593008088, "AvgDurationMs": 63277.312456140346, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLayers] Image layer subresource should return layers from tagged images [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 62, "TotalFlakeCount": 19, "FailureRate": 0.016213389121338913, "AvgDurationMs": 69152.97635294117, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM] OLMv1 operator installation should fail to install a non-existing cluster extension", "TotalTestCount": 3804, "TotalFailureCount": 62, "TotalFlakeCount": 0, "FailureRate": 0.01629863301787592, "AvgDurationMs": 68568.21329446063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with valid namespace labels on when selinux should default to selinux label of namespace if pod has none", "TotalTestCount": 4083, "TotalFailureCount": 61, "TotalFlakeCount": 0, "FailureRate": 0.01493999510164095, "AvgDurationMs": 77913.7505882353, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with multiple image change triggers should run a successful deployment with multiple triggers [apigroup:apps.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 59, "TotalFlakeCount": 11, "FailureRate": 0.01542887029288703, "AvgDurationMs": 57027.492558823535, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs when changing image change trigger should successfully trigger from an updated image [apigroup:apps.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 58, "TotalFlakeCount": 16, "FailureRate": 0.015167364016736401, "AvgDurationMs": 32155.527294117644, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Token Expiration] Using a OAuth client with a non-default token max age [apigroup:oauth.openshift.io] to generate tokens that do not expire works as expected when using a token authorization flow [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 58, "TotalFlakeCount": 56, "FailureRate": 0.014367104285360416, "AvgDurationMs": 28002.430252100847, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] OpenShift alerting rules [apigroup:image.openshift.io] should link to a valid URL if the runbook_url annotation is defined [Suite:openshift/conformance/parallel]", "TotalTestCount": 4920, "TotalFailureCount": 58, "TotalFlakeCount": 0, "FailureRate": 0.011788617886178862, "AvgDurationMs": 30051.97247838616, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge] DNS should answer queries using the local DNS endpoint [Suite:openshift/conformance/parallel]", "TotalTestCount": 4198, "TotalFailureCount": 58, "TotalFlakeCount": 7, "FailureRate": 0.013816102906145784, "AvgDurationMs": 25988.460666666666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs viewing rollout history should print the rollout history [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 55, "TotalFlakeCount": 12, "FailureRate": 0.014382845188284518, "AvgDurationMs": 31293.04797058824, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with custom deployments should run the custom deployment steps [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 55, "TotalFlakeCount": 14, "FailureRate": 0.014382845188284518, "AvgDurationMs": 39267.27208823531, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs initially should not deploy if pods never transition to ready [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 54, "TotalFlakeCount": 17, "FailureRate": 0.01344286781179985, "AvgDurationMs": 44532.52, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs when run iteratively should only deploy the last deployment [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 54, "TotalFlakeCount": 8, "FailureRate": 0.014121338912133892, "AvgDurationMs": 40968.631205882346, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with failing hook should get all logs from retried hooks [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 54, "TotalFlakeCount": 17, "FailureRate": 0.014121338912133892, "AvgDurationMs": 15371.17958823529, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry] Image registry [apigroup:route.openshift.io] should redirect on blob pull [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 54, "TotalFlakeCount": 10, "FailureRate": 0.01322556943423953, "AvgDurationMs": 33850.64966386555, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling [apigroup:apps.openshift.io][apigroup:route.openshift.io] should work with UDP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4081, "TotalFailureCount": 52, "TotalFlakeCount": 9, "FailureRate": 0.01274197500612595, "AvgDurationMs": 38501.008599439774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs generation should deploy based on a status version bump [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 51, "TotalFlakeCount": 8, "FailureRate": 0.013336820083682008, "AvgDurationMs": 22399.70173529412, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:HTPasswdAuth] HTPasswd IDP should successfully configure htpasswd and be responsive [apigroup:user.openshift.io][apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 51, "TotalFlakeCount": 29, "FailureRate": 0.012633143423334159, "AvgDurationMs": 26523.423837535014, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] kube-apiserver should be accessible via service network endpoint [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4232, "TotalFailureCount": 50, "TotalFlakeCount": 6, "FailureRate": 0.011814744801512287, "AvgDurationMs": 9063.101620947635, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs when run iteratively should immediately start a new deployment [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 50, "TotalFlakeCount": 8, "FailureRate": 0.012447099825740602, "AvgDurationMs": 10362.00388405797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:Image] oc tag should work when only imagestreams api is available [apigroup:image.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 50, "TotalFlakeCount": 11, "FailureRate": 0.012853470437017995, "AvgDurationMs": 6091.770454545455, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster shouldn't have failing rules evaluation [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 50, "TotalFlakeCount": 1, "FailureRate": 0.013071895424836602, "AvgDurationMs": 4447.523147058821, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure GIE is enabled after creating an inferencePool CRD [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 50, "TotalFlakeCount": 16, "FailureRate": 0.01224889759921607, "AvgDurationMs": 41117.11780898877, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][Feature:Machines][Serial] Managed cluster should grow and decrease when scaling different machineSets simultaneously [Timeout:30m][apigroup:machine.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 697, "TotalFailureCount": 49, "TotalFlakeCount": 0, "FailureRate": 0.0703012912482066, "AvgDurationMs": 439862.70888324885, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageExtract] Image extract should extract content from an image [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 49, "TotalFlakeCount": 12, "FailureRate": 0.012945838837516512, "AvgDurationMs": 69064.57026470586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:commatrix][apigroup:config.openshift.io][Serial] should validate the communication matrix ports match the node's listening ports [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 49, "TotalFlakeCount": 0, "FailureRate": 0.06930693069306931, "AvgDurationMs": 16287.538115942027, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug ensure that the label is set for node debug [Suite:openshift/conformance/parallel]", "TotalTestCount": 4166, "TotalFailureCount": 48, "TotalFlakeCount": 17, "FailureRate": 0.011521843494959194, "AvgDurationMs": 14729.883830334193, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal][Serial] Baremetal platform should ensure [apigroup:config.openshift.io] cluster baremetal operator and metal3 deployment return back healthy after they are deleted [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 48, "TotalFlakeCount": 0, "FailureRate": 0.06770098730606489, "AvgDurationMs": 37402.66410628021, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure OSSM and OLM related resources are created after creating GatewayClass [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 48, "TotalFlakeCount": 9, "FailureRate": 0.011756061719324026, "AvgDurationMs": 38830.931792717085, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug deployment from a build [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 47, "TotalFlakeCount": 13, "FailureRate": 0.012287581699346406, "AvgDurationMs": 92236.65026470588, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:LDAP] LDAP IDP should authenticate against an ldap server [apigroup:user.openshift.io][apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 46, "TotalFlakeCount": 8, "FailureRate": 0.01139459995045826, "AvgDurationMs": 67631.92221288514, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Rolebinding restrictions tests single project should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 46, "TotalFlakeCount": 3, "FailureRate": 0.011266225814352193, "AvgDurationMs": 3943.4115406162455, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal][Serial] Baremetal platform should skip inspection when disabled by annotation [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 46, "TotalFlakeCount": 0, "FailureRate": 0.06199460916442048, "AvgDurationMs": 5674.165707762559, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should provide named network metrics [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 46, "TotalFlakeCount": 13, "FailureRate": 0.012026143790849673, "AvgDurationMs": 72517.77761764704, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] ClientSecretWithPlus should create oauthclient [apigroup:oauth.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 45, "TotalFlakeCount": 3, "FailureRate": 0.011146891255883081, "AvgDurationMs": 4898.184201680675, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling [apigroup:apps.openshift.io][apigroup:route.openshift.io] should work with TCP (when fully idled) [Suite:openshift/conformance/parallel]", "TotalTestCount": 4081, "TotalFailureCount": 45, "TotalFlakeCount": 5, "FailureRate": 0.011026709139916687, "AvgDurationMs": 27387.67151260505, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling with Deployments [apigroup:route.openshift.io] should work with UDP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 45, "TotalFlakeCount": 16, "FailureRate": 0.011024007839294463, "AvgDurationMs": 31922.342689075627, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume when subPath is used should fail to mount image volume with invalid subPath [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 45, "TotalFlakeCount": 0, "FailureRate": 0.010635783502718033, "AvgDurationMs": 10565.781546134664, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"control-plane-node.kubeconfig\" should be present in all kube-apiserver containers [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4220, "TotalFailureCount": 44, "TotalFlakeCount": 15, "FailureRate": 0.01042654028436019, "AvgDurationMs": 28429.31773067332, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs paused should disable actions on deployments [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 44, "TotalFlakeCount": 9, "FailureRate": 0.011506276150627616, "AvgDurationMs": 18089.446000000004, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs rolled back should rollback to an older deployment [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 44, "TotalFlakeCount": 12, "FailureRate": 0.011506276150627616, "AvgDurationMs": 41448.847529411774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs should respect image stream tag reference policy resolve the image pull spec [apigroup:apps.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 44, "TotalFlakeCount": 9, "FailureRate": 0.011506276150627616, "AvgDurationMs": 7872.88694117647, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl whitelists net.ipv4.ip_unprivileged_port_start [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 44, "TotalFlakeCount": 6, "FailureRate": 0.010399432758213189, "AvgDurationMs": 19235.855885286783, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl whitelists net.ipv4.ping_group_range [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 44, "TotalFlakeCount": 2, "FailureRate": 0.010399432758213189, "AvgDurationMs": 19288.468304239406, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Token Expiration] Using a OAuth client with a non-default token max age [apigroup:oauth.openshift.io] to generate tokens that expire shortly works as expected when using a code authorization flow [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 44, "TotalFlakeCount": 37, "FailureRate": 0.010899182561307902, "AvgDurationMs": 26230.231624649863, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle [apigroup:apps.openshift.io][apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by name [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 44, "TotalFlakeCount": 4, "FailureRate": 0.01095344784665173, "AvgDurationMs": 14066.120463768117, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc set image can set images for pods and deployments [apigroup:image.openshift.io][apigroup:apps.openshift.io][Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 44, "TotalFlakeCount": 13, "FailureRate": 0.01131396245821548, "AvgDurationMs": 13132.17775568182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance readiness test should report failed soon after an annotated objects has failed [apigroup:template.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 44, "TotalFlakeCount": 4, "FailureRate": 0.01131396245821548, "AvgDurationMs": 17210.874346590917, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Conformance][Area:Networking][Feature:Router] The HAProxy router should pass the gRPC interoperability tests [apigroup:route.openshift.io][apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4016, "TotalFailureCount": 44, "TotalFlakeCount": 22, "FailureRate": 0.010956175298804782, "AvgDurationMs": 67450.01892753625, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Managed cluster should verify that nodes have no unexpected reboots [Late] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5169, "TotalFailureCount": 44, "TotalFlakeCount": 3, "FailureRate": 0.008512284774617915, "AvgDurationMs": 25217.04234848485, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Conformance] Prevent openshift node labeling on update by the node TestOpenshiftNodeLabeling [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4199, "TotalFailureCount": 44, "TotalFlakeCount": 17, "FailureRate": 0.01047868540128602, "AvgDurationMs": 14565.906871794872, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] kube-apiserver should be accessible via api-int endpoint [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4232, "TotalFailureCount": 43, "TotalFlakeCount": 2, "FailureRate": 0.010160680529300567, "AvgDurationMs": 11849.685910224442, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] poddisruptionbudgets with unhealthyPodEvictionPolicy should evict according to the IfHealthyBudget policy [Suite:openshift/conformance/parallel]", "TotalTestCount": 4038, "TotalFailureCount": 43, "TotalFlakeCount": 6, "FailureRate": 0.010648836057454186, "AvgDurationMs": 655983.5830050505, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the token URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 43, "TotalFlakeCount": 2, "FailureRate": 0.010651473866732722, "AvgDurationMs": 27341.264313725496, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:PodSecurity][Feature:SCC] required-scc annotation is being applied to workloads [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 43, "TotalFlakeCount": 4, "FailureRate": 0.010163082013708344, "AvgDurationMs": 3544.4506483790524, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle [apigroup:apps.openshift.io][apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by label [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 43, "TotalFlakeCount": 4, "FailureRate": 0.010704505850136919, "AvgDurationMs": 14245.630289855068, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] templates process [apigroup:template.openshift.io][Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 43, "TotalFlakeCount": 6, "FailureRate": 0.011053984575835476, "AvgDurationMs": 32121.793465909086, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex] check registry.redhat.io is available and samples operator can import sample imagestreams run sample related validations [apigroup:config.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 43, "TotalFlakeCount": 39, "FailureRate": 0.01105682694780149, "AvgDurationMs": 99723.73250000003, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs should adhere to Three Laws of Controllers [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 42, "TotalFlakeCount": 5, "FailureRate": 0.010983263598326359, "AvgDurationMs": 56060.662176470585, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with minimum ready seconds set should not transition the deployment to Complete before satisfied [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 42, "TotalFlakeCount": 3, "FailureRate": 0.010983263598326359, "AvgDurationMs": 75732.26661764705, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl whitelists kernel.shm_rmid_forced [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 42, "TotalFlakeCount": 2, "FailureRate": 0.009926731269203497, "AvgDurationMs": 19820.746533665835, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl whitelists net.ipv4.ip_local_port_range [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 42, "TotalFlakeCount": 3, "FailureRate": 0.009926731269203497, "AvgDurationMs": 20375.599426433917, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm storage-admin [apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 42, "TotalFlakeCount": 6, "FailureRate": 0.010286554004408524, "AvgDurationMs": 55604.26285714285, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance impersonation tests [apigroup:user.openshift.io][apigroup:authorization.openshift.io] should pass impersonation creation tests [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 42, "TotalFlakeCount": 2, "FailureRate": 0.010286554004408524, "AvgDurationMs": 15762.287478991599, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageAppend] Image append should create images by appending them [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 42, "TotalFlakeCount": 4, "FailureRate": 0.011096433289299868, "AvgDurationMs": 24537.059676470595, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure LB, service, and dnsRecord are created for a Gateway object [Suite:openshift/conformance/parallel]", "TotalTestCount": 4080, "TotalFailureCount": 42, "TotalFlakeCount": 9, "FailureRate": 0.010294117647058823, "AvgDurationMs": 51014.84456582633, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume should succeed with multiple pods and same image on the same node [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 42, "TotalFlakeCount": 5, "FailureRate": 0.009926731269203497, "AvgDurationMs": 17614.7416957606, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the login URL for the allow all IDP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 41, "TotalFlakeCount": 0, "FailureRate": 0.010156056477582363, "AvgDurationMs": 26158.080420168073, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the login URL for when there is only one IDP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 41, "TotalFlakeCount": 1, "FailureRate": 0.010156056477582363, "AvgDurationMs": 26680.363305322127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestUnprivilegedNewProject [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 41, "TotalFlakeCount": 5, "FailureRate": 0.010041636051922606, "AvgDurationMs": 3594.8010084033613, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle Deployments [apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by label [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 41, "TotalFlakeCount": 6, "FailureRate": 0.01020408163265306, "AvgDurationMs": 13549.404927536236, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance impersonation tests [apigroup:user.openshift.io][apigroup:authorization.openshift.io] should pass impersonation update tests [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 41, "TotalFlakeCount": 6, "FailureRate": 0.010041636051922606, "AvgDurationMs": 19519.311736694686, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure custom gatewayclass can be accepted [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 41, "TotalFlakeCount": 9, "FailureRate": 0.010041636051922606, "AvgDurationMs": 43801.12691876752, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] kube-apiserver should be accessible via api-ext endpoint [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 40, "TotalFlakeCount": 2, "FailureRate": 0.009454029780193808, "AvgDurationMs": 11328.569401496261, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"check-endpoints.kubeconfig\" should be present in all kube-apiserver containers [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4220, "TotalFailureCount": 40, "TotalFlakeCount": 18, "FailureRate": 0.009478672985781991, "AvgDurationMs": 27827.29528678305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ResourceQuota] Object count should properly count the number of imagestreams resources [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 40, "TotalFlakeCount": 57, "FailureRate": 0.00979671809943669, "AvgDurationMs": 6365.162801120447, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][apigroup:apps.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] Evaluate DaemonSet placement in HighlyAvailableArbiterMode topology should not create a DaemonSet on the Arbiter node [Suite:openshift/conformance/parallel]", "TotalTestCount": 4092, "TotalFailureCount": 40, "TotalFlakeCount": 0, "FailureRate": 0.009775171065493646, "AvgDurationMs": 3865.388319327731, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] authorization TestAuthorizationSubjectAccessReview should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 40, "TotalFlakeCount": 4, "FailureRate": 0.00979671809943669, "AvgDurationMs": 20352.05058823529, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SecurityContextConstraints] TestAllowedSCCViaRBAC [apigroup:project.openshift.io][apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 40, "TotalFlakeCount": 5, "FailureRate": 0.00979671809943669, "AvgDurationMs": 16417.737142857146, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc builds new-build [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 40, "TotalFlakeCount": 6, "FailureRate": 0.010285420416559526, "AvgDurationMs": 21606.154602272727, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle] CSRs from machines that are not recognized by the cloud provider are not approved [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 40, "TotalFlakeCount": 1, "FailureRate": 0.009526077637532746, "AvgDurationMs": 31981.146512820516, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance security tests [apigroup:authorization.openshift.io][apigroup:template.openshift.io] should pass security tests [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 40, "TotalFlakeCount": 11, "FailureRate": 0.00979671809943669, "AvgDurationMs": 17425.76442577031, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for oauth.openshift.io/v1, Resource=oauthclientauthorizations [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 39, "TotalFlakeCount": 1, "FailureRate": 0.009551800146950772, "AvgDurationMs": 3701.8643977591037, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] OAuth server has the correct token and certificate fallback semantics [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4055, "TotalFailureCount": 39, "TotalFlakeCount": 11, "FailureRate": 0.009617755856966708, "AvgDurationMs": 5103.199255014326, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the grant URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 39, "TotalFlakeCount": 1, "FailureRate": 0.009660639088432003, "AvgDurationMs": 26448.49588235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the login URL for the bootstrap IDP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 39, "TotalFlakeCount": 2, "FailureRate": 0.009660639088432003, "AvgDurationMs": 25707.50717086834, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Token Expiration] Using a OAuth client with a non-default token max age [apigroup:oauth.openshift.io] to generate tokens that expire shortly works as expected when using a token authorization flow [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 39, "TotalFlakeCount": 39, "FailureRate": 0.009660639088432003, "AvgDurationMs": 26753.545182072838, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc --request-timeout works as expected [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 39, "TotalFlakeCount": 14, "FailureRate": 0.009554140127388535, "AvgDurationMs": 5975.2195798319335, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm policy [apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 39, "TotalFlakeCount": 4, "FailureRate": 0.009551800146950772, "AvgDurationMs": 15828.34176470588, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can show correct whoami result with console [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 39, "TotalFlakeCount": 1, "FailureRate": 0.009287925696594427, "AvgDurationMs": 3299.2521538461533, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain list uncovered GroupVersionResources [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 39, "TotalFlakeCount": 1, "FailureRate": 0.009217679035688962, "AvgDurationMs": 3802.337905236908, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle Deployments [apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by name [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 39, "TotalFlakeCount": 6, "FailureRate": 0.009706321553011448, "AvgDurationMs": 13452.761391304348, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning node validation should have correct cpuset and cpushare set in crio containers [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 39, "TotalFlakeCount": 43, "FailureRate": 0.009579955784819455, "AvgDurationMs": 22939.928795518204, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Serial][sig-cli] oc adm upgrade recommend When the update service has conditional recommendations runs successfully with conditional recommendations to the --version target [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 38, "TotalFlakeCount": 0, "FailureRate": 0.05121293800539083, "AvgDurationMs": 20605.03114155251, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a custom resource.", "TotalTestCount": 4231, "TotalFailureCount": 38, "TotalFlakeCount": 0, "FailureRate": 0.008981328291184117, "AvgDurationMs": 26928.765635910226, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestTriggers_imageChange_nonAutomatic [apigroup:image.openshift.io][apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 38, "TotalFlakeCount": 5, "FailureRate": 0.009309162175404213, "AvgDurationMs": 43065.325462184876, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][apigroup:apps.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] Deployments on HighlyAvailableArbiterMode topology should be created on arbiter nodes when arbiter node is selected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4096, "TotalFailureCount": 38, "TotalFlakeCount": 0, "FailureRate": 0.00927734375, "AvgDurationMs": 3341.117843137257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:LDAP] LDAP should start an OpenLDAP test server [apigroup:user.openshift.io][apigroup:security.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 38, "TotalFlakeCount": 6, "FailureRate": 0.009306882194464855, "AvgDurationMs": 32776.19014005602, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] RBAC proxy for openshift authz RunLegacyEndpointConfirmNoEscalation [apigroup:authorization.openshift.io] should succeed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 38, "TotalFlakeCount": 4, "FailureRate": 0.009306882194464855, "AvgDurationMs": 21800.900168067226, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SecurityContextConstraints] TestPodUpdateSCCEnforcement with service account [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 38, "TotalFlakeCount": 29, "FailureRate": 0.008981328291184117, "AvgDurationMs": 4021.328478802992, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc builds patch buildconfig [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 38, "TotalFlakeCount": 12, "FailureRate": 0.00977114939573155, "AvgDurationMs": 4764.666619318182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle [apigroup:apps.openshift.io][apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by checking previous scale [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 38, "TotalFlakeCount": 7, "FailureRate": 0.009459795867562858, "AvgDurationMs": 14402.963362318842, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc probe can ensure the probe command is functioning as expected on deploymentconfigs [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 38, "TotalFlakeCount": 9, "FailureRate": 0.009459795867562858, "AvgDurationMs": 8267.053362318842, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Idling with a single service and ReplicationController should idle the service and ReplicationController properly [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 38, "TotalFlakeCount": 3, "FailureRate": 0.008981328291184117, "AvgDurationMs": 26499.604663341648, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume should succeed with pod and pull policy of Always [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 38, "TotalFlakeCount": 3, "FailureRate": 0.008981328291184117, "AvgDurationMs": 11999.106957605985, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeature:StorageNetworkPolicy] Storage Network Policy should ensure required NetworkPolicies exist with correct labels [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 38, "TotalFlakeCount": 4, "FailureRate": 0.008981328291184117, "AvgDurationMs": 3298.9237406483785, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"localhost-recovery.kubeconfig\" should be present on all masters and work [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4160, "TotalFailureCount": 37, "TotalFlakeCount": 3, "FailureRate": 0.008894230769230769, "AvgDurationMs": 14073.789895287955, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with multiple image change triggers should run a successful deployment with a trigger used by different containers [apigroup:apps.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 37, "TotalFlakeCount": 5, "FailureRate": 0.009675732217573221, "AvgDurationMs": 41102.94829411764, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] FIPS TestFIPS [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4133, "TotalFailureCount": 37, "TotalFlakeCount": 12, "FailureRate": 0.008952334865714976, "AvgDurationMs": 9935.880238726792, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl pod should not start for sysctl not on whitelist net.ipv4.ip_dynaddr [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 37, "TotalFlakeCount": 3, "FailureRate": 0.008744977546679272, "AvgDurationMs": 5007.635436408977, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] scopes TestScopedTokens should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:oauth.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 37, "TotalFlakeCount": 0, "FailureRate": 0.009064184223419891, "AvgDurationMs": 5421.61607843137, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestInvalidRoleRefs should succeed [apigroup:authorization.openshift.io][apigroup:user.openshift.io][apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 2, "FailureRate": 0.009061964241978937, "AvgDurationMs": 18782.98495798319, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestProjectIsNamespace should succeed [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 1, "FailureRate": 0.009061964241978937, "AvgDurationMs": 3280.573473389356, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI] TestProjectWatchWithSelectionPredicate should succeed [apigroup:project.openshift.io][apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 179, "FailureRate": 0.009061964241978937, "AvgDurationMs": 18062.220224089648, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc probe can ensure the probe command is functioning as expected on pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 37, "TotalFlakeCount": 4, "FailureRate": 0.008744977546679272, "AvgDurationMs": 9601.85730673317, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] policy scc-subject-review, scc-review [apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 2, "FailureRate": 0.009061964241978937, "AvgDurationMs": 3673.825574229691, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle] Pods cannot access the /config/master API endpoint [Suite:openshift/conformance/parallel]", "TotalTestCount": 3940, "TotalFailureCount": 37, "TotalFlakeCount": 2, "FailureRate": 0.009390862944162437, "AvgDurationMs": 8903.807667560322, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling with Deployments [apigroup:route.openshift.io] should work with TCP (when fully idled) [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 37, "TotalFlakeCount": 18, "FailureRate": 0.009064184223419891, "AvgDurationMs": 19251.749383753508, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router converges when multiple routers are writing conflicting status [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 116, "FailureRate": 0.009061964241978937, "AvgDurationMs": 66505.01719887955, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] should override timeoutGracePeriodSeconds when annotation is set [Suite:openshift/conformance/parallel]", "TotalTestCount": 4185, "TotalFailureCount": 37, "TotalFlakeCount": 53, "FailureRate": 0.008841099163679809, "AvgDurationMs": 14911.846633416457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] supplemental groups Ensure supplemental groups propagate to docker should propagate requested groups to the container [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 37, "TotalFlakeCount": 0, "FailureRate": 0.009061964241978937, "AvgDurationMs": 9151.410840336135, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ClusterResourceQuota] Cluster resource quota should control resource limits across namespaces [apigroup:quota.openshift.io][apigroup:image.openshift.io][apigroup:monitoring.coreos.com][apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 36, "TotalFlakeCount": 10, "FailureRate": 0.00881704628949302, "AvgDurationMs": 27946.980700280114, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Token Expiration] Using a OAuth client with a non-default token max age [apigroup:oauth.openshift.io] to generate tokens that do not expire works as expected when using a code authorization flow [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 36, "TotalFlakeCount": 30, "FailureRate": 0.008917513004706465, "AvgDurationMs": 27479.07655462185, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:PodSecurity][Feature:SCC] SCC admission fails for incorrect/non-existent required-scc annotation [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 36, "TotalFlakeCount": 1, "FailureRate": 0.008508626802174428, "AvgDurationMs": 3066.842817955113, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm build-chain [apigroup:build.openshift.io][apigroup:image.openshift.io][apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 36, "TotalFlakeCount": 3, "FailureRate": 0.008819206271435572, "AvgDurationMs": 6016.100616246498, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can output expected --dry-run text [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 36, "TotalFlakeCount": 4, "FailureRate": 0.008508626802174428, "AvgDurationMs": 3422.2590274314216, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle [apigroup:apps.openshift.io][apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by all [Suite:openshift/conformance/parallel]", "TotalTestCount": 3971, "TotalFailureCount": 36, "TotalFlakeCount": 6, "FailureRate": 0.009065726517250064, "AvgDurationMs": 13784.453652173912, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance impersonation tests [apigroup:user.openshift.io][apigroup:authorization.openshift.io] should pass impersonation deletion tests [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 36, "TotalFlakeCount": 0, "FailureRate": 0.00881704628949302, "AvgDurationMs": 15889.342913165267, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLookup] Image policy should perform lookup when the Deployment gets the resolve-names annotation later [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 36, "TotalFlakeCount": 4, "FailureRate": 0.00881704628949302, "AvgDurationMs": 8622.698123249302, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] APIServer CR fields validation additionalCORSAllowedOrigins [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 1, "FailureRate": 0.008572128337007103, "AvgDurationMs": 2874.954005602242, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs keep the deployer pod invariant valid should deal with config change in case the deployment is still running [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 35, "TotalFlakeCount": 1, "FailureRate": 0.008712969878018421, "AvgDurationMs": 16973.896000000004, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs won't deploy RC with unresolved images when patched with empty image [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 35, "TotalFlakeCount": 8, "FailureRate": 0.009152719665271966, "AvgDurationMs": 19707.36538235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestDeployScale [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 35, "TotalFlakeCount": 50, "FailureRate": 0.00857422831945125, "AvgDurationMs": 3340.099915966387, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestDeploymentConfigDefaults [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 35, "TotalFlakeCount": 0, "FailureRate": 0.00857422831945125, "AvgDurationMs": 2933.4221568627454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the token request URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 35, "TotalFlakeCount": 3, "FailureRate": 0.008669804310131285, "AvgDurationMs": 25956.927675070023, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] RBAC proxy for openshift authz RunLegacyLocalRoleEndpoint should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 1, "FailureRate": 0.008572128337007103, "AvgDurationMs": 3340.8123529411764, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about build.openshift.io api-resources [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 35, "TotalFlakeCount": 2, "FailureRate": 0.00857422831945125, "AvgDurationMs": 3212.1511204481803, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc can route traffic to services [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 3, "FailureRate": 0.008572128337007103, "AvgDurationMs": 17174.326414565825, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for podsecuritypolicysubjectreviews of security.openshift.io, if the resource is present [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 3, "FailureRate": 0.008572128337007103, "AvgDurationMs": 3329.1791316526615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance creation with invalid object reports error should report a failure on creation [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 1, "FailureRate": 0.008572128337007103, "AvgDurationMs": 4100.703837535014, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance object kinds test should create and delete objects from varying API groups [apigroup:template.openshift.io][apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 3, "FailureRate": 0.008572128337007103, "AvgDurationMs": 5654.756134453782, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:Image] signature TestImageRemoveSignature [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 35, "TotalFlakeCount": 1, "FailureRate": 0.008572128337007103, "AvgDurationMs": 5259.746610644257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should start and expose a secured proxy and unsecured metrics [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3813, "TotalFailureCount": 35, "TotalFlakeCount": 56, "FailureRate": 0.00917912404930501, "AvgDurationMs": 15221.05611764706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Conformance][Area:Networking][Feature:Router][apigroup:route.openshift.io][apigroup:config.openshift.io] The HAProxy router should pass the http2 tests [apigroup:image.openshift.io][apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4016, "TotalFailureCount": 35, "TotalFlakeCount": 27, "FailureRate": 0.008715139442231075, "AvgDurationMs": 59722.80472463768, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service with InternalTrafficPolicy set to Local", "TotalTestCount": 3799, "TotalFailureCount": 35, "TotalFlakeCount": 0, "FailureRate": 0.009212950776520137, "AvgDurationMs": 22162.273449275355, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressFirewall] when using openshift ovn-kubernetes should ensure egressfirewall is created [Suite:openshift/conformance/parallel]", "TotalTestCount": 4229, "TotalFailureCount": 35, "TotalFlakeCount": 516, "FailureRate": 0.00827618822416647, "AvgDurationMs": 24076.546857855355, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume should fail when image does not exist [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 35, "TotalFlakeCount": 0, "FailureRate": 0.008272276057669581, "AvgDurationMs": 8440.796209476308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for route.openshift.io/v1, Resource=routes [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 1, "FailureRate": 0.008327210384521186, "AvgDurationMs": 3167.492016806723, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for user.openshift.io/v1, Resource=groups [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008327210384521186, "AvgDurationMs": 3231.509551820728, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apimachinery] server-side-apply should function properly should clear fields when they are no longer being applied in built-in APIs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 34, "TotalFlakeCount": 1, "FailureRate": 0.008035925313164737, "AvgDurationMs": 3428.535411471321, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs ignores deployer and lets the config with a NewReplicationControllerCreated reason should let the deployment config with a NewReplicationControllerCreated reason [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 34, "TotalFlakeCount": 3, "FailureRate": 0.00846402788150361, "AvgDurationMs": 5703.972782608696, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] scopes TestScopeEscalations should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:build.openshift.io][apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008329250367466928, "AvgDurationMs": 5335.401904761906, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] scopes TestTokensWithIllegalScopes should succeed [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 2, "FailureRate": 0.008327210384521186, "AvgDurationMs": 3803.4674789915975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a rolebinding that also contains system:non-existing users should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 3, "FailureRate": 0.008327210384521186, "AvgDurationMs": 3099.2698599439773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about image.openshift.io api-resources [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 1, "FailureRate": 0.008327210384521186, "AvgDurationMs": 3059.3994677871146, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can get version information from API [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008035925313164737, "AvgDurationMs": 3071.861620947631, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug dissect deployment config debug [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4044, "TotalFailureCount": 34, "TotalFlakeCount": 3, "FailureRate": 0.008407517309594461, "AvgDurationMs": 54196.61609550561, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry] Image --dry-run should not update resources [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 3, "FailureRate": 0.008327210384521186, "AvgDurationMs": 4087.528907563024, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLookup] Image policy should update standard Kube object image fields when local names are on [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 5, "FailureRate": 0.008327210384521186, "AvgDurationMs": 8926.474145658265, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagSTIWithConfigChange [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 34, "TotalFlakeCount": 7, "FailureRate": 0.008329250367466928, "AvgDurationMs": 4400.912408963584, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:Image] signature TestImageAddSignature [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 34, "TotalFlakeCount": 5, "FailureRate": 0.008327210384521186, "AvgDurationMs": 5012.350588235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal] Baremetal platform should have baremetalhost resources [Suite:openshift/conformance/parallel]", "TotalTestCount": 4183, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008128137700215157, "AvgDurationMs": 2951.50842892768, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow connections to pods from infra cluster pod via LoadBalancer service across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008035925313164737, "AvgDurationMs": 3042.3943890274318, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Conformance][Area:Networking][Feature:Router] The HAProxy router should be able to connect to a service that is idled because a GET on the route will unidle it [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 3702, "TotalFailureCount": 34, "TotalFlakeCount": 2, "FailureRate": 0.009184224743381956, "AvgDurationMs": 35201.69049689442, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] pod-resize-resource-quota-test", "TotalTestCount": 4199, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008097165991902834, "AvgDurationMs": 11521.445564102562, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume should handle multiple image volumes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 34, "TotalFlakeCount": 0, "FailureRate": 0.008035925313164737, "AvgDurationMs": 12132.568304239403, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ResourceQuota] Object count check the quota after import-image with --all option [Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4006, "TotalFailureCount": 33, "TotalFlakeCount": 33, "FailureRate": 0.008237643534697954, "AvgDurationMs": 9741.808363636366, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for oauth.openshift.io/v1, Resource=oauthaccesstokens [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.008082292432035268, "AvgDurationMs": 3420.9633053221287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for security.openshift.io/v1, Resource=rangeallocations [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.008082292432035268, "AvgDurationMs": 3229.9173389355733, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for template.openshift.io/v1, Resource=templates [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.008082292432035268, "AvgDurationMs": 3630.98974789916, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs keep the deployer pod invariant valid should deal with cancellation of running deployment [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.008215085884988798, "AvgDurationMs": 17525.06162318841, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl whitelists net.ipv4.tcp_syncookies [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 0, "FailureRate": 0.007799574568659891, "AvgDurationMs": 19368.643840399003, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the authorize URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.008174386920980926, "AvgDurationMs": 27463.90140056022, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SecurityContextConstraints] TestPodDefaultCapabilities [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 3, "FailureRate": 0.007799574568659891, "AvgDurationMs": 16719.378977556113, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm release extract image-references [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 5, "FailureRate": 0.007799574568659891, "AvgDurationMs": 4859.0914214463855, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc annotate pod [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 0, "FailureRate": 0.007799574568659891, "AvgDurationMs": 4846.021820448877, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about operator.openshift.io api-resources [apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 2, "FailureRate": 0.008082292432035268, "AvgDurationMs": 3076.8296638655474, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for template.openshift.io [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 0, "FailureRate": 0.008082292432035268, "AvgDurationMs": 3674.743081232493, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle] TestAdminAck should succeed [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 33, "TotalFlakeCount": 2, "FailureRate": 0.008082292432035268, "AvgDurationMs": 2797.665910364145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageInfo] Image info should display information about images [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3758, "TotalFailureCount": 33, "TotalFlakeCount": 2, "FailureRate": 0.008781266631186802, "AvgDurationMs": 13753.354911764705, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagSTI [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 33, "TotalFlakeCount": 3, "FailureRate": 0.008084272415482607, "AvgDurationMs": 4378.847535014006, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 0, "FailureRate": 0.007799574568659891, "AvgDurationMs": 209377.80092269328, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeature:StorageNetworkPolicy] Storage Network Policy should verify required labels for CSO related Operators [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 33, "TotalFlakeCount": 1, "FailureRate": 0.007799574568659891, "AvgDurationMs": 3328.884713216957, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for template.openshift.io/v1, Resource=brokertemplateinstances [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 3, "FailureRate": 0.007837374479549351, "AvgDurationMs": 3236.2798879551824, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ControlPlaneSecurity] should have privileged securityContext for control plane init and main containers [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 32, "TotalFlakeCount": 0, "FailureRate": 0.007563223824155046, "AvgDurationMs": 3390.1350374064837, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the logout URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 32, "TotalFlakeCount": 0, "FailureRate": 0.007926678226405746, "AvgDurationMs": 26474.66658263306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [apigroup:oauth.openshift.io] OAuthClientWithRedirectURIs must validate request URIs according to oauth-client definition [Suite:openshift/conformance/parallel]", "TotalTestCount": 4009, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.007982040409079572, "AvgDurationMs": 3962.1406303724934, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] authorization TestBrowserSafeAuthorizer should succeed [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 8, "FailureRate": 0.007837374479549351, "AvgDurationMs": 3634.802268907563, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] scopes TestScopedImpersonation should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 32, "TotalFlakeCount": 3, "FailureRate": 0.007839294463498285, "AvgDurationMs": 4579.421456582632, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] self-SAR compatibility TestBootstrapPolicySelfSubjectAccessReviews should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 1, "FailureRate": 0.007837374479549351, "AvgDurationMs": 3991.6375070028007, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:PodSecurity][Feature:SCC] creating pod controllers [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 32, "TotalFlakeCount": 5, "FailureRate": 0.007563223824155046, "AvgDurationMs": 3758.4732917705737, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig should prune completed builds based on the successfulBuildsHistoryLimit setting [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.008368200836820083, "AvgDurationMs": 221948.52244117644, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm new-project [apigroup:project.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4084, "TotalFailureCount": 32, "TotalFlakeCount": 9, "FailureRate": 0.007835455435847209, "AvgDurationMs": 122849.48252100842, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm who-can [apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 1, "FailureRate": 0.007837374479549351, "AvgDurationMs": 5677.112913165267, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can create and interact with a list of resources [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 32, "TotalFlakeCount": 4, "FailureRate": 0.007563223824155046, "AvgDurationMs": 6360.469551122193, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can patch resources [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.007837374479549351, "AvgDurationMs": 4891.431120448178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc builds get buildconfig [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 32, "TotalFlakeCount": 7, "FailureRate": 0.008228336333247621, "AvgDurationMs": 4793.52494318182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for template.openshift.io [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.007837374479549351, "AvgDurationMs": 3173.7915126050416, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagDockerWithConfigChange [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.007839294463498285, "AvgDurationMs": 4096.189299719888, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should have a AlertmanagerReceiversNotConfigured alert in firing state [Suite:openshift/conformance/parallel]", "TotalTestCount": 3797, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.00842770608375033, "AvgDurationMs": 3669.7718072289167, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should have important platform topology metrics [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 32, "TotalFlakeCount": 9, "FailureRate": 0.008366013071895426, "AvgDurationMs": 4118.204441176471, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][OCPFeatureGate:GatewayAPIController][Feature:Router][apigroup:gateway.networking.k8s.io] Ensure default gatewayclass is accepted [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 32, "TotalFlakeCount": 3, "FailureRate": 0.007837374479549351, "AvgDurationMs": 22115.292240896353, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume should succeed if image volume is not existing but unused [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 32, "TotalFlakeCount": 0, "FailureRate": 0.007563223824155046, "AvgDurationMs": 10287.641097256857, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeature:StorageNetworkPolicy] Storage Network Policy should verify required labels for CSI related Operators [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 32, "TotalFlakeCount": 2, "FailureRate": 0.007563223824155046, "AvgDurationMs": 3169.8129426433898, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions kube-controller-manager", "TotalTestCount": 6807, "TotalFailureCount": 32, "TotalFlakeCount": 0, "FailureRate": 0.0047010430439253705, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Serial][sig-cli] oc adm upgrade recommend When the update service has conditional recommendations runs successfully when listing all updates [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.04177897574123989, "AvgDurationMs": 20889.547260273976, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for build.openshift.io/v1, Resource=buildconfigs [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 31, "TotalFlakeCount": 3, "FailureRate": 0.007594316511513963, "AvgDurationMs": 3192.306302521007, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for oauth.openshift.io/v1, Resource=oauthclients [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 2, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3202.2333893557425, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for user.openshift.io/v1, Resource=identities [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3179.3652941176465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for user.openshift.io/v1, Resource=users [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3231.2912885154055, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] ClusterOperators [apigroup:config.openshift.io] should define at least one related object that is not a namespace [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 2, "FailureRate": 0.007592456527063434, "AvgDurationMs": 2979.7633893557413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:Authentication] TestFrontProxy should succeed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007326873079650201, "AvgDurationMs": 3355.7721695760606, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] [Headers][apigroup:route.openshift.io][apigroup:config.openshift.io][apigroup:oauth.openshift.io] expected headers returned from the root URL [Suite:openshift/conformance/parallel]", "TotalTestCount": 4037, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007678969531830567, "AvgDurationMs": 26860.217198879553, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] well-known endpoint should be reachable [apigroup:route.openshift.io] [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4009, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007732601646295834, "AvgDurationMs": 3141.2518338108885, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] authorization TestAuthorizationSubjectAccessReviewAPIGroup should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 3, "FailureRate": 0.007592456527063434, "AvgDurationMs": 7286.732885154061, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a rolebinding when subject is permitted by RBR should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3082.0567787114846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SecurityContextConstraints] TestPodUpdateSCCEnforcement [apigroup:user.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007592456527063434, "AvgDurationMs": 5525.5643417366955, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:UserAPI] groups should work [apigroup:user.openshift.io][apigroup:project.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007592456527063434, "AvgDurationMs": 6575.01974789916, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc --request-timeout works as expected for deployment [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 15, "FailureRate": 0.007326873079650201, "AvgDurationMs": 6189.891047381546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm user-creation [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 6237.528963585432, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about api-resources [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007326873079650201, "AvgDurationMs": 4124.195810473815, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for authorization.openshift.io [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 5045.813529411765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for console.openshift.io [apigroup:console.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007594316511513963, "AvgDurationMs": 4639.838655462184, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for route.openshift.io [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3087.4063305322134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper spec+status for CRDs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007326873079650201, "AvgDurationMs": 25602.565461346636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc expose can ensure the expose command is functioning as expected [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007592456527063434, "AvgDurationMs": 7722.573053221289, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc idle Deployments [apigroup:route.openshift.io][apigroup:project.openshift.io][apigroup:image.openshift.io] by all [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007715281234444997, "AvgDurationMs": 12466.37353623189, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc statefulset creates and deletes statefulsets [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007326873079650201, "AvgDurationMs": 8704.522219451372, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] template-api TestTemplateTransformationFromConfig [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 3, "FailureRate": 0.007592456527063434, "AvgDurationMs": 3012.8157983193273, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:Image] oc tag should preserve image reference for external images [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 4587.211568627452, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge][Conformance][Area:Networking][Feature:Router][apigroup:route.openshift.io] The HAProxy router should pass the h2spec conformance tests [apigroup:authorization.openshift.io][apigroup:user.openshift.io][apigroup:security.openshift.io][apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4018, "TotalFailureCount": 31, "TotalFlakeCount": 9, "FailureRate": 0.007715281234444997, "AvgDurationMs": 52889.72214492754, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.007326873079650201, "AvgDurationMs": 219212.4293765586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster workloads with limits should have resources modified if CPUPartitioningMode = AllNodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007592456527063434, "AvgDurationMs": 8733.05305322129, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should Implement packages API server and list packagemanifest info with namespace not NULL [apigroup:packages.operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 31, "TotalFlakeCount": 1, "FailureRate": 0.007594316511513963, "AvgDurationMs": 325.0647899159663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions cluster-autoscaler", "TotalTestCount": 6807, "TotalFailureCount": 31, "TotalFlakeCount": 0, "FailureRate": 0.004554135448802703, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ResourceQuota] Object count when exceed openshift.io/image-tags will ban to create new image references in the project [Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4006, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.0074887668497254116, "AvgDurationMs": 6074.061246753246, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for build.openshift.io/v1, Resource=builds [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.007349338559529643, "AvgDurationMs": 3201.670336134454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for image.openshift.io/v1, Resource=images [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 3, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 3256.1982352941172, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] poddisruptionbudgets with unhealthyPodEvictionPolicy should evict according to the AlwaysAllow policy [Suite:openshift/conformance/parallel]", "TotalTestCount": 4038, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.007429420505200594, "AvgDurationMs": 338482.5627272726, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with env in params referencing the configmap should expand the config map key to a value [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 30, "TotalFlakeCount": 10, "FailureRate": 0.007845188284518828, "AvgDurationMs": 20960.60102941177, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestTriggers_imageChange [apigroup:apps.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.007349338559529643, "AvgDurationMs": 3269.5420168067217, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Late][Jira:\"kube-apiserver\"] all registered tls artifacts must have no metadata violation regressions [Suite:openshift/conformance/parallel]", "TotalTestCount": 5233, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.005732849226065355, "AvgDurationMs": 21343.148869778866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Late][Jira:\"kube-apiserver\"] all tls artifacts must be registered [Suite:openshift/conformance/parallel]", "TotalTestCount": 5233, "TotalFailureCount": 30, "TotalFlakeCount": 5, "FailureRate": 0.005732849226065355, "AvgDurationMs": 21114.83739557739, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Late][Jira:\"kube-apiserver\"] collect certificate data [Suite:openshift/conformance/parallel]", "TotalTestCount": 5233, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.005732849226065355, "AvgDurationMs": 20921.992039312045, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a RBAC rolebinding when subject is not already bound and is not permitted by any RBR should fail [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 3231.699915966385, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm must-gather runs successfully [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.007369196757553427, "AvgDurationMs": 153588.6875070028, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm ui-project-commands [apigroup:project.openshift.io][apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 77571.5256582633, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can create deploymentconfig and clusterquota [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.007349338559529643, "AvgDurationMs": 9917.87428571429, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for project.openshift.io [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 3154.032577030812, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for build.openshift.io [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.007349338559529643, "AvgDurationMs": 3431.422408963585, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for image.openshift.io [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 3719.4274509803918, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for podsecuritypolicyreviews of security.openshift.io, if the resource is present [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 3452.234425770308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] template-api TestTemplate [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.0073475385745775165, "AvgDurationMs": 2776.8495798319327, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagCustom [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 1, "FailureRate": 0.007349338559529643, "AvgDurationMs": 6008.500028011204, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagCustomWithConfigChange [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 30, "TotalFlakeCount": 2, "FailureRate": 0.007349338559529643, "AvgDurationMs": 5972.464901960784, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal] Baremetal platform should have hostfirmwaresetting resources [Suite:openshift/conformance/parallel]", "TotalTestCount": 4184, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.007170172084130019, "AvgDurationMs": 3425.1693765586037, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking IPerf2 [Feature:Networking-Performance] should run iperf2", "TotalTestCount": 4230, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.0070921985815602835, "AvgDurationMs": 50427.03628428927, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery should allow recovery if controller expansion fails with final error", "TotalTestCount": 4231, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.007090522335145356, "AvgDurationMs": 86817.06723192018, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions console", "TotalTestCount": 6807, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.004407227853680035, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions machine-api", "TotalTestCount": 6807, "TotalFailureCount": 30, "TotalFlakeCount": 0, "FailureRate": 0.004407227853680035, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer] TestTLSDefaults [Suite:openshift/conformance/parallel]", "TotalTestCount": 4228, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.006859035004730369, "AvgDurationMs": 2975.726907730674, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer] should serve openapi v3 [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.00690640628721124, "AvgDurationMs": 846.5095384615386, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs keep the deployer pod invariant valid should deal with cancellation after deployer pod succeeded [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007583682008368201, "AvgDurationMs": 22323.66829411765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] ClusterOperators [apigroup:config.openshift.io] should define at least one namespace in their lists of related objects [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 2958.230420168067, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] ocp payload should be based on existing source OLM version should contain the source commit id [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 29, "TotalFlakeCount": 3, "FailureRate": 0.00690640628721124, "AvgDurationMs": 3677.6767179487174, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] authorization TestClusterReaderCoverage should succeed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4230, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.006855791962174941, "AvgDurationMs": 3031.979650872818, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm groups [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 7628.080924369748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm role-reapers [apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 2, "FailureRate": 0.007102620622091599, "AvgDurationMs": 11014.06280112045, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm role-selectors [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 6893.521456582632, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc builds complex build start-build [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 29, "TotalFlakeCount": 3, "FailureRate": 0.007456929802005657, "AvgDurationMs": 8454.956448863637, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug ensure debug does not depend on a container actually existing for the selected resource for deployment [Suite:openshift/conformance/parallel]", "TotalTestCount": 4166, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.006961113778204513, "AvgDurationMs": 3562.0196658097684, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for security.internal.openshift.io [apigroup:security.internal.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3092.5808683473388, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for securitycontextconstraints of security.openshift.io, if the resource is present [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 2, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3153.6247619047626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for user.openshift.io [apigroup:user.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3823.2312324929985, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for podsecuritypolicyselfsubjectreviews of security.openshift.io, if the resource is present [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3098.1266666666675, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for project.openshift.io [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3216.835154061625, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for route.openshift.io [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3178.043193277311, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal] Baremetal platform should not allow updating BootMacAddress [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.0068541715906405105, "AvgDurationMs": 3003.137057356608, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow connections to pods from infra cluster pod via NodePort across different infra nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.0068541715906405105, "AvgDurationMs": 2895.3472817955108, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster workloads in annotated namespaces should be modified if CPUPartitioningMode = AllNodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 0, "FailureRate": 0.007102620622091599, "AvgDurationMs": 6428.669915966385, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster workloads in non-annotated namespaces should be allowed if CPUPartitioningMode = AllNodes with a warning annotation [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 6766.198319327731, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] an end user can use OLM Report Upgradeable in OLM ClusterOperators status [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 29, "TotalFlakeCount": 1, "FailureRate": 0.007102620622091599, "AvgDurationMs": 3554.9214285714284, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for template.openshift.io/v1, Resource=templateinstances [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006857702669605682, "AvgDurationMs": 3160.8101120448177, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apimachinery] server-side-apply should function properly should clear fields when they are no longer being applied in FeatureGates [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006857702669605682, "AvgDurationMs": 3115.1034733893557, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] [Conformance] sysctl pod should not start for sysctl not on whitelist kernel.msgmax [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 3, "FailureRate": 0.006617820846135665, "AvgDurationMs": 5175.682618453865, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] RBAC proxy for openshift authz RunLegacyClusterRoleBindingEndpoint should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006857702669605682, "AvgDurationMs": 3452.87700280112, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm images [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006857702669605682, "AvgDurationMs": 6263.29761904762, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm must-gather runs successfully with options [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006857702669605682, "AvgDurationMs": 13442.971792717088, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm serviceaccounts [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 3, "FailureRate": 0.006617820846135665, "AvgDurationMs": 6702.778054862843, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can get version information from CLI [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006617820846135665, "AvgDurationMs": 3218.4857356608472, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc can get list of nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006617820846135665, "AvgDurationMs": 3232.644987531173, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug dissect deployment debug [Suite:openshift/conformance/parallel]", "TotalTestCount": 4166, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.00672107537205953, "AvgDurationMs": 51046.95218508999, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug ensure debug does not depend on a container actually existing for the selected resource [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.0069703759024147376, "AvgDurationMs": 4528.167826086957, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for apps.openshift.io [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006859382655560999, "AvgDurationMs": 3180.345350140057, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for image.openshift.io [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006857702669605682, "AvgDurationMs": 4122.966078431375, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for network.operator.openshift.io [apigroup:network.operator.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 2, "FailureRate": 0.006857702669605682, "AvgDurationMs": 3116.6792436974797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for oauth.openshift.io [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006857702669605682, "AvgDurationMs": 4182.769383753502, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain spec+status for apps.openshift.io [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006859382655560999, "AvgDurationMs": 3116.0768347338935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc label pod [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 2, "FailureRate": 0.006617820846135665, "AvgDurationMs": 4807.701047381546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc project --show-labels works for projects [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006857702669605682, "AvgDurationMs": 3256.0605882352934, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc secret creates and retrieves expected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 1, "FailureRate": 0.006617820846135665, "AvgDurationMs": 4261.508304239401, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cloud-provider][Feature:OpenShiftCloudControllerManager][Late] Cluster scoped load balancer healthcheck port and path should be 10256/healthz [Suite:openshift/conformance/parallel]", "TotalTestCount": 5244, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.005339435545385202, "AvgDurationMs": 5127.666167076166, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:Templates] templateinstance cross-namespace test should create and delete objects across namespaces [apigroup:user.openshift.io][apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 4, "FailureRate": 0.006857702669605682, "AvgDurationMs": 7639.437899159663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLookup] Image policy should perform lookup when the object has the resolve-names annotation [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 28, "TotalFlakeCount": 4, "FailureRate": 0.006857702669605682, "AvgDurationMs": 9549.749383753502, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal] Baremetal platform should have preprovisioning images for workers [Suite:openshift/conformance/parallel]", "TotalTestCount": 4183, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006693760459000717, "AvgDurationMs": 3129.249551122194, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 operator installation should install an openshift catalog cluster extension", "TotalTestCount": 3611, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.007754084741068956, "AvgDurationMs": 11488.911124260354, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006617820846135665, "AvgDurationMs": 33536.654563591015, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 28, "TotalFlakeCount": 0, "FailureRate": 0.006617820846135665, "AvgDurationMs": 32270.051995012473, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"lb-ext.kubeconfig\" should be present on all masters and work [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4159, "TotalFailureCount": 27, "TotalFlakeCount": 2, "FailureRate": 0.006491945179129598, "AvgDurationMs": 14581.141937172772, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"lb-int.kubeconfig\" should be present on all masters and work [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4159, "TotalFailureCount": 27, "TotalFlakeCount": 5, "FailureRate": 0.006491945179129598, "AvgDurationMs": 15031.907146596855, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-api-machinery][Feature:APIServer] local kubeconfig \"localhost.kubeconfig\" should be present on all masters and work [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4160, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006490384615384616, "AvgDurationMs": 14712.713638743458, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer][Late] API LBs follow /readyz of kube-apiserver and don't send request early [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005147759771210677, "AvgDurationMs": 111.29189189189192, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer][Late] API LBs follow /readyz of kube-apiserver and stop sending requests [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005147759771210677, "AvgDurationMs": 111.14346437346438, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer][Late] kube-apiserver terminates within graceful termination period [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005147759771210677, "AvgDurationMs": 110.27626535626537, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer][Late] kubelet terminates kube-apiserver gracefully [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005147759771210677, "AvgDurationMs": 111.28334152334152, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for image.openshift.io/v1, Resource=imagestreams [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006612784717119765, "AvgDurationMs": 3221.4591596638656, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][apigroup:apps.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] Deployments on HighlyAvailableArbiterMode topology should be created on master nodes when no node selected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4094, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006595017098192477, "AvgDurationMs": 3339.883305322129, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] ClusterOperators [apigroup:config.openshift.io] should define valid related objects [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 27, "TotalFlakeCount": 3, "FailureRate": 0.006612784717119765, "AvgDurationMs": 2971.6487955182074, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Late] clients should not use APIs that are removed in upcoming releases [apigroup:apiserver.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5057, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005339133873838244, "AvgDurationMs": 175.73660167130922, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:UserAPI] users can manipulate groups [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4055, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006658446362515413, "AvgDurationMs": 6318.070916905444, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] custom build with buildah being created from new-build should complete build with custom builder image [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 27, "TotalFlakeCount": 2, "FailureRate": 0.007060669456066946, "AvgDurationMs": 150989.94073529416, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][volumes] build volumes should mount given secrets and configmaps into the build pod for source strategy builds [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 27, "TotalFlakeCount": 4, "FailureRate": 0.006721433905899925, "AvgDurationMs": 76769.73776811597, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl exec should be able to execute 1000 times in a container", "TotalTestCount": 4231, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.00638147010163082, "AvgDurationMs": 41159.277132169576, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc can run inside of a busybox container [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 27, "TotalFlakeCount": 5, "FailureRate": 0.006719761075161772, "AvgDurationMs": 12298.298115942029, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug does not require a real resource on the server [Suite:openshift/conformance/parallel]", "TotalTestCount": 4165, "TotalFailureCount": 27, "TotalFlakeCount": 2, "FailureRate": 0.006482593037214886, "AvgDurationMs": 3319.01456185567, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc env can set environment variables for deployment [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 27, "TotalFlakeCount": 1, "FailureRate": 0.006614404703576678, "AvgDurationMs": 8472.898879551822, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for config.openshift.io [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 27, "TotalFlakeCount": 1, "FailureRate": 0.006612784717119765, "AvgDurationMs": 3171.7400840336136, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cloud-provider][Feature:OpenShiftCloudControllerManager][Late] Deploy an external cloud provider [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5057, "TotalFailureCount": 27, "TotalFlakeCount": 1, "FailureRate": 0.005339133873838244, "AvgDurationMs": 2461.1546239554323, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry] Image --dry-run should not delete resources [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006612784717119765, "AvgDurationMs": 4260.561204481795, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestMultipleImageChangeBuildTriggers [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006614404703576678, "AvgDurationMs": 5446.678879551819, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow connections to pods from guest cluster PodNetwork pod via LoadBalancer service across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.00638147010163082, "AvgDurationMs": 3000.0657605985043, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow connections to pods from guest hostNetwork pod via NodePort across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.00638147010163082, "AvgDurationMs": 2957.4330673316713, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] Should properly block MCN updates by impersonation of the MCD SA [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006612784717119765, "AvgDurationMs": 482.33218487394953, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][Late] should not have pod creation failures due to systemd timeouts [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.005147759771210677, "AvgDurationMs": 491.42601965601966, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] expected Master and Arbiter node counts Should validate that there are Master and Arbiter nodes as specified in the cluster [Suite:openshift/conformance/parallel]", "TotalTestCount": 4097, "TotalFailureCount": 27, "TotalFlakeCount": 0, "FailureRate": 0.006590187942396876, "AvgDurationMs": 99.43249299719888, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API health endpoints should contain the required checks for the openshift-apiserver APIs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006145119357125975, "AvgDurationMs": 1662.8801496259352, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for apps.openshift.io/v1, Resource=deploymentconfigs [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006472491909385114, "AvgDurationMs": 3321.397420289855, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestTriggers_manual [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006369426751592357, "AvgDurationMs": 3140.020280112044, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Cluster topology single node tests Verify that OpenShift components deploy one replica in SingleReplica topology mode [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006145119357125975, "AvgDurationMs": 2609.1844139650875, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OAuthServer] OAuth Authenticator accepts sha256 access tokens [apigroup:user.openshift.io][apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006367866764633847, "AvgDurationMs": 3430.404705882354, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] The default cluster RBAC policy should have correct RBAC rules [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 26, "TotalFlakeCount": 9, "FailureRate": 0.006191950464396285, "AvgDurationMs": 3624.4834358974354, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SecurityContextConstraints] TestAllowedSCCViaRBAC with service account [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 26, "TotalFlakeCount": 1, "FailureRate": 0.006367866764633847, "AvgDurationMs": 10136.389383753505, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] Multi-stage image builds should succeed [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006799163179916318, "AvgDurationMs": 96060.75426470589, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][pullsecret] docker build using a pull secret Building from a template should create a docker build that pulls using a secret run it [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006685523270763692, "AvgDurationMs": 126560.0059943182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][volumes] build volumes should mount given secrets and configmaps into the build pod for docker strategy builds [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 26, "TotalFlakeCount": 7, "FailureRate": 0.006472491909385114, "AvgDurationMs": 71691.59000000001, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about route.openshift.io api-resources and api-version [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006367866764633847, "AvgDurationMs": 3165.6389355742313, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc run can use --image flag correctly for deployment [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006145119357125975, "AvgDurationMs": 3316.688354114713, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:OpenShiftControllerManager] TestAutomaticCreationOfPullSecrets [apigroup:config.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006367866764633847, "AvgDurationMs": 4042.196526610645, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-etcd] etcd leader changes are not excessive [Late] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5135, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.005063291139240506, "AvgDurationMs": 468.84810606060614, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLayers] Image layer subresource should identify a deleted image as missing [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006470881035340966, "AvgDurationMs": 3350.9745507246384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] Alerts shouldn't exceed the series limit of total series sent via telemetry from each cluster [Suite:openshift/conformance/parallel]", "TotalTestCount": 5199, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.005000961723408348, "AvgDurationMs": 331.05884520884524, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] OpenShift alerting rules [apigroup:image.openshift.io] should link to an HTTP(S) location if the runbook_url annotation is defined [Suite:openshift/conformance/parallel]", "TotalTestCount": 4920, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.005284552845528455, "AvgDurationMs": 409.867665706052, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow direct connections to pods from guest cluster pod in host network across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4230, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006146572104018913, "AvgDurationMs": 2935.152119700748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network-edge] DNS should answer A and AAAA queries for a dual-stack service [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006367866764633847, "AvgDurationMs": 5351.786722689075, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router] The HAProxy router should expose prometheus metrics for a route [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3795, "TotalFailureCount": 26, "TotalFlakeCount": 2, "FailureRate": 0.006851119894598155, "AvgDurationMs": 44032.843554216866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] required pods on the Arbiter node Should verify that the correct number of pods are running on the Arbiter node [Suite:openshift/conformance/parallel]", "TotalTestCount": 4097, "TotalFailureCount": 26, "TotalFlakeCount": 0, "FailureRate": 0.006346106907493288, "AvgDurationMs": 99.9021848739496, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should have imagePullPolicy:IfNotPresent on thier deployments [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 26, "TotalFlakeCount": 1, "FailureRate": 0.006191950464396285, "AvgDurationMs": 810.6376666666666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] an end user can use OLM can subscribe to the operator [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 26, "TotalFlakeCount": 4, "FailureRate": 0.006683804627249357, "AvgDurationMs": 37706.92477272725, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer] should serve openapi v3 discovery [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00590876861262113, "AvgDurationMs": 87.80107231920198, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] TTLAfterFinished job should be deleted once it finishes after TTL seconds", "TotalTestCount": 4185, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.005973715651135006, "AvgDurationMs": 27168.18389027431, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:DeploymentConfig] deploymentconfigs with enhanced status should include various info in status [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 25, "TotalFlakeCount": 2, "FailureRate": 0.00653765690376569, "AvgDurationMs": 17719.458029411766, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should ensure platform components have system-* priority class associated [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.005953798523457966, "AvgDurationMs": 263.15697435897437, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should set requests but not limits [Suite:openshift/conformance/parallel]", "TotalTestCount": 4185, "TotalFailureCount": 25, "TotalFlakeCount": 45, "FailureRate": 0.005973715651135006, "AvgDurationMs": 4609.151645885287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] RBAC proxy for openshift authz RunLegacyLocalRoleBindingEndpoint should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00612294881214793, "AvgDurationMs": 4129.272268907563, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] scopes TestUnknownScopes should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 47, "FailureRate": 0.00612294881214793, "AvgDurationMs": 5365.006022408964, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] self-SAR compatibility TestSelfSubjectAccessReviewsNonExistingNamespace should succeed [apigroup:user.openshift.io][apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00612294881214793, "AvgDurationMs": 3717.76694677871, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a rolebinding when subject is already bound should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00612294881214793, "AvgDurationMs": 3256.0276470588237, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] result image should have proper labels set S2I build from a template should create a image from \"test-s2i-build.json\" template with proper Docker labels [apigroup:build.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.006428387760349704, "AvgDurationMs": 66776.59889204548, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm must-gather runs successfully for audit logs [apigroup:config.openshift.io][apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3997, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.006254691018263698, "AvgDurationMs": 77910.28825214901, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm must-gather when looking at the audit logs [apigroup:config.openshift.io] [sig-node] kubelet runs apiserver processes strictly sequentially in order to not risk audit log corruption [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 1, "FailureRate": 0.00612294881214793, "AvgDurationMs": 63504.84918767506, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can show correct whoami result [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 1, "FailureRate": 0.00590876861262113, "AvgDurationMs": 2984.337082294263, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc observe works as expected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 2, "FailureRate": 0.00590876861262113, "AvgDurationMs": 11775.577481296754, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc run can use --image flag correctly [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.006124448799608035, "AvgDurationMs": 3157.322605042017, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Annotation trigger reconciles after the image is overwritten [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 25, "TotalFlakeCount": 5, "FailureRate": 0.006426735218508998, "AvgDurationMs": 6915.931335227273, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-installer][Feature:baremetal] Baremetal/OpenStack/vSphere/None/AWS/Azure/GCP platforms have a metal3 deployment [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00590876861262113, "AvgDurationMs": 3097.525211970074, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow direct connections to pods from guest cluster pod in pod network across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00590876861262113, "AvgDurationMs": 2947.338653366583, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] Should properly block MCN updates from a MCD that is not the associated one [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 1, "FailureRate": 0.00612294881214793, "AvgDurationMs": 1366.555406162465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - remove memory limits", "TotalTestCount": 4199, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.005953798523457966, "AvgDurationMs": 5390.315666666667, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one container - decrease CPU only", "TotalTestCount": 4199, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.005953798523457966, "AvgDurationMs": 26574.87815384615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster platform workloads should be annotated correctly for DaemonSets [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00612294881214793, "AvgDurationMs": 1818.6111484593846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster platform workloads should be annotated correctly for Deployments [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 25, "TotalFlakeCount": 225, "FailureRate": 0.00612294881214793, "AvgDurationMs": 6369.624005602238, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][Jira:OLM] clusterextension PolarionID:83069-[OTP]olmv1 static networkpolicy.", "TotalTestCount": 3952, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.006325910931174089, "AvgDurationMs": 7160.237312661499, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 25, "TotalFlakeCount": 0, "FailureRate": 0.00590876861262113, "AvgDurationMs": 32785.53755610973, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056179775280898875, "AvgDurationMs": 47127.852354368915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] JSON Patch [apigroup:operator.openshift.io] should delete an entry from an array with a test precondition provided [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 296.95957983193284, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer] authenticated browser should get a 200 from / [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 85.45947630922693, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ServerSideApply] Server-Side Apply should work for oauth.openshift.io/v1, Resource=oauthauthorizetokens [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3318.5594397759096, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apimachinery] server-side-apply should function properly should clear fields when they are no longer being applied on CRDs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 3397.620374064838, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should ensure control plane operators do not make themselves unevictable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 256.75773067331676, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should have operators on the cluster version [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.005878030859662013, "AvgDurationMs": 215.07126050420166, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should only include cluster daemonsets that have maxUnavailable or maxSurge update of 10 percent or maxUnavailable of 33 percent [Suite:openshift/conformance/parallel]", "TotalTestCount": 4153, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005778954972309174, "AvgDurationMs": 202.00343589743593, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization] RBAC proxy for openshift authz RunLegacyClusterRoleEndpoint should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3374.9122128851545, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a rolebinding when subject is not already bound and is not permitted by any RBR should fail [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3095.842464985994, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][timing] capture build stages and durations should record build stages and durations for docker [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0061712522499357166, "AvgDurationMs": 79836.47414772723, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm node-logs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4187, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.005732027704800574, "AvgDurationMs": 7507.130384615384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about project.openshift.io api-resources [apigroup:project.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3122.7490196078425, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc api-resources can output expected information about snapshot.storage.k8s.io api-resources [apigroup:snapshot.storage.k8s.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005879470847623714, "AvgDurationMs": 2980.4466386554636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can describe an OAuth access token [apigroup:oauth.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3716.030168067227, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc explain should contain proper fields description for rangeallocations of security.openshift.io, if the resource is present [apigroup:security.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 3206.244369747899, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc set image can set images for pods and deployments [apigroup:image.openshift.io][Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.006169665809768638, "AvgDurationMs": 12793.703011363634, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] templates different namespaces [apigroup:user.openshift.io][apigroup:project.openshift.io][apigroup:template.openshift.io][apigroup:authorization.openshift.io][Skipped:Disconnected] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3890, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.006169665809768638, "AvgDurationMs": 69414.03551136362, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-devex][Feature:OpenShiftControllerManager] TestDockercfgTokenDeletedController [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 4055.8043977591033, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-etcd][apigroup:config.openshift.io][OCPFeatureGate:HighlyAvailableArbiter] Ensure etcd health and quorum in HighlyAvailableArbiterMode should have all etcd pods running and quorum met [Suite:openshift/conformance/parallel]", "TotalTestCount": 4097, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005857944837686112, "AvgDurationMs": 101.05680672268909, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageLookup] Image policy should update OpenShift object image fields when local names are on [apigroup:image.openshift.io][apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.005879470847623714, "AvgDurationMs": 7657.458515406164, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers] Image change build triggers TestSimpleImageChangeBuildTriggerFromImageStreamTagDocker [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.005879470847623714, "AvgDurationMs": 4140.250952380952, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][sig-builds][Feature:Builds] Prometheus when installed on the cluster should start and expose a secured proxy and verify build metrics [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.006276150627615063, "AvgDurationMs": 73441.50558823529, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] services when running openshift cluster on KubeVirt virtual machines should allow connections to pods from guest podNetwork pod via NodePort across different guest nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 3024.289600997507, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce except clause while egress access to server in CIDR block [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 34128.62950124688, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should ensure an IP overlapping both IPBlock.CIDR and IPBlock.Except is allowed [Feature:NetworkPolicy]", "TotalTestCount": 4230, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005673758865248227, "AvgDurationMs": 53043.752219451344, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should fallback to local terminating endpoints when there are no ready endpoints with externalTrafficPolicy=Local", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 40133.61104738154, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should serve endpoints on same port and different protocols [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 29725.49002493765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressFirewall] egressFirewall should have no impact outside its namespace [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 10, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 12252.84149625935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod should start with all sysctl on whitelist [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.005973120955699353, "AvgDurationMs": 9506.332463768116, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:vlan] should create pingable pods with ipvlan interface on an in-container master [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 15222.69019607843, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller pod connected to ClusterUserDefinedNetwork CR \u0026 managed NADs cannot be deleted when being used [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 3, "FailureRate": 0.005878030859662013, "AvgDurationMs": 16944.865602240898, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork is isolated from the default network with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 2, "FailureRate": 0.005878030859662013, "AvgDurationMs": 105535.85232492995, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to remove the external certificate then also the route is reachable and serves the default certificate [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 2, "FailureRate": 0.005878030859662013, "AvgDurationMs": 18353.612885154063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use new external certificate then also the route is reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 5, "FailureRate": 0.005878030859662013, "AvgDurationMs": 18130.52868347339, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the secret is updated then also routes are reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 32838.179635854336, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Managed cluster should report ready nodes the entire duration of the test run [Late][apigroup:monitoring.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4802, "TotalFailureCount": 24, "TotalFlakeCount": 1, "FailureRate": 0.0049979175343606835, "AvgDurationMs": 2591.1878711484583, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:ImageVolume] ImageVolume when subPath is used should handle image volume with subPath [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 11285.231571072318, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io] CPU Partitioning cluster infrastructure should be configured correctly [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005878030859662013, "AvgDurationMs": 801.674173669468, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM] OLMv1 CRDs should be installed", "TotalTestCount": 3804, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.006309148264984227, "AvgDurationMs": 188.17341107871724, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with catalogsources at version v1alpha1 [apigroup:operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.005879470847623714, "AvgDurationMs": 284.2259103641457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.0056724178681162845, "AvgDurationMs": 10491.124089775562, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][Late] Metrics should report short attach times [Suite:openshift/conformance/parallel]", "TotalTestCount": 4990, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.004809619238476954, "AvgDurationMs": 728.3267407407409, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][Late] Metrics should report short mount times [Suite:openshift/conformance/parallel]", "TotalTestCount": 4990, "TotalFailureCount": 24, "TotalFlakeCount": 0, "FailureRate": 0.004809619238476954, "AvgDurationMs": 731.2944197530863, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API health endpoints should contain the required checks for the oauth-apiserver APIs [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 1765.287680798005, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.0053838951310861425, "AvgDurationMs": 72813.27060679611, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] JSON Patch [apigroup:operator.openshift.io] should delete multiple entries from an array when multiple test precondition provided [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 298.8844257703082, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] OrderedNamespaceDeletion namespace deletion should delete pod first [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 115973.48221945137, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:APIServer] anonymous browsers should get a 403 from / [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 11, "FailureRate": 0.005436067123611439, "AvgDurationMs": 120.08384039900248, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps][Feature:OpenShiftControllerManager] TestTriggers_configChange [apigroup:apps.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005634492895639392, "AvgDurationMs": 3796.8108123249312, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should ensure control plane pods do not run in best-effort QoS [Suite:openshift/conformance/parallel]", "TotalTestCount": 4185, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005495818399044205, "AvgDurationMs": 250.0868329177057, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should ensure pods use downstream images from our release image with proper ImagePullPolicy [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 1140.7268347338936, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:PodSecurity] restricted-v2 SCC should mutate empty securityContext to match restricted PSa profile [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 2986.3471820448876, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:RoleBindingRestrictions] RoleBindingRestrictions should be functional Create a rolebinding when there are no restrictions should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 1, "FailureRate": 0.005633112907176096, "AvgDurationMs": 2995.476666666667, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] Optimized image builds should succeed [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.006014644351464435, "AvgDurationMs": 77291.59223529411, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] s2i build with a root user image should create a root build and pass with a privileged SCC [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 23, "TotalFlakeCount": 3, "FailureRate": 0.006014644351464435, "AvgDurationMs": 63266.99841176469, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] verify /run filesystem contents do not have unexpected content using a simple Docker Strategy Build [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.006014644351464435, "AvgDurationMs": 68565.5699411765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][timing] capture build stages and durations should record build stages and durations for s2i [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 23, "TotalFlakeCount": 1, "FailureRate": 0.005914116739521728, "AvgDurationMs": 67344.45133522728, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][valueFrom] process valueFrom in build strategy environment variables should successfully resolve valueFrom in s2i build environment variables [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005914116739521728, "AvgDurationMs": 65065.077982954535, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc basics can process templates [apigroup:template.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 5506.2619047619055, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc builds complex build webhooks CRUD [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 23, "TotalFlakeCount": 1, "FailureRate": 0.005914116739521728, "AvgDurationMs": 8707.601392045448, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc completion returns expected help messages [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 3575.778927680797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc observe works as expected with cluster operators [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 281.5128011204482, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc service creates and deletes services [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 5471.87493765586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc status returns expected help messages [apigroup:project.openshift.io][apigroup:build.openshift.io][apigroup:image.openshift.io][apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 23, "TotalFlakeCount": 1, "FailureRate": 0.005634492895639392, "AvgDurationMs": 3501.6385714285702, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli][OCPFeatureGate:UpgradeStatus] oc adm upgrade status reports correctly when the cluster is not updating [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 2296.5304761904763, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][Feature:Machines] Managed cluster should have machine resources [apigroup:machine.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3804, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.006046267087276551, "AvgDurationMs": 319.11093294460636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should provide ingress metrics [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 23, "TotalFlakeCount": 2, "FailureRate": 0.006013071895424837, "AvgDurationMs": 4233.6172352941185, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] OpenShift alerting rules [apigroup:image.openshift.io] should have a runbook_url annotation if the alert is critical [Suite:openshift/conformance/parallel]", "TotalTestCount": 4711, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.004882190617703248, "AvgDurationMs": 411.0894782608696, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] OpenShift alerting rules [apigroup:image.openshift.io] should have a valid severity label [Suite:openshift/conformance/parallel]", "TotalTestCount": 4711, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.004882190617703248, "AvgDurationMs": 411.24849275362305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] OpenShift alerting rules [apigroup:image.openshift.io] should have description and summary annotations [Suite:openshift/conformance/parallel]", "TotalTestCount": 4711, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.004882190617703248, "AvgDurationMs": 407.74797101449275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] Should have MCN properties matching associated node properties for nodes in default MCPs [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 428.09252100840337, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should not mistakenly treat 'protocol: SCTP' as 'protocol: TCP', even if the plugin doesn't support SCTP [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 54943.59062344141, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should be able to handle large requests: http", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 20991.74625935162, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should update endpoints: udp", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 106756.44009975063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should support externalTrafficPolicy=Local for type=NodePort", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 7050.670324189527, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] external gateway address when using openshift ovn-kubernetes should match the address family of the pod [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 1, "FailureRate": 0.005436067123611439, "AvgDurationMs": 132001.6274064838, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:bond] should create a pod with bond interface [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 9702.08789915967, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod sysctl should not affect existing pods [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 10249.021400560227, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod sysctl should not affect newly created pods [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 14562.953025210081, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:vlan] should create pingable pods with vlan interface on an in-container master [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005633112907176096, "AvgDurationMs": 15265.721932773115, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork isolates overlapping CIDRs with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 23, "TotalFlakeCount": 2, "FailureRate": 0.005649717514124294, "AvgDurationMs": 192837.60322128845, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions isolates overlapping CIDRs with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 23, "TotalFlakeCount": 2, "FailureRate": 0.005649717514124294, "AvgDurationMs": 199778.76450980396, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork isolates overlapping CIDRs with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 23, "TotalFlakeCount": 2, "FailureRate": 0.005649717514124294, "AvgDurationMs": 191543.6188235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 New Catalog Install should fail to install if it has an invalid reference", "TotalTestCount": 3611, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.006369426751592357, "AvgDurationMs": 37712.41923076923, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with installplans at version v1alpha1 [apigroup:operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005634492895639392, "AvgDurationMs": 255.43812324929974, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with operatorgroups at version v1 [apigroup:operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005634492895639392, "AvgDurationMs": 258.76190476190476, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with subscriptions at version v1alpha1 [apigroup:operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005634492895639392, "AvgDurationMs": 321.09417366946786, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume snapshot CSI Volume Snapshots secrets [Feature:VolumeSnapshotDataSource] volume snapshot create/delete with secrets", "TotalTestCount": 4199, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005477494641581329, "AvgDurationMs": 191028.4096153846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, late binding, with topology", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 72245.49882793013, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005477494641581329, "AvgDurationMs": 64645.66584615382, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 35196.425112219455, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 38646.49184538653, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PVC Protection Verify that PVC in active use by a pod is not removed immediately", "TotalTestCount": 4204, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005470980019029496, "AvgDurationMs": 13711.158628428928, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 12171.833366583538, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 23, "TotalFlakeCount": 0, "FailureRate": 0.005436067123611439, "AvgDurationMs": 12803.215461346634, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005149812734082397, "AvgDurationMs": 50215.01599514563, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] JSON Patch [apigroup:operator.openshift.io] should delete an entry from an array with multiple field owners [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 458.8211764705882, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] JSON Patch [apigroup:operator.openshift.io] should error when the test precondition provided doesn't match [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 294.3139495798319, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] buildconfig secret injector should inject secrets to the appropriate buildconfigs [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 3755.623445378151, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] remove all builds when build configuration is removed oc delete buildconfig should start builds and delete the buildconfig [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005389514943655071, "AvgDurationMs": 9162.37568627451, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] s2i build with a root user image should create a root build and fail without a privileged SCC [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4017, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005476723923325865, "AvgDurationMs": 2366.556666666666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] verify /run filesystem contents are writeable using a simple Docker Strategy Build [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005753138075313808, "AvgDurationMs": 63991.64188235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][subscription-content] builds installing subscription content [apigroup:build.openshift.io] should succeed for RHEL 7 base images [Suite:openshift/conformance/parallel]", "TotalTestCount": 4054, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005426739023186976, "AvgDurationMs": 2342.12452722063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][subscription-content] builds installing subscription content [apigroup:build.openshift.io] should succeed for RHEL 8 base images [Suite:openshift/conformance/parallel]", "TotalTestCount": 4054, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005426739023186976, "AvgDurationMs": 2434.443954154728, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][subscription-content] builds installing subscription content [apigroup:build.openshift.io] should succeed for RHEL 9 base images [Suite:openshift/conformance/parallel]", "TotalTestCount": 4054, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005426739023186976, "AvgDurationMs": 2388.9558166189113, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][valueFrom] process valueFrom in build strategy environment variables should successfully resolve valueFrom in docker build environment variables [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00565698122910774, "AvgDurationMs": 75004.83795454547, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][webhook] TestWebhook [apigroup:build.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 3232.3424369747904, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][webhook] TestWebhookGitHubPing [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 7472.569607843136, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][webhook] TestWebhookGitHubPushWithImage [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 2752.5197759103653, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][webhook] TestWebhookGitHubPushWithImageStream [apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 8, "FailureRate": 0.005389514943655071, "AvgDurationMs": 4075.7458543417374, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc env can set environment variables [apigroup:apps.openshift.io][apigroup:image.openshift.io][apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005389514943655071, "AvgDurationMs": 8478.850560224088, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc help works as expected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 2408.089376558603, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc rsh specific flags should work well when access to a remote shell [Suite:openshift/conformance/parallel]", "TotalTestCount": 4038, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005448241703813769, "AvgDurationMs": 9021.264747474743, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-coreos] [Conformance] CoreOS bootimages TestBootimagesPresent [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/parallel/minimal]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 105.14173669467787, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should have non-Pod host cAdvisor metrics [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005751633986928104, "AvgDurationMs": 3884.899647058824, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Connectivity Pod Lifecycle should be able to have zero downtime on a Blue Green deployment using Services and Readiness Gates", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 36179.47067331672, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack proxy implementation should not be vulnerable to the invalid conntrack state bug [Privileged]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 64384.515835411476, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS HostNetwork spec.Hostname field is silently ignored and the node hostname is used when hostNetwork is set to true for a Pod", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 4442.706009975064, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should allow egress access to server in CIDR block [Feature:NetworkPolicy]", "TotalTestCount": 4230, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005200945626477541, "AvgDurationMs": 40895.424962593504, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access from updated pod [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 57707.0186533666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should deny ingress access to updated pod [Feature:NetworkPolicy]", "TotalTestCount": 4230, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005200945626477541, "AvgDurationMs": 51007.397581047386, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce egress policy allowing traffic to a server in a different namespace based on PodSelector and NamespaceSelector [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 38621.90628428927, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce ingress policy allowing any port traffic to a server on a specific protocol [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 64008.76359102243, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on Multiple PodSelectors and NamespaceSelectors [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 40779.62319201996, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on PodSelector and NamespaceSelector [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 39282.17004987532, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should properly isolate pods that are selected by a policy allowing SCTP, even if the plugin doesn't support SCTP [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 39237.943690773085, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should support allow-all policy [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 58747.286882793014, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should support denying of egress traffic on the client side (even if the server explicitly allows this traffic) [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 50206.377605985035, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for node-Service: http", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 38167.73276807981, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] NoSNAT Should be able to send traffic between Pods without SNAT", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 14239.599177057356, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Proxy version v1 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 5236.858952618454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Proxy version v1 should proxy logs on node with explicit kubelet port using proxy subresource ", "TotalTestCount": 4272, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005149812734082397, "AvgDurationMs": 2101.926699029127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 14395.609725685783, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to update service type to NodePort listening on same port number but different protocols", "TotalTestCount": 4230, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005200945626477541, "AvgDurationMs": 29245.130374064825, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be rejected when no endpoints exist", "TotalTestCount": 4230, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005200945626477541, "AvgDurationMs": 5353.976159600998, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should connect to the ports exposed by restartable init containers", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 10168.511546134665, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should create endpoints for unready pods", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 10895.060473815463, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should delete a collection of services [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 1603.585211970075, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should fail health check node port if there are only terminating endpoints", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 9103.249451371572, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should not be able to connect to terminating and unready endpoints if PublishNotReadyAddresses is false", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 62641.10992518704, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should release NodePorts on delete", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 4943.440748129675, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router converges when multiple routers are writing conflicting upgrade validation status [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005388194954690179, "AvgDurationMs": 73709.22641456583, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router reports the expected host names in admitted routes' statuses [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 12621.425098039213, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Whereabouts] should assign unique IP addresses to each pod in the event of a race condition case [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 16994.868739495803, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tap] should create a pod with a tap interface [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 6, "FailureRate": 0.005389514943655071, "AvgDurationMs": 9169.790504201683, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod sysctls should not affect node [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.0054753608760577405, "AvgDurationMs": 13178.574202898551, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:vlan] should create pingable pods with macvlan interface on an in-container master [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 15173.539439775908, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions does not mirror EndpointSlices in namespaces not using user defined primary networks L3 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 23318.374173669465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] Network Policies when using openshift ovn-kubernetes pods within namespace should be isolated when deny policy is present in L2 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005404077622205846, "AvgDurationMs": 17692.264901960778, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes UDN Pod should react to k8s.ovn.org/open-default-ports annotations changes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 8, "FailureRate": 0.005388194954690179, "AvgDurationMs": 50567.39540616247, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions is isolated from the default network with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 103880.0330812325, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions is isolated from the default network with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 6, "FailureRate": 0.005388194954690179, "AvgDurationMs": 105193.85966386553, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork is isolated from the default network with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 4, "FailureRate": 0.005388194954690179, "AvgDurationMs": 105748.95638655465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork isolates overlapping CIDRs with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005404077622205846, "AvgDurationMs": 91901.47361344541, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising a cluster user defined network [apigroup:user.openshift.io][apigroup:security.openshift.io] Over a VRF-Lite configuration Pods should be able to communicate on a secondary network [Timeout:30m] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 2547.6463305322127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to remove the external certificate and again re-add the same external certificate then also the route is reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 2, "FailureRate": 0.005388194954690179, "AvgDurationMs": 15647.526302521013, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use new external certificate, but RBAC permissions are not added route update is rejected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 4, "FailureRate": 0.005388194954690179, "AvgDurationMs": 7910.0305602240915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use new external certificate, but secret does not exist route update is rejected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 3, "FailureRate": 0.005388194954690179, "AvgDurationMs": 7669.60731092437, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use same external certificate then also the route is reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 15105.394509803926, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the secret is deleted and re-created again but RBAC permissions are dropped then routes are not reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 3, "FailureRate": 0.005388194954690179, "AvgDurationMs": 13923.811120448181, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the secret is deleted then routes are not reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 1, "FailureRate": 0.005388194954690179, "AvgDurationMs": 13880.075266106445, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the secret is updated but RBAC permissions are dropped then routes are not reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005388194954690179, "AvgDurationMs": 15032.44226890756, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] decrease memory limit below usage", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 58752.21107692306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with clusterserviceversions at version v1alpha1 [apigroup:operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 361.5051820728291, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator] OLM should be installed with packagemanifests at version v1 [apigroup:packages.operators.coreos.com] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4082, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005389514943655071, "AvgDurationMs": 238.04630252100839, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 65582.4042394015, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes", "TotalTestCount": 3141, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.0070041388092964025, "AvgDurationMs": 49682.01642659281, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-12", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 100290.44605128204, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 7323.396234413965, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 5475.71518703242, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 20414.499251870322, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 18699.315610972568, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 20409.13359102245, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 22032.97114713216, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 18937.433441396508, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 17761.581047381547, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 19242.78024937656, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 25856.179128205134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 44191.37530769232, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 45119.20405128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and a pre-bound PV: test write access", "TotalTestCount": 4199, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.00523934270064301, "AvgDurationMs": 24038.489846153847, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 12047.637680798009, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 9772.566209476314, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 5479.158553615959, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] Nodes should fail with invalid version annotation [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 22, "TotalFlakeCount": 0, "FailureRate": 0.005199716379106594, "AvgDurationMs": 94.14912718204488, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9914.634563591026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0049157303370786515, "AvgDurationMs": 18798.727281553398, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0049157303370786515, "AvgDurationMs": 40096.27400485437, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail update of a custom resource that does not satisfy a x-kubernetes-validations transition rule", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 3970.028877805486, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch] Managed cluster should expose cluster services outside the cluster [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005548216644649934, "AvgDurationMs": 8080.30855882353, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] build without output image building from templates should create an image from a docker template without an output image reference defined [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0053998457186937516, "AvgDurationMs": 71813.34500000003, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] result image should have proper labels set Docker build from a template should create a image from \"test-docker-build.json\" template with proper Docker labels [apigroup:build.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.0053998457186937516, "AvgDurationMs": 74692.6180397727, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Connectivity Pod Lifecycle should be able to connect from a Pod to a terminating Pod", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 24015.60074812968, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack should be able to preserve UDP traffic when initial unready endpoints get ready", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 23672.26379052369, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS HostNetwork spec.Hostname field is not silently ignored and is used for hostname for a Pod", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 4812.941795511222, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should support configurable pod DNS nameservers [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5463.437556109727, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should support configurable pod resolv.conf", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 8501.422294264341, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should work with a search path containing an underscore and a search path with a single dot", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 8832.227680798009, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should work with the pod containing more than 6 DNS search paths and longer than 256 search list characters", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 8245.991546134665, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 7373.2112219451365, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1808.0749875311724, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1201.7127182044892, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should support a Service with multiple endpoint IPs specified in multiple EndpointSlices [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12193.85284289277, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should support a Service with multiple ports specified in multiple EndpointSlices [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9400.238628428931, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1969.0848129675815, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSliceMirroring should mirror a custom Endpoint with multiple subsets and same IP address", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 11155.166159600998, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] EndpointSliceMirroring should mirror a custom Endpoints resource through create update and delete [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 2429.4787281795516, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Ingress API should support creating Ingress API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 2185.0299501246895, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] IngressClass API should support creating IngressClass API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1826.1450623441397, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol API should support creating NetworkPolicy API operations", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5869.516384039899, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol API should support creating NetworkPolicy API with endport field", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 3382.1495261845384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should allow ingress access from updated namespace [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 56682.320473815475, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should deny egress from all pods in a namespace [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 51028.653142144634, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should deny egress from pods based on PodSelector [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 39591.20755610972, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce multiple egress policies with egress allow-all policy taking precedence [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 63570.71456359101, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce multiple ingress policies with ingress allow-all policy taking precedence [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 65282.4464837905, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policies to check ingress and egress policies can be controlled independently based on PodSelector [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 52841.31528678304, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy based on Ports [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 38962.37795511221, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow ingress traffic for a target [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 46008.326334164594, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow ingress traffic from pods in all namespaces [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 34438.57351620947, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should enforce policy to allow traffic based on NamespaceSelector with MatchLabels using default ns label [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 38445.20201995014, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should not allow access by TCP when a policy specifies only UDP [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 39612.19573566085, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should stop enforcing policies after they are deleted [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 73943.53261845386, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should support a 'default-deny-all' policy [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 57314.02910224441, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should support a 'default-deny-ingress' policy [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 52137.48857855362, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Netpol NetworkPolicy between server and client should work with Ingress, Egress specified together [Feature:NetworkPolicy]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 73453.43037406483, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 21466.710099750624, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 21564.86009975062, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 21580.274937655868, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 24007.222319201992, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should be able to handle large requests: udp", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 21002.352443890275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for client IP based session affinity: http [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 40638.71890274313, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for client IP based session affinity: udp [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 43602.19428927682, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for endpoint-Service: http", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 37767.900673316726, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for endpoint-Service: udp", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 38097.5182543641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for multiple endpoint-Services with same selector", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 85022.99628428929, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for node-Service: udp", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 46046.90261845386, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for pod-Service: http", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 37958.64822942642, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should function for pod-Service: udp", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 38233.68426433915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should support basic nodePort: udp functionality", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 34045.397605985054, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking Granular Checks: Services should update endpoints: http", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 107294.11668329177, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Proxy version v1 A set of valid responses are returned for both pod and service Proxy [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5312.248653366584, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Proxy version v1 should proxy logs on node using proxy subresource ", "TotalTestCount": 4272, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0049157303370786515, "AvgDurationMs": 1975.7447572815538, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Proxy version v1 should proxy through a service and a pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 6877.825311720699, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] ServiceCIDR and IPAddress API should support IPAddress API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1579.6187281795508, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] ServiceCIDR and IPAddress API should support ServiceCIDR API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1298.2708977556113, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should allow pods to hairpin back to themselves through services", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5913.7408478803, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9159.30246882793, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 11171.302369077304, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 14374.834713216962, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9438.169077306733, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to connect to terminating and unready endpoints if PublishNotReadyAddresses is true", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 60377.49596009973, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to create a functioning NodePort service [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 14078.684389027429, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 20779.002019950116, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be possible to connect to a service via ExternalIP when the external IP is not assigned to a node", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12539.073915211971, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should be updated after adding or deleting ports ", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 16284.168154613459, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should check NodePort out-of-range", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1418.3673815461352, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should complete a service status lifecycle [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1683.9670822942646, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should connect to the named ports exposed by restartable init containers", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 10330.26246882793, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should fallback to local terminating endpoints when there are no ready endpoints with internalTrafficPolicy=Local", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 36878.50907730673, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should fallback to terminating endpoints when there are no ready endpoints with externallTrafficPolicy=Cluster", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 36970.40369077306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should fallback to terminating endpoints when there are no ready endpoints with internalTrafficPolicy=Cluster", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 37458.05064837905, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should find a service from listing all namespaces [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0049157303370786515, "AvgDurationMs": 1227.2954854368932, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 15594.640723192022, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12090.420673316707, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should preserve source pod IP for traffic thru service cluster IP [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 8767.025985037411, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should prevent NodePort collisions", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1561.9931172069823, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should provide secure master service [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0049157303370786515, "AvgDurationMs": 1128.3664077669898, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should respect internalTrafficPolicy=Local Pod (hostNetwork: true) to Pod", "TotalTestCount": 4230, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004964539007092199, "AvgDurationMs": 13870.91620947631, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should respect internalTrafficPolicy=Local Pod and Node, to Pod (hostNetwork: true)", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 24893.257531172076, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should respect internalTrafficPolicy=Local Pod to Pod", "TotalTestCount": 4230, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004964539007092199, "AvgDurationMs": 14926.455087281798, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should serve a basic endpoint from pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 15545.998379052367, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should serve endpoints on same port and different protocol for internal traffic on Type LoadBalancer ", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 14420.334912718205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should serve multiport endpoints from pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 14641.610423940152, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should test the lifecycle of an Endpoint [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1655.8670573566087, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Services should work after the service has been recreated", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 20227.090473815457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] load balancer should be managed by OpenShift [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 2634.9206733167084, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router converges when multiple routers are writing status [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 2, "FailureRate": 0.005143277002204262, "AvgDurationMs": 30702.69322128852, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router should override the route host with a custom value [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005490196078431373, "AvgDurationMs": 29069.008000000005, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router should serve the correct routes when running with the haproxy config manager [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005226480836236934, "AvgDurationMs": 2284.07979710145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] when FIPS is enabled the HAProxy router should not work when configured with a 1024-bit RSA key [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 5960.196582633052, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] sysctl allowlist update should start a pod with custom sysctl only when the sysctl is added to whitelist [Suite:openshift/conformance/parallel]", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 26625.18187179487, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure CRD of experimental group can not be created [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2351.7928571428565, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure CRD of standard group can not be created [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2342.9278711484594, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure existing CRDs can not be deleted [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2612.4425210084037, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions does not mirror EndpointSlices in namespaces not using user defined primary networks L2 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 22767.25022408964, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L2 primary UDN, cluster-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 35172.51985994397, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L3 primary UDN, cluster-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 34963.34014005601, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork does not mirror EndpointSlices in namespaces not using user defined primary networks L3 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 23172.258151260507, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L3 primary UDN, cluster-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 35615.636330532216, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] Network Policies when using openshift ovn-kubernetes allow ingress traffic to one pod from a particular namespace in L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 32763.654761904756, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] Network Policies when using openshift ovn-kubernetes pods within namespace should be isolated when deny policy is present in L3 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 17030.979859943975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller should create NAD in new created namespaces that apply to namespace-selector [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 5399.9443697479, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes UserDefinedNetwork CRD controller pod connected to UserDefinedNetwork cannot be deleted when being used [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 15300.461764705882, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes UserDefinedNetwork CRD controller should create NetworkAttachmentDefinition according to spec [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2059.3332212885152, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes UserDefinedNetwork CRD controller should delete NetworkAttachmentDefinition when UserDefinedNetwork is deleted [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2273.4666106442573, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork can perform east/west traffic between nodes for two pods connected over a L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 17793.775742296915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork is isolated from the default network with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 3, "FailureRate": 0.005143277002204262, "AvgDurationMs": 106088.66579831934, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork isolates overlapping CIDRs with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 92993.9773389356, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions can perform east/west traffic between nodes two pods connected over a L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 17149.383249299713, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork can perform east/west traffic between nodes two pods connected over a L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005158437730287398, "AvgDurationMs": 17568.404733893552, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes when primary network exist, ClusterUserDefinedNetwork status should report not-ready [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2382.2607843137257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [EgressIP] Advertising EgressIP [apigroup:user.openshift.io][apigroup:security.openshift.io] For cluster user defined networks When the network topology is Layer 3 UDN pods should have the assigned EgressIPs and EgressIPs can be created, updated and deleted [apigroup:route.openshift.io] When the network is IPv6 [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2585.440896358543, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [EgressIP] Advertising EgressIP [apigroup:user.openshift.io][apigroup:security.openshift.io] For the default network Pods should have the assigned EgressIPs and EgressIPs can be created, updated and deleted [apigroup:route.openshift.io] When the network is IPv4 [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2595.747002801121, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [EgressIP] Advertising EgressIP [apigroup:user.openshift.io][apigroup:security.openshift.io] For the default network Pods should have the assigned EgressIPs and EgressIPs can be created, updated and deleted [apigroup:route.openshift.io] When the network is IPv6 [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2614.4683753501395, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising a cluster user defined network [apigroup:user.openshift.io][apigroup:security.openshift.io] Over the default VRF When the network topology is Layer 2 External host should be able to query route advertised pods by the pod IP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2755.211568627451, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising a cluster user defined network [apigroup:user.openshift.io][apigroup:security.openshift.io] Over the default VRF When the network topology is Layer 2 Pods should communicate with external host without being SNATed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2734.861428571429, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising a cluster user defined network [apigroup:user.openshift.io][apigroup:security.openshift.io] Over the default VRF When the network topology is Layer 3 External host should be able to query route advertised pods by the pod IP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2659.7127731092446, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising the default network [apigroup:user.openshift.io][apigroup:security.openshift.io] External host should be able to query route advertised pods by the pod IP [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2635.112240896358, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising the default network [apigroup:user.openshift.io][apigroup:security.openshift.io] Pods should communicate with external host without being SNATed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.005143277002204262, "AvgDurationMs": 2558.227843137255, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with invalid setup the router should not support external certificate if inline certificate is also present [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005143277002204262, "AvgDurationMs": 7129.896610644257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and routes are reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 2, "FailureRate": 0.005143277002204262, "AvgDurationMs": 33241.25425770308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use new external certificate, but secret is not of type kubernetes.io/tls route update is rejected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005143277002204262, "AvgDurationMs": 8372.680336134454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the route is updated to use same external certificate, but RBAC permissions are dropped route update is rejected [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 1, "FailureRate": 0.005143277002204262, "AvgDurationMs": 7713.141400560223, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with valid setup the router should support external certificate and the secret is deleted and re-created again then routes are reachable [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 2, "FailureRate": 0.005143277002204262, "AvgDurationMs": 32319.06761904762, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][endpoints] admission [apigroup:config.openshift.io] blocks manual creation of EndpointSlices pointing to the cluster or service network [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 21, "TotalFlakeCount": 3, "FailureRate": 0.005143277002204262, "AvgDurationMs": 3372.8225210084033, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Lifecycle sleep action zero value when create a pod with lifecycle hook using sleep action with a duration of zero seconds prestop hook using sleep action with zero duration", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 7572.974463840399, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one container, one restartable init container - decrease init container CPU", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 26108.07625641026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted with a /healthz http liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 244193.0708205128, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator][Jira:OLM] OLMv0 within a namespace PolarionID:43114-[OTP][Skipped:Disconnected]Subscription status should show the message for InstallPlan failure conditions", "TotalTestCount": 425, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.04941176470588235, "AvgDurationMs": 201477.574047619, "PeriodStart": "2025-12-01", "PeriodEnd": "2025-12-03", "DaysWithData": 3 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock fsgroup as mount option Delegate FSGroup to CSI driver [LinuxOnly] should not pass FSGroup to CSI driver if it is set in pod and driver supports VOLUME_MOUNT_GROUP", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 100420.64950124687, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should require VolumeAttach for drivers with attachment", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 46455.761321695754, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should not expand volume if resizingOnDriver=off, resizingOnSC=on", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 206315.94014962603, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should not update fsGroup if update from File to None", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 104879.9251870324, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should not update fsGroup if update from detault to None", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 104614.59837905238, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from detault to File", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 104717.10713216958, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume snapshot CSI Volume Snapshots [Feature:VolumeSnapshotDataSource] volumesnapshotcontent and pvc in Bound state with deletion timestamp set should not get deleted while snapshot finalizer exists", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 196218.35497435887, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity disabled", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 49273.290598503736, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity unused", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 49466.1249127182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, late binding, no topology", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 70695.24633416455, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI PodInfoOnMount Update should be passed when update from false to true", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 54658.4521945137, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should not be passed when podInfoOnMount=false", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 45173.41157107231, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 29868.661795511227, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 80112.97458852867, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 1045.4608974358978, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 979.8927182044889, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 1046.638564102564, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 64552.73997506234, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 41741.23371794873, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 56582.906359102235, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 75605.73397435894, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSIStorageCapacity should support CSIStorageCapacities API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1899.1847880299247, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 4082.951546134663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 32129.981571072327, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5403.685486284287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5399.584663341647, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5424.030448877804, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] files with FSGroup ownership should support (root,0644,tmpfs)", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5395.798079800498, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] nonexistent volume subPath should have the correct mode and owner using FSGroup", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5395.401596009976, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] volume on tmpfs should have the correct mode using FSGroup", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5398.981571072319, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : configmap", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 33201.70698254364, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5457.622842892768, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 32905.31401496258, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 24591.91675810474, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 18701.24174563591, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 20836.601795511222, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 19261.863615960097, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 1048.1773316708234, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 32088.83653366584, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 24339.567356608488, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 20223.977107231913, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 19578.61995012469, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 37553.474114713215, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 20483.19972568579, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 22179.054282051275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 35379.538051282056, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 18612.512589743594, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 29478.831051282057, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 25532.66059850375, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 19106.489923076915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test phase transition timestamp is set", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 22641.140256410254, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PVC and non-pre-bound PV: test write access", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 24518.525717948713, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs should create a non-pre-bound PV and PVC: test write access ", "TotalTestCount": 4199, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.0050011907597046915, "AvgDurationMs": 18013.753, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 10385.237605985038, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: block] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 13597.642942643395, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: block] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12926.692094763092, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 14027.556059850374, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12549.841371571076, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12126.137381546132, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9283.189002493764, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 12664.186334164588, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 9117.812044887783, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5405.371645885287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5416.261970074813, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide podname as non-root with fsgroup and defaultMode [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5424.568453865335, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 5426.8061596009975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.004963365634601749, "AvgDurationMs": 25369.366009975074, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions etcd", "TotalTestCount": 6807, "TotalFailureCount": 21, "TotalFlakeCount": 0, "FailureRate": 0.003085059497576025, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 10072.8493765586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0046816479400749065, "AvgDurationMs": 64349.40415048543, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0046816479400749065, "AvgDurationMs": 1857.0002184466027, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Discovery Custom resource should have storage version hash", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 2419.7513715710734, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] s2i build with a quota Building from a template should create an s2i build with a quota and run it [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.005142710208279763, "AvgDurationMs": 84546.72451704547, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl validation should create/apply a CR with unknown fields for CRD with no validation schema", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 13067.08463840399, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl validation should detect unknown metadata fields in both the root and embedded object of a CR", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 15520.361870324192, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] oc debug ensure it works with image streams [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 20, "TotalFlakeCount": 3, "FailureRate": 0.00522875816993464, "AvgDurationMs": 14299.797882352947, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]", "TotalTestCount": 4185, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0047789725209080045, "AvgDurationMs": 20434.6510723192, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] ServiceCIDR should be blocked [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 624.4616957605984, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] load balancer should not be managed by OpenShift [Suite:openshift/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 2470.5937157107232, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressRouterCNI] should ensure ipv4 egressrouter cni resources are created [apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004977600796416127, "AvgDurationMs": 9076.890695652173, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressRouterCNI] when using openshift ovn-kubernetes should ensure ipv6 egressrouter cni resources are created [apigroup:operator.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4018, "TotalFailureCount": 20, "TotalFlakeCount": 1, "FailureRate": 0.004977600796416127, "AvgDurationMs": 9253.56576811594, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Multus] should use multus to create net1 device from network-attachment-definition [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 6934.947422969186, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router] The HAProxy router should enable openshift-monitoring to pull metrics [Suite:openshift/conformance/parallel]", "TotalTestCount": 3973, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.005033979360684621, "AvgDurationMs": 8546.803541666666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router] The HAProxy router should expose the profiling endpoints [Suite:openshift/conformance/parallel]", "TotalTestCount": 3973, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.005033979360684621, "AvgDurationMs": 8159.390911458334, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:image.openshift.io] The HAProxy router should serve a route that points to two services and respect weights [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 20, "TotalFlakeCount": 1, "FailureRate": 0.00522875816993464, "AvgDurationMs": 117627.47399999999, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:operator.openshift.io] The HAProxy router should serve routes that were created from an ingress [apigroup:route.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 20, "TotalFlakeCount": 2, "FailureRate": 0.005284015852047556, "AvgDurationMs": 13635.80238235294, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Whereabouts] should use whereabouts net-attach-def to limit IP ranges for newly created pods [apigroup:k8s.cni.cncf.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 48086.27204481793, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod should not start for sysctls not on whitelist [apigroup:k8s.cni.cncf.io] net.ipv4.conf.IFNAME.arp_filter [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 17451.084621848742, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:tuning] pod should not start for sysctls not on whitelist [apigroup:k8s.cni.cncf.io] net.ipv4.conf.all.send_redirects [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 17351.584565826324, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure CRD of experimental group is not installed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 2689.912969187675, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure existing CRDs can not be updated [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 3493.3524089635853, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:GatewayAPI][Feature:Router][apigroup:gateway.networking.k8s.io] Verify Gateway API CRDs and ensure required CRDs should already be installed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 2616.3393277310925, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L2 primary UDN, host-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 32143.696442577035, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L3 primary UDN, host-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 32229.019075630244, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork does not mirror EndpointSlices in namespaces not using user defined primary networks L2 dualstack primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 22837.535854341735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L2 primary UDN, cluster-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 35154.242857142846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L2 primary UDN, host-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 33058.4394677871, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] EndpointSlices mirroring when using openshift ovn-kubernetes created using UserDefinedNetwork mirrors EndpointSlices managed by the default controller for namespaces with user defined primary networks L3 primary UDN, host-networked pods [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 33057.46028011203, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] Network Policies when using openshift ovn-kubernetes allow ingress traffic to one pod from a particular namespace in L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0049127978383689515, "AvgDurationMs": 33550.01557422968, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller should create NAD according to spec in each target namespace and report active namespaces [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 3281.6622408963585, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller when CR is deleted, should delete all managed NAD in each target namespace [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 3308.564453781512, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller when namespace-selector is mutated should create NAD in namespaces that apply to mutated namespace-selector [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 4893.421120448178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes ClusterUserDefinedNetwork CRD Controller when namespace-selector is mutated should delete managed NAD in namespaces that no longer apply to namespace-selector [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 3531.2241176470593, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using ClusterUserDefinedNetwork can perform east/west traffic between nodes two pods connected over a L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0049127978383689515, "AvgDurationMs": 18131.027086834736, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions can perform east/west traffic between nodes for two pods connected over a L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0049127978383689515, "AvgDurationMs": 16259.69473389356, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using NetworkAttachmentDefinitions isolates overlapping CIDRs with L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 20, "TotalFlakeCount": 1, "FailureRate": 0.0049127978383689515, "AvgDurationMs": 91681.21218487396, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork can perform east/west traffic between nodes for two pods connected over a L2 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4071, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.0049127978383689515, "AvgDurationMs": 17550.363837535013, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes created using UserDefinedNetwork is isolated from the default network with L3 primary UDN [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 1, "FailureRate": 0.004898359049718345, "AvgDurationMs": 104542.12330532212, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkSegmentation][Feature:UserDefinedPrimaryNetworks] when using openshift ovn-kubernetes when primary network exist, UserDefinedNetwork status should report not-ready [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 1591.2825490196078, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [EgressIP] Advertising EgressIP [apigroup:user.openshift.io][apigroup:security.openshift.io] For cluster user defined networks When the network topology is Layer 3 UDN pods should have the assigned EgressIPs and EgressIPs can be created, updated and deleted [apigroup:route.openshift.io] When the network is IPv4 [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 2621.933193277312, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteAdvertisements][Feature:RouteAdvertisements][apigroup:operator.openshift.io] when using openshift ovn-kubernetes [PodNetwork] Advertising a cluster user defined network [apigroup:user.openshift.io][apigroup:security.openshift.io] Over the default VRF When the network topology is Layer 3 Pods should communicate with external host without being SNATed [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 2637.845434173669, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with invalid setup the router should not support external certificate if the route termination type is Passthrough [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 7058.8889355742285, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with invalid setup the router should not support external certificate if the secret is in a different namespace [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 7282.903053221287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with invalid setup the router should not support external certificate if the secret is not of type kubernetes.io/tls [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 6925.64949579832, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:RouteExternalCertificate][Feature:Router][apigroup:route.openshift.io] with invalid setup the router should not support external certificate without proper permissions [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004898359049718345, "AvgDurationMs": 7013.342268907562, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][endpoints] admission [apigroup:config.openshift.io] blocks manual creation of Endpoints pointing to the cluster or service network [Suite:openshift/conformance/parallel]", "TotalTestCount": 4083, "TotalFailureCount": 20, "TotalFlakeCount": 3, "FailureRate": 0.004898359049718345, "AvgDurationMs": 3266.8342857142866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests and increase CPU limits", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 12707.866384615392, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one container - increase CPU (NotRequired) \u0026 memory (RestartContainer)", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 36517.976666666655, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one restartable init container - increase CPU \u0026 memory", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 24730.70351282052, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended (pod generation) [Feature:PodObservedGenerationTracking] [FeatureGate:PodObservedGenerationTracking] [Beta] Pod Generation issue 500 podspec updates and verify generation and observedGeneration eventually converge", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 307882.10282051296, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] ResourceSlice Controller creates slices [ConformanceCandidate]", "TotalTestCount": 2954, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.006770480704129994, "AvgDurationMs": 86117.71483146066, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 244046.74582051285, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Static provisioning should honor pv retain reclaim policy when deleting pv then pvc", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 92083.74563591027, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Static provisioning should honor pv retain reclaim policy when deleting pvc then pv", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 91274.00441396509, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should pass SELinux mount option for RWOP volume and Pod with SELinux context set [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 72397.49251282049, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should require VolumeAttach for ephemermal volume and drivers with attachment", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 35886.60917705736, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume by restarting pod if attach=off, nodeExpansion=on", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 80356.70269326685, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should not have staging_path missing in node expand volume pod if attach=on, nodeExpansion=on", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 94915.98299251872, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI online volume expansion with secret should expand volume without restarting pod if attach=on, nodeExpansion=on, csiNodeExpandSecret=on", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 88840.41117206981, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery should allow recovery if controller expansion fails with infeasible error", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 161991.22488778058, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from File to default", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 105816.56087281796, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should modify fsGroup if fsGroupPolicy=default", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 64801.538204488774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should not be plumbed down when csiServiceAccountTokenEnabled=false", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 45758.33887780547, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume snapshot CSI Snapshot Controller metrics [Feature:VolumeSnapshotDataSource] snapshot controller should emit dynamic CreateSnapshot, CreateSnapshotAndReady, and DeleteSnapshot metrics", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 25319.89366666666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity storage capacity exhausted, immediate binding", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 71607.56296758108, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI workload information using mock driver contain ephemeral=true when using inline volume", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 31533.753615960108, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 28898.78603491271, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 1057.6430769230767, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 40020.877132169575, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 100694.59841025644, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 49434.448453865334, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004866180048661801, "AvgDurationMs": 52237.13311557789, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 1014.6749625935158, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 1044.6136923076924, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 63677.697692307695, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 40255.978952618454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 40186.783990024946, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 36535.567630922706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 102370.99107692306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 1023.204717948718, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 75656.94784615388, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 1038.1192051282046, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 959.6338403990026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 48998.07301745636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSIInlineVolumes should support CSIVolumeSource in Pod API [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 1314.37650872818, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSINodes CSI Conformance should run through the lifecycle of a csinode [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 1413.3185286783043, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5389.454438902745, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5475.400922693268, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5432.776284289276, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 25843.114089775558, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5334.425660847881, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5427.096059850374, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide podname as non-root with fsgroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5367.675885286783, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide podname as non-root with fsgroup and defaultMode [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5402.864912718204, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5430.90259351621, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes pod should support memory backed volumes of specified size", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 3812.481820448879, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5383.718778054866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5398.637381546134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5424.599700748129, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5441.795261845385, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 6169.258204488777, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 7472.105037406484, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 14547.839476309226, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 29295.67990024938, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 16284.224912718206, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 20004.260374064834, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 20570.45209476309, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 20636.833067331663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 18945.604314214466, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 23633.59551122195, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 17578.487157107236, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 37435.45768079798, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 17400.81775561097, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 18976.104538653362, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 32776.86311720698, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 20244.25169576061, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 37114.76563591022, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 18777.4997755611, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 17990.30132169577, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 32567.702169576063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 40606.05443589744, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 22190.365179487177, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 19133.445256410258, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 34847.543307692315, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 41222.90056410255, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 35673.3712820513, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 18760.477589743583, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 9417.287057356605, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 44028.5815897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PVC Protection Verify \"immediate\" deletion of a PVC that is not in active use by a pod", "TotalTestCount": 4229, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004729250413809411, "AvgDurationMs": 14656.412568578551, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test phase transition timestamp multiple updates", "TotalTestCount": 4199, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004763038818766373, "AvgDurationMs": 23569.54012820513, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: block] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 10047.946334164591, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 13453.545012468827, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 10485.811620947632, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 9495.836807980055, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 12521.239775561096, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 9215.686658354114, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 12706.800024937655, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 12331.40940149626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 34869.847680798004, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5427.45546134663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5356.822793017456, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5434.919426433916, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5462.989077306733, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5449.302119700748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 5387.443566084787, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] StorageClasses CSI Conformance should run through the lifecycle of a StorageClass [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 1478.8217705735658, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 25297.493690773063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Container restart should verify that container can restart successfully after configmaps modified", "TotalTestCount": 4231, "TotalFailureCount": 20, "TotalFlakeCount": 0, "FailureRate": 0.004727014890096904, "AvgDurationMs": 98765.51074812964, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004447565543071161, "AvgDurationMs": 17561.040946601945, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST NOT fail validation for create of a custom resource that satisfies the x-kubernetes-validations rules", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 4096.243241895261, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail validation for create of a custom resource that does not satisfy the x-kubernetes-validations rules", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 3884.262892768079, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] build have source revision metadata started build should contain source revision information [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004885574697865775, "AvgDurationMs": 57758.93349431816, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] build with empty source started build should build even with an empty source in build config [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 19, "TotalFlakeCount": 1, "FailureRate": 0.004885574697865775, "AvgDurationMs": 57576.051903409105, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig should prune errored builds based on the failedBuildsHistoryLimit setting [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 19, "TotalFlakeCount": 1, "FailureRate": 0.004968619246861925, "AvgDurationMs": 29997.585588235303, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig should prune failed builds based on the failedBuildsHistoryLimit setting [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004968619246861925, "AvgDurationMs": 36255.98976470589, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][pullsearch] docker build where the registry is not specified Building from a Dockerfile whose FROM image ref does not specify the image registry should create a docker build that has buildah search from our predefined list of image registries and succeed [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004885574697865775, "AvgDurationMs": 48855.891903409094, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][valueFrom] process valueFrom in build strategy environment variables should fail resolving unresolvable valueFrom in sti build environment variable references [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 19, "TotalFlakeCount": 1, "FailureRate": 0.004885574697865775, "AvgDurationMs": 9538.735482954544, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Internal connectivity for TCP and UDP on ports 9000-9999 is allowed [Serial:Self] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4191, "TotalFailureCount": 19, "TotalFlakeCount": 1, "FailureRate": 0.004533524218563589, "AvgDurationMs": 16997.487955112207, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router] The HAProxy router should expose a health check on the metrics port [Suite:openshift/conformance/parallel]", "TotalTestCount": 3973, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.00478228039265039, "AvgDurationMs": 7220.470026041665, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router should override the route host for overridden domains with a custom value [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 19, "TotalFlakeCount": 1, "FailureRate": 0.0049673202614379085, "AvgDurationMs": 30049.091911764703, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router should serve the correct routes when scoped to a single namespace and label set [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0049673202614379085, "AvgDurationMs": 28599.812999999995, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one container - increase CPU \u0026 memory with an extended resource", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 20480.575666666675, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 244190.8615461346, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock fsgroup as mount option Delegate FSGroup to CSI driver [LinuxOnly] should pass FSGroup to CSI driver if it is set in pod and driver supports VOLUME_MOUNT_GROUP", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 102519.37461346634, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy changes using mock driver should honor pv reclaim policy after it is changed from retain to deleted", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 88534.0280548628, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv delete reclaim policy when deleting pv then pvc", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 28350.808902743134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv retain reclaim policy when deleting pvc then pv", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 88035.19556109725, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Static provisioning should honor pv delete reclaim policy when deleting pv then pvc", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 32100.642443890272, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Static provisioning should honor pv delete reclaim policy when deleting pvc", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 32619.078329177053, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not unstage RWOP volume when starting a second pod with the same SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 78388.4881025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should not require VolumeAttach for drivers without attachment", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 33748.89216957607, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume by restarting pod if attach=on, nodeExpansion=on", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 96787.3735660848, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI online volume expansion should expand volume without restarting pod if attach=on, nodeExpansion=on", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 85127.94596009972, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery recovery should be possible for node-only expanded volumes with final error", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 133368.33187032418, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery recovery should be possible for node-only expanded volumes with infeasible error", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 135270.9149127182, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery recovery should not be possible in partially expanded volumes", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 129321.99817955113, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from None to File", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 104633.8882543641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy Update [LinuxOnly] should update fsGroup if update from None to default", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 104363.3011970075, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume snapshot CSI Snapshot Controller metrics [Feature:VolumeSnapshotDataSource] snapshot controller should emit pre-provisioned CreateSnapshot, CreateSnapshotAndReady, and DeleteSnapshot metrics", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 25115.68105128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should be passed when podInfoOnMount=true", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 44710.20551122195, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 50460.202443890266, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004622871046228711, "AvgDurationMs": 894.5632160804023, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 42306.47707692308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 36570.43680798006, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 56791.46067331672, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 52762.084713216944, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 40885.02693266833, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 39171.70995012468, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 39202.807506234414, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 40488.24453865336, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 99083.09135897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 40108.168333333335, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 100357.56458974358, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 79915.58014962592, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 37311.03423940151, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 36442.03416458852, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 79073.45591022444, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 36615.43087281796, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 50184.51850374064, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 969.3491025641026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 101944.97784615382, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSIInlineVolumes should run through the lifecycle of a CSIDriver [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 1465.4742144638406, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5327.863216957606, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5440.50301745636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5349.706259351619, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 3709.0568079800496, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] volume on default medium should have the correct mode using FSGroup", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5317.3904738154615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : projected", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 33074.14194513716, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Flexvolumes should be mountable when attachable [Feature:Flexvolumes]", "TotalTestCount": 3428, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.005542590431738623, "AvgDurationMs": 1009.8256104651159, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] HostPath should give a volume the correct mode [LinuxOnly] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5455.269451371572, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 7554.826708229428, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 7563.631720698253, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 6098.087331670822, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 25719.36346633416, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 16151.322992518708, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 9959.416907730674, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 16184.000723192019, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 16166.770723192018, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 16143.92219451372, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 16106.051620947632, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 21099.00239401496, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19738.86568578553, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 38439.68920199502, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 20916.13306733167, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19293.06451371571, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19314.255511221938, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 33774.66144638404, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19278.904289276797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 17824.90922693267, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19420.92149625935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19204.399152119702, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 18089.503416458854, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19300.16007481297, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 19230.495187032422, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 23447.303765586028, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 17590.323815461346, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 1048.888846153846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 23438.77482051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 967.3793266832918, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 20767.16107692308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 18688.426128205127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 19345.898230769228, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 19056.85230769231, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 20166.87492307692, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 891.4598503740651, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 1009.5984538653368, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 890.7921025641026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 899.5316923076925, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 22161.554769230774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 21829.257974358963, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 22155.255897435887, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 19114.830256410263, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 19170.48184615384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 9441.180573566084, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PVC Protection Verify that scheduling of a pod that uses PVC that is being deleted fails and the pod becomes Unschedulable", "TotalTestCount": 4204, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004519505233111323, "AvgDurationMs": 13839.268104738156, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes CSI Conformance should run through the lifecycle of a PV and a PVC [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 4072.8481795511225, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS when invoking the Recycle reclaim policy should test that a PV becomes Available and is clean after the PVC is deleted.", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 33764.6285897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 3 PVs and 3 PVCs: test write access", "TotalTestCount": 4199, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.004524886877828055, "AvgDurationMs": 29896.753717948715, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithformat] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 11046.43508728179, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 10170.413715710722, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 13524.08246882793, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-bindmounted] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 9213.780049875311, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 9517.262119700747, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 9346.324688279305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 11903.82578553616, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5407.147032418953, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5324.537605985037, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5323.049625935164, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5390.666957605988, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5592.186832917706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5370.750074812968, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5532.273516209475, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5397.037680798005, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 6005.883391521197, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5348.267381546134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 5434.336084788032, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 19, "TotalFlakeCount": 0, "FailureRate": 0.0044906641455920585, "AvgDurationMs": 25282.429700748136, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004213483146067416, "AvgDurationMs": 18303.134805825237, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should support orphan deletion of custom resources", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 111110.74423940151, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should update/patch PodDisruptionBudget status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 4601.426982543642, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should not delete PVCs when there is another controller", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 54781.692194513715, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] build can reference a cluster service with a build being created from new-build should be able to run a build that references a cluster service [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3527, "TotalFailureCount": 18, "TotalFlakeCount": 7, "FailureRate": 0.005103487383045081, "AvgDurationMs": 205666.13463878323, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] build without output image building from templates should create an image from a S2i template without an output image reference defined [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 18, "TotalFlakeCount": 1, "FailureRate": 0.004628439187451787, "AvgDurationMs": 55262.81542613636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] imagechangetriggers imagechangetriggers should trigger builds of all types [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 18, "TotalFlakeCount": 1, "FailureRate": 0.004628439187451787, "AvgDurationMs": 4356.283522727274, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] oc new-app should fail with a --name longer than 58 characters [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 18, "TotalFlakeCount": 1, "FailureRate": 0.004628439187451787, "AvgDurationMs": 3534.960113636364, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] oc new-app should succeed with a --name of 58 characters [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3527, "TotalFailureCount": 18, "TotalFlakeCount": 9, "FailureRate": 0.005103487383045081, "AvgDurationMs": 154461.2542205323, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] oc new-app should succeed with an imagestream [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 18, "TotalFlakeCount": 1, "FailureRate": 0.004707112970711297, "AvgDurationMs": 4280.240852941177, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig buildconfigs should have a default history limit set when created via the group api [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004707112970711297, "AvgDurationMs": 2738.0749705882363, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig should prune builds after a buildConfig change [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004707112970711297, "AvgDurationMs": 15859.625529411767, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds] prune builds based on settings in the buildconfig should prune canceled builds based on the failedBuildsHistoryLimit setting [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3824, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004707112970711297, "AvgDurationMs": 18421.148558823534, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-builds][Feature:Builds][valueFrom] process valueFrom in build strategy environment variables should fail resolving unresolvable valueFrom in docker build environment variable references [apigroup:build.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3889, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004628439187451787, "AvgDurationMs": 9545.905710227275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl validation should create/apply a valid CR for CRD with validation schema", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 13070.296583541147, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] The HAProxy router should run even if it has no access to update status [apigroup:image.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 3825, "TotalFailureCount": 18, "TotalFlakeCount": 24, "FailureRate": 0.004705882352941176, "AvgDurationMs": 18936.17694117647, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Node Lifecycle should run through the lifecycle of a node [Conformance]", "TotalTestCount": 4203, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004282655246252677, "AvgDurationMs": 1392.89058524173, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu requests and limits - resize with equivalents", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 23131.30748717949, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] pod-resize-limit-ranger-test", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 10641.959128205131, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should not be ready with an exec readiness probe timeout [MinimumKubeletVersion:1.20] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 63813.382568578556, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv delete reclaim policy when deleting pvc", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 28010.670548628423, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy using mock driver Dynamic provisioning should honor pv retain reclaim policy when deleting pv then pvc", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 88914.31059850375, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should add SELinux mount option to existing mount options [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 71943.82553846155, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for Pod without SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 72336.28469230767, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWOP volume when starting a second pod with different SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 91917.96700000002, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion Expansion with recovery should record target size in allocated resources", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 86801.78049875311, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should not modify fsGroup if fsGroupPolicy=None", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 65090.105885286786, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should be plumbed down when csiServiceAccountTokenEnabled=true", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 45441.36885286782, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume service account token CSIServiceAccountToken token should not be plumbed down when CSIDriver is not deployed", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 44944.402618453874, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, insufficient capacity", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 29989.44506234414, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity storage capacity unlimited", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 33225.29802992519, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI PodInfoOnMount Update should not be passed when update from true to false", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 57635.26029925188, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock workload info CSI workload information using mock driver should not be passed when CSIDriver does not exist", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 46336.211920199516, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 28991.84231920199, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 890.1425436408978, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 34393.37526184538, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 898.056608478803, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 40784.590153846155, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 41542.83461538462, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 945.8685286783041, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should block a second pod from using an in-use ReadWriteOncePod volume on the same node", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 46537.34433915211, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 48343.48673316706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 41564.13700000001, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 874.7724937655861, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 24387.969052369073, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 889.0302493765587, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 37484.75755610974, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 37844.87319201995, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 960.1303076923078, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root with FSGroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5339.357805486285, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5307.675885286781, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5305.7791521197005, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5262.820099750625, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5298.108229426432, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 6948.414314214466, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5324.0196259351615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5316.529625935162, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5308.309052369077, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5296.4456857855375, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5279.675885286782, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] new files should be created with FSGroup ownership when container is non-root", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5248.974588528681, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes when FSGroup is specified [LinuxOnly] new files should be created with FSGroup ownership when container is root", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5345.425511221944, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Ephemeralstorage When pod refers to non-existent ephemeral storage should allow deletion of pod with invalid volume : secret", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 33068.98703241896, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Flexvolumes should be mountable when non-attachable", "TotalTestCount": 3428, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.005250875145857643, "AvgDurationMs": 926.8248546511627, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] HostPath should support subPath [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5368.94221945137, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 905.5396758104738, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 7416.7031920199515, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 9342.16251870324, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 14757.882169576062, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 16116.015311720694, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 890.0162094763093, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 14607.763566084786, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 34574.01837905235, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 14759.468428927687, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 870.5611970074814, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19737.90610972569, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 871.2452119700748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 945.0323940149624, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 37503.7374563591, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 20248.195885286787, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 18999.977406483787, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19970.1312967581, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 947.0556109725687, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 18897.711471321694, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 870.0200997506234, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19387.712718204497, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19198.93059850374, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 23963.28391521197, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19722.74371571072, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 17860.40049875313, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19528.95102244389, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 19208.243042394017, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4078, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004413928396272682, "AvgDurationMs": 33880.78927648579, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 895.1583076923075, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 27125.838435897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 22251.231102564107, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 20481.827307692303, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 23816.046009975063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 9281.948603491273, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 876.9935162094764, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 24817.88623076923, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 10034.374812967586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4078, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004413928396272682, "AvgDurationMs": 33692.332041343674, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 864.6025685785536, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 26855.35053846154, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 21179.3971025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 22230.586512820508, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 18518.20576923077, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 24769.929564102567, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 26147.62248717948, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 892.5731421446384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 31543.2463076923, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 32643.629127182045, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PV Protection Verify \"immediate\" deletion of a PV that is not bound to a PVC", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 1765.2964089775562, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PV Protection Verify that PV bound to a PVC is not removed immediately", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 4205.509950124687, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV and a pre-bound PVC: test write access", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 22739.55538461539, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with multiple PVs and PVCs all in same ns should create 2 PVs and 4 PVCs: test write access", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 22900.180282051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-expansion loopback local block volume should support online expansion on node", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 65720.09501246884, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local Pod with node different from PV's NodeAffinity should fail scheduling due to different NodeAffinity", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5983.839600997506, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: blockfswithoutformat] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 13237.437481296756, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: dir-link-bindmounted] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 9565.766284289275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: tmpfs] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 9781.462069825433, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5322.851645885288, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5316.745137157108, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root with FSGroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5322.240773067331, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 25223.278802992518, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5307.700648379051, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5272.898827930175, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 7052.556683291771, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 32099.346009975063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5292.513466334165, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 5328.101795511221, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets should be immutable if `immutable` field is set [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 1179.735860349127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 25310.37850374064, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 25370.494987531172, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] VolumeAttachment Conformance should apply changes to a volumeattachment status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.004254313401087214, "AvgDurationMs": 1272.856483790524, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] VolumeAttributesClass [FeatureGate:VolumeAttributesClass] should run through the lifecycle of a VolumeAttributesClass", "TotalTestCount": 4199, "TotalFailureCount": 18, "TotalFlakeCount": 0, "FailureRate": 0.0042867349368897354, "AvgDurationMs": 2222.146846153847, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.003979400749063671, "AvgDurationMs": 40151.528252427175, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (cpu, memory quota set) against a pod with same priority class.", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 2432.4106733167077, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should provide basic identity", "TotalTestCount": 4229, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.0040198628517379995, "AvgDurationMs": 104510.70660847882, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support inline execution and attach with websockets or fallback to spdy", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 57764.84089775561, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service and client is hostNetwork", "TotalTestCount": 3799, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004474861805738352, "AvgDurationMs": 16686.991130434777, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:operator.openshift.io] The HAProxy router should set Forwarded headers appropriately [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 17, "TotalFlakeCount": 1, "FailureRate": 0.004491413474240423, "AvgDurationMs": 13277.765882352944, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io][apigroup:operator.openshift.io] The HAProxy router should support reencrypt to services backed by a serving certificate automatically [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004491413474240423, "AvgDurationMs": 19329.78547058824, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Downward API tests for pod level resources should provide default limits.cpu/memory from pod level resources", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 5811.849769230769, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Downward API tests for pod level resources should provide default limits.cpu/memory from pod level resources or node allocatable", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 5758.623794871795, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container - increase memory request (NoRestart memory resize policy)", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 12810.39482051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with CPU requests + limits, cpu requests - remove memory requests", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 5135.45535897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one container - increase CPU \u0026 memory", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 19552.564205128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock honor pv reclaim policy CSI honor pv reclaim policy changes using mock driver should honor pv reclaim policy after it is changed from deleted to retain", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 28328.785935162086, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for CSI driver that does not support SELinux mount [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 72607.12697435898, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume attach CSI attach test using mock driver should preserve attachment policy when no CSIDriver present", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 46248.315361596004, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI Volume expansion should expand volume without restarting pod if nodeExpansion=off", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 86084.49523690771, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume expansion CSI online volume expansion should expand volume without restarting pod if attach=off, nodeExpansion=on", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 57251.40506234416, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume fsgroup policies CSI FSGroupPolicy [LinuxOnly] should modify fsGroup if fsGroupPolicy=File", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 64668.44269326683, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, have capacity", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 49904.73755610972, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock volume storage capacity CSIStorageCapacity CSIStorageCapacity used, no capacity", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 29887.59394014962, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 50792.26314214465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 79853.82802992518, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 58556.31815461348, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 40265.04012468828, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 873.8250256410254, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 873.0643333333336, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 37986.14087281796, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 50527.47728179549, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 36245.81890274315, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 79382.03182044889, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5301.863142144637, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5322.428578553616, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5307.85132169576, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 1173.7662094763095, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5306.595860349125, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5321.613640897757, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5286.0937905236915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 3475.422344139649, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] HostPath should support r/w [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5473.448054862845, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 34475.39294264339, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 9622.937730673317, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 16143.738428927683, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 29063.542468827927, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 25229.597680798004, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 18411.539925187033, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 880.7425685785536, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 868.0302493765587, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 862.6464089775562, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 23397.76658354115, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 19214.054513715706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 17767.32391521197, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 37594.079576059856, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 878.4822194513714, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 17452.389301745632, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 884.8529426433915, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 868.5343142144638, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 17510.252917705748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 865.7383541147133, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 890.8647132169574, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 875.4479999999996, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 22304.554897435904, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 21090.6047948718, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 32262.459600997503, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 867.9067331670824, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 20116.811743589744, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4199, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004048582995951417, "AvgDurationMs": 32362.937205128208, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes CSI Conformance should apply changes to a pv/pvc status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 1406.9795261845386, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes NFS with Single PV - PVC pairs create a PV: test phase transition timestamp is set and phase is Available", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 11446.74598503741, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local Pod with node different from PV's NodeAffinity should fail scheduling due to different NodeSelector", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5945.810798004988, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected combined should project all components that make up the projection API [Projection] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5380.437705735659, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected configMap should be consumable from pods in volume as non-root with FSGroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5302.869002493763, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5359.883092269326, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should provide podname as non-root with fsgroup [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5284.305910224441, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5264.980349127183, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 7099.786758104739, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 6011.308778054863, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 5375.234538653365, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 31917.950224438897, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] VolumeAttachment Conformance should run through the lifecycle of a VolumeAttachment [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.004017962656582368, "AvgDurationMs": 1656.2058603491266, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions network", "TotalTestCount": 6807, "TotalFailureCount": 17, "TotalFlakeCount": 0, "FailureRate": 0.0024974291170853534, "AvgDurationMs": 0.000020242914979757088, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003745318352059925, "AvgDurationMs": 3671.5027912621354, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/json,application/vnd.kubernetes.protobuf\"", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 1094.4081795511222, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should schedule multiple jobs concurrently [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 92062.27009975062, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should not evict unready pods with Default UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 6585.4619201995, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet AvailableReplicas should get updated accordingly when MinReadySeconds is enabled", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 127995.5350623441, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications with PVCs", "TotalTestCount": 4229, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003783400331047529, "AvgDurationMs": 118829.34538653369, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs with a OnScaledown policy", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 39434.3896508728, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 2761.673615960099, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 22887.89344139651, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for pods for Subdomain [Conformance]", "TotalTestCount": 3799, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0042116346406949196, "AvgDurationMs": 12593.668927536228, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for services [Conformance]", "TotalTestCount": 3799, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0042116346406949196, "AvgDurationMs": 13619.024028985505, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:operator.openshift.io] The HAProxy router should respond with 503 to unrecognized hosts [Suite:openshift/conformance/parallel]", "TotalTestCount": 3785, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.004227212681638045, "AvgDurationMs": 7880.8469705882335, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - remove CPU limits", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 5270.831794871794, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, three containers - no change for c1, increase c2 resources, decrease c3 (net decrease for pod)", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 31323.314769230765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, one restartable init container - decrease CPU \u0026 increase memory", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 26475.933974358974, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Guaranteed QoS pod, three containers (c1, c2, c3) - increase: CPU (c1,c3), memory (c2, c3) ; decrease: CPU (c2)", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 29931.524333333342, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted by liveness probe because startup probe delays it", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 244402.3443142145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted by liveness probe because startup probe delays it", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 244069.0336923077, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted with a tcp:8080 liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 244283.73843589742, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container with readiness probe that fails should never be ready and never restart", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 60972.51428205128, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 868.7681795511221, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.0038104310550130982, "AvgDurationMs": 890.568358974359, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 16, "TotalFlakeCount": 0, "FailureRate": 0.003781611912077523, "AvgDurationMs": 862.9197755610973, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API priority and fairness should support FlowSchema API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 2047.8110972568586, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 9699.836857855364, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST evaluate a CRD Validation Rule with oldSelf = nil for new values when optionalOldSelf is true", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 2240.4684615384613, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST fail to update a resource due to JSONSchema errors on changed fields", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 2188.331512820513, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST fail to update a resource due to JSONSchema errors on unchanged uncorrelatable fields", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 2147.1099999999997, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0035112359550561797, "AvgDurationMs": 20643.20672330097, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should create/apply a CR with unknown fields for CRD with no validation schema [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 4283.394713216958, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0035112359550561797, "AvgDurationMs": 1643.8148543689322, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes through scope selectors.", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 14906.434438902748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should mark indexes as failed when the FailIndex action is matched in podFailurePolicy [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 6619.3652867830415, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 13203.645286783041, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0035112359550561797, "AvgDurationMs": 2970.426140776699, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should adopt matching orphans and release non-matching pods", "TotalTestCount": 4229, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0035469378103570584, "AvgDurationMs": 37405.031496259355, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should not deadlock when a pod's predecessor fails", "TotalTestCount": 4229, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0035469378103570584, "AvgDurationMs": 63829.12897755612, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications for partiton1 and delete pod-0 without failing container", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 58041.55917705735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 84256.9164837905, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should not delete PVC with OnScaledown policy if another controller owns the PVC", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 59672.40952618452, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should guarantee kube-root-ca.crt exist in any namespace [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 2138.9171072319214, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl cluster-info dump should check if cluster-info dump succeeds", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 1852.122543640898, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a ClusterIP service", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 14361.188405797104, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Conntrack should be able to preserve UDP traffic when server pod cycles for a NodePort service", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 17685.70620289855, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide /etc/hosts entries for the cluster [Conformance]", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 7520.883072463766, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for ExternalName services [Conformance]", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 17541.608637681158, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for pods for Hostname [Conformance]", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 9984.322434782607, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should resolve DNS of partial qualified names for the cluster [LinuxOnly]", "TotalTestCount": 3799, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003948407475651487, "AvgDurationMs": 6928.782115942029, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Mount propagation should propagate mounts within defined scopes", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 33059.019564102564, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] NodeLease NodeLease the kubelet should report node status infrequently", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 88661.30314214464, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, mixed containers - add requests", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 16491.90928205128, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory limits only", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 12936.111307692308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with memory requests + limits, cpu requests - remove CPU requests", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 3999.2997948717943, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container, one restartable init container - decrease init container memory requests only", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 20616.42974358975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, three containers - decrease c1 resources, increase c2 resources, no change for c3 (net increase for pod)", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 25250.918333333342, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 244440.2011720698, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 149428.67972568577, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers must create the user namespace if set to false [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 6430.796641025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support seccomp default which is unconfined [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003545261167572678, "AvgDurationMs": 5970.491720698252, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted with a non-local redirect http liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 244127.4521025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with a /healthz http liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.00357227911407478, "AvgDurationMs": 23681.894051282055, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 3767, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.003981948500132732, "AvgDurationMs": 23375.430928143705, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "ensure 1 worker node at least gets ready", "TotalTestCount": 6822, "TotalFailureCount": 15, "TotalFlakeCount": 0, "FailureRate": 0.0021987686895338612, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 7685.677381546136, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 6603.425660847879, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 6707.790623441396, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 9973.052144638401, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST fail to update a resource due to CRD Validation Rule errors on changed fields", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 2198.27382051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains a x-kubernetes-validations rule that refers to a property that do not exist", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 958.6892269326684, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should delete jobs and pods created by cronjob", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 32420.611970074813, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [FeatureGate:VolumeAttributesClass] should verify ResourceQuota's volume attributes class scope (quota set to pvc count: 1) against 2 pvcs with same volume attributes class.", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 3933.6248974358973, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [FeatureGate:VolumeAttributesClass] should verify ResourceQuota's volume attributes class scope (quota set to pvc count: 1) against a pvc with different volume attributes class.", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 20203.64182051283, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim with a storage class", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 8926.523241895262, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0032771535580524347, "AvgDurationMs": 7856.3725, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should work for CRDs", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 7803.196733167081, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Servers with support for Table transformation should return generic metadata details across all namespaces for nodes", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1051.7207980049873, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0032771535580524347, "AvgDurationMs": 1187.222063106796, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0032771535580524347, "AvgDurationMs": 5640.607233009709, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/json\"", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1057.97753117207, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should be able to schedule after more than 100 missed schedule", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 33405.05713216956, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should remove from active list jobs that have been deleted", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 67455.28087281797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 92998.49189526188, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment Deployment should have a working scale subresource [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 4836.302967581046, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 2376.389775561097, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should not evict unready pods with IfHealthyBudget UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 6476.6460598503745, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should allow to use a pod failure policy to ignore failure matching on DisruptionTarget condition [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 53786.15528678302, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should surface a failure condition on a common issue like exceeded quota", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 3156.6036907730677, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should adopt matching pods on creation [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 4490.24740648379, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 4974.091995012469, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Increasing .start.ordinal", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 33884.10284289275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] NodeAuthenticator The kubelet can delegate ServiceAccount tokens to the API server", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 3731.988154613467, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] SelfSubjectReview testing SSR in different API groups authentication/v1beta1", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1054.7290773067334, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts no secret-based service account token should be auto-generated", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 11003.257506234417, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0032771535580524347, "AvgDurationMs": 1401.1835679611647, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent configmap should exit with the Forbidden error, not a NotFound error", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1031.8913965087288, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends NO DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 11258.493665835413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects NO client request should support a client that connects, sends DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 12089.097605985035, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects a client request should support a client that connects, sends NO DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 11376.818952618452, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 4009.1617206982546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should return command exit codes should support port-forward", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 17260.157182044884, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support exec", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 18983.325561097256, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] kubectl kuberc given preferences should be applied", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1735.2893516209476, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for the cluster [Conformance]", "TotalTestCount": 3739, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003744316662209147, "AvgDurationMs": 8705.322695924766, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should provide DNS for the cluster [Provider:GCE]", "TotalTestCount": 3739, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003744316662209147, "AvgDurationMs": 2423.718213166145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]", "TotalTestCount": 3741, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0037423148890670943, "AvgDurationMs": 15740.07877742947, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide hostIPs as an env var [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 6000.793865336659, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 5597.317206982542, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Lifecycle Sleep Hook when create a pod with lifecycle hook using sleep action valid prestop hook using sleep action", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 57071.53177057358, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] NodeLease NodeLease should have OwnerReferences set", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1114.6876309226936, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, mixed containers - scale up cpu and memory", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 15868.246717948718, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and increase CPU limits", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 12746.77856410256, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and increase memory limits", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 12601.111256410257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests and limits", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 12689.807333333336, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests only", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 12752.76005128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container, one restartable init container - decrease init container CPU only", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 18689.870102564102, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PodRejectionStatus Kubelet should reject pod when the node didn't have enough resource", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 3546.9999501246875, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 244656.06169576064, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should override timeoutGracePeriodSeconds when StartupProbe field is set [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 25051.372568578558, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] RuntimeClass should support RuntimeClasses API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 1847.028079800499, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy nil in SecurityContext when if the container's primary UID belongs to some groups in the image when scheduled node supports SupplementalGroupsPolicy it should add SupplementalGroups to them [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 4217.742205128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers metrics should report count of started and failed user namespaced pods [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 6832.067282051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls [MinimumKubeletVersion:1.21] [Environment:NotInUserNS] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 5737.850149625935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion allow almost all printable ASCII characters as environment variable names", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 5898.3575311720715, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 5892.96970074813, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:Example] Secret should create a pod that reads a secret", "TotalTestCount": 4231, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0033089104230678325, "AvgDurationMs": 6678.870274314214, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with a GRPC liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.003334127173136461, "AvgDurationMs": 75368.65335897438, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for RWO volume with SELinuxMount disabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly]", "TotalTestCount": 3728, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.0037553648068669528, "AvgDurationMs": 71269.65991869918, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions image-registry", "TotalTestCount": 6807, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.00205670633171735, "AvgDurationMs": 0.000060728744939271244, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions kube-apiserver", "TotalTestCount": 6807, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.00205670633171735, "AvgDurationMs": 0.00006072874493927123, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions monitoring", "TotalTestCount": 6807, "TotalFailureCount": 14, "TotalFlakeCount": 0, "FailureRate": 0.00205670633171735, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API priority and fairness should ensure that requests can be classified by adding FlowSchema and PriorityLevelConfiguration", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 3436.535610972569, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API priority and fairness should support PriorityLevelConfiguration API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1989.6425187032419, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7002.8416708229415, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]", "TotalTestCount": 4229, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030740127689761173, "AvgDurationMs": 30902.707231920198, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST NOT ratchet errors raised by transition rules", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 2153.560512820513, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003043071161048689, "AvgDurationMs": 1430.6745145631069, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003043071161048689, "AvgDurationMs": 4204.930097087379, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Discovery should accurately determine present and missing resources", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1496.7140897755612, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should detect duplicates in a CR when preserving unknown fields [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4293.949077306735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with same priority class.", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 2459.0497256857857, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003043071161048689, "AvgDurationMs": 24933.28621359223, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003043071161048689, "AvgDurationMs": 13970.768665048543, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with cross namespace pod affinity scope using scope-selectors.", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 2623.2005985037404, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Servers with support for Table transformation should return chunks of table results for list calls", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1335.6351371571075, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should delete failed finished jobs with limit of one job", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 100726.67406483793, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 8988.648254364089, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment deployment should delete old replica sets [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7864.538827930173, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment should validate Deployment Status endpoints [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4511.712668329174, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment test Deployment ReplicaSet orphaning and adoption regarding controllerRef", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 8033.008802992519, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: enough pods, absolute =\u003e should allow an eviction", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4340.498852867833, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: maxUnavailable allow single eviction, percentage =\u003e should allow an eviction", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5598.604438902744, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should block an eviction until the PDB is updated to allow it [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 8590.731720698255, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 9511.141396508727, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should fail to exceed backoffLimit", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 17307.147082294272, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should recreate pods only after they have failed if pod replacement policy is set to Failed", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 21319.487556109732, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job with successPolicy succeededIndexes rule should succeeded even when some indexes remain pending [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 6987.286533665836, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet Replace and Patch tests [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7326.513366583543, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4741.662668329178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4798.961620947629, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 68037.99214463838, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet MinReadySeconds should be honored when enabled", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 11176.069127182045, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] Certificates API [Privileged:ClusterAdmin] should support building a client with a CSR", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 11659.019351620947, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] SelfSubjectReview testing SSR in different API groups authentication/v1", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1112.192394014962, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1308.831371571072, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] Getting an existing secret should exit with the Forbidden error", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1105.1055112219453, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding Shutdown client connection while the remote stream is writing data to the port-forward connection port-forward should keep working after detect broken connection", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 17129.557830423943, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5890.500523690773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl validation should create/apply an invalid/valid CR with arbitrary-extra properties for CRD with partially-specified validation schema", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 16477.246807980056, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support inline execution and attach", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 57921.36650872819, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client kubectl subresource flag should not be used in a bulk GET", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1016.7157605985037, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] kubectl debug custom profile should be applied on static profiles on ephemeral container", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7825.672443890274, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] kubectl kuberc given preferences should be ignored when flags are explicitly passed", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4819.954812967581, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Metrics should grab all metrics from kubelet /metrics/resource endpoint", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 4881.653915211971, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5540.742019950125, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test when running a container with a new image should not be able to pull from private registry without secret [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 3824.627057356609, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Containers should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5743.378603491271, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Ephemeral Containers [NodeConformance] should update the ephemeral containers in an existing pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7843.294488778052, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7005.607531172071, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Lifecycle Sleep Hook when create a pod with lifecycle hook using sleep action reduce GracePeriodSeconds during runtime", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 37270.79723192021, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] BestEffort pod - try requesting memory, expect error", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 4006.000025641026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container - decrease memory request (RestartContainer memory resize policy)", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 16490.509743589748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU limits only", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 17625.786846153846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests and increase memory limits", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 12789.603769230773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests - decrease memory request", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 9333.698692307693, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests - increase cpu request", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 8980.367230769236, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container, one restartable init container - increase init container CPU \u0026 memory", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 17888.18828205129, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container, one restartable init container - increase init container CPU only", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 18611.63782051282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended (pod generation) [Feature:PodObservedGenerationTracking] [FeatureGate:PodObservedGenerationTracking] [Beta] Pod Generation pod generation should start at 1 and increment per update", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 15249.995641025636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pod Container lifecycle evicted pods should be terminal", "TotalTestCount": 4185, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003106332138590203, "AvgDurationMs": 79883.56755610975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted with a GRPC liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 244333.5473566084, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should *not* be restarted with a non-local redirect http liveness probe", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 244356.33354114712, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 23810.20817955112, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with a GRPC liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 75218.33433915213, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with a failing exec liveness probe that took longer than the timeout", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 53701.8172817955, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 54622.7670074813, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should mark readiness on pods to false and disable liveness probes while pod is in progress of terminating", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 92668.87169576061, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Secrets should patch a secret [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 1252.860748129676, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy was set to Strict in PodSpec when the container's primary UID belongs to some groups in the image when scheduled node does not support SupplementalGroupsPolicy it should reject the pod [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 1512.0431025641026, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsNonRoot should not run with an explicit root user ID [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5242.202069825437, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers must not create the user namespace if set to true [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 5740.903153846154, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers should set FSGroup to user inside the container with hostUsers=false [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 5948.419794871795, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with privileged should run the container as privileged when true [LinuxOnly] [Feature:HostAccess]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 6227.14940149626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support container.SecurityContext.RunAsUser [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 5854.722668329178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 7020.5620947630905, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should succeed in writing subpaths in container [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030725596785629874, "AvgDurationMs": 38536.81506234414, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should *not* be restarted with a GRPC liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 244238.84992307695, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted by liveness probe after startup probe enables it", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 63405.45228205129, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with a exec \"cat /tmp/health\" liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 53495.81020512821, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should have monotonically increasing restart count", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 149068.1322307692, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should mark readiness on pods to false while pod is in progress of terminating when a pod has a readiness probe", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 39713.318461538474, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should override timeoutGracePeriodSeconds when StartupProbe field is set", "TotalTestCount": 4199, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0030959752321981426, "AvgDurationMs": 25372.709666666673, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator][Jira:OLM] OLMv0 should PolarionID:83105-[OTP][Skipped:Disconnected]olmv0 static networkpolicy on ocp", "TotalTestCount": 3759, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.003458366586858207, "AvgDurationMs": 10761.835078534032, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for RWO volume with only SELinuxChangePolicy enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly] [FeatureGate:SELinuxChangePolicy] [Beta]", "TotalTestCount": 3728, "TotalFailureCount": 13, "TotalFlakeCount": 0, "FailureRate": 0.0034871244635193135, "AvgDurationMs": 72543.35524390244, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to create and update mutating webhook configurations with match conditions [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6896.920498753116, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 9797.07982543641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 19532.0593266833, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate everything except 'skip-me' configmaps [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6991.3747381546145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6945.472244389028, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 9164.22922693267, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Discovery should locate the groupVersion and a resource within each APIGroup [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1985.7805985037412, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] OpenAPIV3 should publish OpenAPI V3 for CustomResourceDefinition", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5764.396209476311, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] OpenAPIV3 should round trip OpenAPI V3 for all built-in group versions", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 4082.98957605985, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with different priority class (ScopeSelectorOpNotIn).", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 2252.005436408977, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should apply changes to a resourcequota status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 148712.5746134663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 1144.3606310679613, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 7142.308155339806, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should remove a field if it is owned but removed in the apply request", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1306.1404987531175, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should work for subresources", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1355.4931172069832, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should support ValidatingAdmissionPolicyBinding API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1917.7438653366582, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 11417.48186893204, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] health handlers should contain necessary checks", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1354.2263840399005, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should delete successful finished jobs with limit of one successful job", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 98934.625361596, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should support CronJob API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1784.250773067332, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment deployment should support rollover [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 21488.155411471325, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment iterative rollouts should eventually progress", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 52320.67855361597, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment should not disrupt a cloud load-balancer's connectivity during rollout", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 66217.27850374069, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 11352.627705735662, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 3505.570972568578, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should evict ready pods with AlwaysAllow UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 26902.056758104743, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should evict ready pods with Default UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 27123.978927680797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should evict ready pods with IfHealthyBudget UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 26901.93857855362, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should observe PodDisruptionBudget status updated [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 4663.94386533666, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should allow to use a pod failure policy to ignore failure matching on exit code", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 42006.695112219444, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should delete pods when suspended", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 8285.942443890277, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should manage the lifecycle of a job [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 11418.700972568577, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should update the status ready field", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 7295.811197007481, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 4920.4199251870305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should list and delete a collection of ReplicaSets [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5883.3130423940165, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should get and update a ReplicationController scale [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 2768.6599002493754, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should serve a basic image on each replica with a private image", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1627.8525685785537, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications for partiton1 and delete pod-0 with failing container", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 58105.352668329186, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs with a WhenDeleted policy", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 31491.12837905237, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Setting .start.ordinal", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 34545.75122194515, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] SelfSubjectReview should support SelfSubjectReview API operations", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 923.5689526184539, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should mount projected service account token [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5819.732244389027, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should set ownership and permission when RunAsUser or FsGroup is present [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 19957.961670822944, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ValidatingAdmissionPolicy can restrict access by-node", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5135.550822942641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] Getting a non-existent secret should exit with the Forbidden error, not a NotFound error", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1052.0911471321701, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl apply apply set/view last-applied", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 2908.886733167082, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes control plane services is included in cluster-info [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1189.3789027431421, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl create quota should create a quota with scopes", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1208.8191770573565, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for cronjob", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1708.5983042394018, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl events should show event when pod is created", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6803.960249376556, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 1016.0358980582522, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl logs logs should be able to retrieve and filter logs [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 9960.851820448881, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl rollout undo undo should rollback and update deployment env", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 10485.988104738155, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Events should manage the lifecycle of an event [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1774.3647381546139, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] AppArmor load AppArmor profiles can disable an AppArmor profile, using unconfined", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 915.2956359102245, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6306.826708229427, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 937.6496601941748, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test on terminated container should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5351.509800498753, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5986.138653366582, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 3802.3806982543633, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Lease lease API should be available [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 1543.5899999999997, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] NodeLease NodeLease the kubelet should create and update a lease in the kube-node-lease namespace", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 11236.625536159601, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod with memory requests + limits - decrease memory limit", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 20016.1808974359, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container - decrease CPU (NotRequired) \u0026 memory (RestartContainer)", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 27523.913307692306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease memory requests only", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 17432.770410256413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests and limits", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 12938.195512820512, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, three containers - increase c1 resources, no change for c2, decrease c3 resources (no net change for pod)", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 22272.847410256414, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended (pod generation) [Feature:PodObservedGenerationTracking] [FeatureGate:PodObservedGenerationTracking] [Beta] Pod Generation custom-set generation on new pods and graceful delete", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 4092.8207692307706, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended (pod generation) [Feature:PodObservedGenerationTracking] [FeatureGate:PodObservedGenerationTracking] [Beta] Pod Generation pod observedGeneration field set in pod conditions", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 5825.095282051284, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Delete Grace Period should be submitted and removed", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 9352.96815461347, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should delete a collection of pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 7672.440523690774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should get a host IP [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 3969.24451371571, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should patch a pod status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 3777.699600997507, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should run through the lifecycle of Pods and PodStatus [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6708.050847880301, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted by liveness probe after startup probe enables it", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 63673.73605985038, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5881.2516957605985, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsNonRoot should not run without a specified user ID", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 3536.6760847880305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5670.428653366583, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with readonly rootfs when readOnlyRootFilesystem=true [LinuxOnly] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5723.009775561099, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5607.9353366583555, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5891.4683042394, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5875.388329177058, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context when if the container's primary UID belongs to some groups in the image [LinuxOnly] should add pod.Spec.SecurityContext.SupplementalGroups to them [LinuxOnly] in resultant supplementary groups for the container processes", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5863.758778054862, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls with slashes as separator [MinimumKubeletVersion:1.23] [Environment:NotInUserNS]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 5754.936832917704, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should fail substituting values in a volume subpath with absolute path [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 7675.821770573565, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:Example] Downward API should create a pod that prints his name and namespace", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 6687.360149625935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:Example] Liveness liveness pods should be automatically restarted", "TotalTestCount": 4231, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028362089340581422, "AvgDurationMs": 74027.81346633413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should not be ready with an exec readiness probe timeout [MinimumKubeletVersion:1.20]", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 61039.28928205128, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container with readiness probe should not be ready before initial delay and never restart", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 25871.507307692307, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:KubeletFineGrainedAuthz] [Beta] when calling kubelet API check /healthz enpoint is not accessible via nodes/configz RBAC", "TotalTestCount": 4199, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028578232912598238, "AvgDurationMs": 4739.721461538461, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 12, "TotalFlakeCount": 0, "FailureRate": 0.0028089887640449437, "AvgDurationMs": 8718.992111650487, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 7222.3986034912705, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6716.3552119700735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AggregatedDiscovery should support aggregated discovery interface [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1048.391970074813, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST NOT fail to update a resource due to JSONSchema errors on unchanged correlatable fields", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 2108.6288974358968, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST fail to update a resource due to CRD Validation Rule errors on unchanged uncorrelatable fields", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 2116.998358974359, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.0025749063670411983, "AvgDurationMs": 1024.0376699029127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceFieldSelectors [Privileged:ClusterAdmin] CustomResourceFieldSelectors MUST list and watch custom resources matching the field selector [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 9799.139127182047, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should create/apply a valid CR for CRD with validation schema [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 4411.9034663341645, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should detect unknown metadata fields of a typed object [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1096.9028927680797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.0025749063670411983, "AvgDurationMs": 2063.352766990292, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should support cascading deletion of custom resources", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 25851.62960099751, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should manage the lifecycle of a ResourceQuota [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1192.7465586034914, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 24105.57937655861, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should type check validation expressions", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1358.9044638403993, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should validate against a Deployment [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 2177.327581047382, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.0025749063670411983, "AvgDurationMs": 21358.79609223301, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should set the cronjob-scheduled-timestamp annotation on a job", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 32270.61523690774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment deployment reaping should cascade to its replica sets and pods", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5225.190448877805, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should evict unready pods with AlwaysAllow UnhealthyPodEvictionPolicy", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6424.62902743142, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should allow to delegate reconciliation to external controller", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 4006.6370324189515, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should delete a job [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 7162.444164588527, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should execute all indexes despite some failing when using backoffLimitPerIndex [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 26279.05461346633, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should fail when exceeds active deadline", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6220.941795511222, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should not create pods when created in suspend state", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 11799.83054862843, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should record the failure-count in the Pod annotation when using backoffLimitPerIndex", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 17598.161870324187, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should terminate job execution when the number of failed indexes exceeds maxFailedIndexes [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6555.319875311721, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job with successPolicy should succeeded when all indexes succeeded [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6957.179351620946, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should validate Replicaset Status endpoints [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5084.239975062344, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6806.77620947631, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should validate Statefulset Status endpoints [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 22165.305586034905, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 35764.69234413966, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should create a serviceAccountToken and ensure a successful TokenReview [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1056.5982044887785, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] SubjectReview should support SubjectReview API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1075.610598503741, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] A node shouldn't be able to delete another node", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1066.3802493765584, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects NO client request should support a client that connects, sends DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 11910.337506234417, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on localhost that expects a client request should support a client that connects, sends DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 12173.948154613465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl apply should reuse port when apply to an existing SVC", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 2143.9980049875303, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod Kubectl run running a failing command", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 20138.73591022443, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support exec using resource/name", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 17305.621346633423, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl logs default container logs the second container is the default-container by annotation should log default container if not specified", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 11585.352693266834, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified in annotations", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 937.5152618453866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified on the pod", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 917.1594763092271, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5706.18685785536, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 10743.375211970073, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5866.252144638404, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 46101.62950124689, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet when scheduling an agnhost Pod with hostAliases should write entries to /etc/hosts [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5682.190174563592, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 9827.434513715709, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] BestEffort QoS pod - empty resize", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 6733.947487179487, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests only", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 12674.20307692308, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, two containers with cpu \u0026 memory requests + limits - reorder containers", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 5649.508128205127, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PodOSRejection [NodeConformance] Kubelet [LinuxOnly] should reject pod when the node OS doesn't match pod's OS", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 2706.4493516209473, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 1236.2197256857855, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pod Container Status should never report container start when an init container fails", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 64965.99167082295, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should be submitted and removed [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6564.148603491273, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 8486.384139650872, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PreStop should call prestop when killing a pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 11797.591022443889, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be ready immediately after startupProbe succeeds", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 37743.479625935164, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with a local redirect http liveness probe", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 23708.540648379047, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should mark readiness on pods to false while pod is in progress of terminating when a pod has a readiness probe", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 39686.70867830423, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Secrets should be consumable as environment variable names variable names with various prefixes [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5862.706184538654, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 6016.846483790526, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Secrets should fail to create secret due to empty secret key [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 996.5325187032421, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy was set to Merge in PodSpec when the container's primary UID belongs to some groups in the image when scheduled node supports SupplementalGroupsPolicy it should add SupplementalGroups to them [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 4222.007230769232, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy was set to Strict in PodSpec when the container's primary UID belongs to some groups in the image when scheduled node supports SupplementalGroupsPolicy it should NOT add SupplementalGroups to them [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 4090.338384615385, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5638.785685785534, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support seccomp runtime/default [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 5733.989700748131, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when true [LinuxOnly] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002599858189553297, "AvgDurationMs": 7017.1239152119715, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with a local redirect http liveness probe", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 23698.287256410247, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with an exec liveness probe with timeout [MinimumKubeletVersion:1.20]", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 54626.96187179487, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should mark readiness on pods to false and disable liveness probes while pod is in progress of terminating", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 92879.15353846153, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart https hook properly [MinimumKubeletVersion:1.23]", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 11521.346461538466, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:KubeletFineGrainedAuthz] [Beta] when calling kubelet API check /healthz enpoint is accessible via nodes/healthz RBAC", "TotalTestCount": 4199, "TotalFailureCount": 11, "TotalFlakeCount": 0, "FailureRate": 0.002619671350321505, "AvgDurationMs": 4901.730307692307, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API priority and fairness should ensure that requests can't be drowned out (priority)", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 918.1963092269327, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 7490.222892768081, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 7084.215561097257, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should reject mutating webhook configurations with invalid match conditions [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 6755.479027431421, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AggregatedDiscovery should support aggregated discovery interface for CRDs [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4054.2334663341653, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AggregatedDiscovery should support raw aggregated discovery request for CRDs [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4046.917880299252, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CRDValidationRatcheting [Privileged:ClusterAdmin] [FeatureGate:CRDValidationRatcheting] MUST NOT fail to update a resource due to CRD Validation Rule errors on unchanged correlatable fields", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 2179.8898974358976, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 9481.696408977554, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023408239700374533, "AvgDurationMs": 5501.7030582524285, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains an x-kubernetes-validations rule that exceeds the estimated cost limit", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 963.6887281795508, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3524.5499251870315, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should create/apply an invalid CR with extra properties for CRD with validation schema [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4242.822394014963, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should detect unknown and duplicate fields of a typed object [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1105.5773067331672, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should orphan pods created by rc if deleteOptions.OrphanDependents is nil", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 33222.49087281798, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Generated clientset should create v1 cronJobs, delete cronJobs, watch cronJobs", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1192.3322942643392, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's multiple priority class scope (quota set to pod count: 2) against 2 pods with same priority classes.", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2558.579077306732, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against a pod with different priority class (ScopeSelectorOpExists).", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2486.2266832917703, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a persistent volume claim", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 8711.092443890271, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope using scope-selectors.", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3529.7740149625934, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Server request timeout default timeout should be used if the specified timeout in the request URL is 0s", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 955.249825436409, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should ignore conflict errors if force apply is used", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1437.8190523690769, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023408239700374533, "AvgDurationMs": 963.1814805825243, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Servers with support for Table transformation should return pod details", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1035.9252867830426, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should allow expressions to refer variables. [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2216.12463840399, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should type check a CRD", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2229.1055112219447, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023408239700374533, "AvgDurationMs": 1227.0389805825246, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] server version should find the server version [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 937.2356359102246, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should not emit unexpected warnings", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 94354.61279301746, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] CronJob should support timezone", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 980.5316708229426, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4328.308254364089, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Deployment deployment should support proportional scaling [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 9844.852743142143, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: enough pods, replicaSet, percentage =\u003e should allow an eviction", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5402.224588528678, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: no PDB =\u003e should allow an eviction", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3897.827306733168, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: too few pods, absolute =\u003e should not allow an eviction", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3714.7842892768076, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController should observe that the PodDisruptionBudget status is not updated for unmanaged pods", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 64899.172418952636, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should apply changes to a job status [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4421.157730673317, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should create pods with completion indexes for an Indexed Job", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 11209.028877805487, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should remove pods when job is deleted", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 6777.936783042394, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicationController should release no longer matching pods [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023408239700374533, "AvgDurationMs": 3802.5321601941737, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 22526.027381546133, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should implement legacy replacement when the update strategy is OnDelete", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 50894.07743142146, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Removing .start.ordinal", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 32787.21975062345, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2619.942094763092, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] ServiceAccounts should update a ServiceAccount [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1038.4393017456352, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] Getting a secret for a workload the node has access to should succeed", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3154.788952618453, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] Getting an existing configmap should exit with the Forbidden error", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1243.673216957606, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 that expects a client request should support a client that connects, sends DATA, and disconnects", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 11998.479152119702, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023408239700374533, "AvgDurationMs": 1217.7976941747575, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl apply should apply a new configuration to an existing RC", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1792.460099750623, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl copy should copy a file from a running Pod", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5531.048578553615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl create quota should create a quota without scopes", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1212.1482294264345, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 6461.829800498752, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl get componentstatuses should get componentstatuses", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1849.4773815461347, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3838.1867082294266, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod Kubectl run running a successful command", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 20439.86693266833, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] kubectl debug custom profile should be applied on static profiles while copying from pod", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 10323.172793017457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-etcd][apigroup:config.openshift.io][OCPFeatureGate:DualReplica] Two Node with Fencing should have etcd pods and containers configured correctly [Suite:openshift/conformance/parallel]", "TotalTestCount": 485, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.020618556701030927, "AvgDurationMs": 125.91034722222227, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-etcd][apigroup:config.openshift.io][OCPFeatureGate:DualReplica] Two Node with Fencing should have podman etcd containers running on each node [Suite:openshift/conformance/parallel]", "TotalTestCount": 485, "TotalFailureCount": 10, "TotalFlakeCount": 1, "FailureRate": 0.020618556701030927, "AvgDurationMs": 3268.071805555556, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster shouldn't report any alerts in firing state apart from Watchdog and AlertmanagerReceiversNotConfigured [Early][apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4735, "TotalFailureCount": 10, "TotalFlakeCount": 4, "FailureRate": 0.0021119324181626186, "AvgDurationMs": 3915.286115942029, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should be consumable as environment variable names with various prefixes [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5915.0405735660825, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1251.7639650872816, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 10545.234114713216, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test on terminated container should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5226.188703241895, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test on terminated container should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5089.663142144638, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5125.3773815461345, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test when running a container with a new image should not be able to pull image from invalid registry [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4846.912319201994, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 29060.4089276808, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5667.847082294263, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 6160.570897755612, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5982.276184538654, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Events should be sent by kubelets and the scheduler about pods scheduling and running ", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 7840.743441396509, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5267.466608478804, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 3793.883815461347, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1062.1078054862846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5252.852543640899, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Lifecycle Sleep Hook when create a pod with lifecycle hook using sleep action ignore terminated container", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 8552.899201995013, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU requests and decrease CPU limits", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 13414.127641025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container, one restartable init container - increase init container memory only", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 20608.152666666665, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended (pod generation) [Feature:PodObservedGenerationTracking] [FeatureGate:PodObservedGenerationTracking] [Beta] Pod Generation pod rejected by kubelet should have updated generation and observedGeneration", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 3598.3736153846157, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pod Container Status should never report success for a pending container", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 64966.12800498754, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pod TerminationGracePeriodSeconds is negative pod with negative grace period", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1228.9969326683288, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should be updated [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 4414.40633416459, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should support pod readiness gates [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 16046.060872817958, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PrivilegedPod [NodeConformance] should enable privileged commands [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5187.748628428926, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should override timeoutGracePeriodSeconds when LivenessProbe field is set [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 25104.675361596008, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 25334.52713216957, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 1424.9936408977558, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy nil in SecurityContext when if the container's primary UID belongs to some groups in the image when scheduled node does not support SupplementalGroupsPolicy it should add SupplementalGroups to them [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 3770.2867435897433, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context SupplementalGroupsPolicy [LinuxOnly] [Feature:SupplementalGroupsPolicy] [FeatureGate:SupplementalGroupsPolicy] [Beta] when SupplementalGroupsPolicy was set to Merge in PodSpec when the container's primary UID belongs to some groups in the image when scheduled node does not support SupplementalGroupsPolicy it should add SupplementalGroups to them [LinuxOnly]", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 3917.780384615384, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsNonRoot should run with an image specified user ID", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 7898.9132917705765, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsUser should run the container with uid 0 [LinuxOnly] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5678.472967581047, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers must create the user namespace in the configured hostUID/hostGID range [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 904.4568717948721, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5831.297082294265, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context when creating containers with AllowPrivilegeEscalation should allow privilege escalation when not explicitly set and uid != 0 [LinuxOnly] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 6998.893840399002, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should not launch unsafe, but not explicitly enabled sysctls on the node [MinimumKubeletVersion:1.21]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2901.75620947631, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] User Namespaces for Pod Security Standards [LinuxOnly] with UserNamespacesSupport and UserNamespacesPodSecurityStandards enabled should allow pod [Feature:UserNamespacesPodSecurityStandards] [FeatureGate:UserNamespacesSupport] [Beta] [FeatureGate:UserNamespacesPodSecurityStandards] [Alpha] [Feature:OffByDefault]", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 5964.4868205128205, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5801.195586034912, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 5903.3068827930165, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should fail substituting values in a volume subpath with backticks [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 7478.647306733166, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports external claim referenced by multiple pods", "TotalTestCount": 2954, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.003385240352064997, "AvgDurationMs": 49141.61887640449, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted startup probe fails", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 73583.46623076923, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be restarted with a failing exec liveness probe that took longer than the timeout", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 53454.74592307691, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should override timeoutGracePeriodSeconds when LivenessProbe field is set", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 25078.588282051278, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 11268.740820512823, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 11415.876743589743, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop https hook properly [MinimumKubeletVersion:1.23]", "TotalTestCount": 4199, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.0023815194093831865, "AvgDurationMs": 11441.406410256412, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io][OCPFeatureGate:DualReplica] Two Node with Fencing topology should have BareMetalHost operational status set to detached if they exist [Suite:openshift/conformance/parallel]", "TotalTestCount": 485, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.020618556701030927, "AvgDurationMs": 126.64416666666668, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io][OCPFeatureGate:DualReplica] Two Node with Fencing topology should have infrastructure platform type set correctly [Suite:openshift/conformance/parallel]", "TotalTestCount": 485, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.020618556701030927, "AvgDurationMs": 125.9715972222222, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node][apigroup:config.openshift.io][OCPFeatureGate:DualReplica] Two Node with Fencing topology should only have two control plane nodes and no arbiter nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 485, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.020618556701030927, "AvgDurationMs": 158.60006944444447, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] LimitRange should list, patch and delete a LimitRange by collection [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 10, "TotalFlakeCount": 0, "FailureRate": 0.002363507445048452, "AvgDurationMs": 2019.026259351621, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to create and update validating webhook configurations with match conditions [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 6874.126184538655, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 7500.355486284286, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AggregatedDiscovery should support raw aggregated discovery endpoint Accept headers [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1059.6460349127183, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource that exceeds the runtime cost limit for x-kubernetes-validations rule execution", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 4866.233117206982, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] FieldValidation should detect unknown metadata fields in both the root and embedded object of a CR [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 4303.3235411471305, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against 2 pods with different priority class.", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 2172.7332418952615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota [Feature:PodPriority] should verify ResourceQuota's priority class scope (quota set to pod count: 1) against 2 pods with same priority class.", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 2457.8165835411464, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002106741573033708, "AvgDurationMs": 7718.309271844659, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002106741573033708, "AvgDurationMs": 7944.811796116503, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should give up ownership of a field if forced applied by a controller", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 2768.1571571072313, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Servers with support for API chunking should return chunks of results for list calls [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 22410.24735660848, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/vnd.kubernetes.protobuf,application/json\"", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1027.4063840399, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should allow to use the pod failure policy on exit code to fail the job early [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 7049.179476309226, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should create pods for an Indexed job with completion indexes and specified hostname [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 11344.657880299252, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job with successPolicy succeededCount rule should succeeded even when some indexes remain pending [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 6622.252892768078, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] ReplicaSet should serve a basic image on each replica with a private image", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1647.3435910224441, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 20042.562743142145, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs after adopting pod (WhenDeleted)", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 30957.616807980048, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Non-retain StatefulSetPersistentVolumeClaimPolicy should delete PVCs after adopting pod (WhenScaled)", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 46485.482394014965, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Scaling StatefulSetStartOrdinal Decreasing .start.ordinal", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 32135.44645885287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding with a pod being removed should stop port-forwarding", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 12518.245361596015, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl create quota should reject quota with invalid scopes", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1018.7289526184538, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 8954.801970074808, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 4875.3112967581055, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should contain last line of the log", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 29028.156433915214, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should return command exit codes execing into a container with a failing command", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 17419.3210723192, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client kubectl subresource flag GET on status subresource of built-in type (node) returns identical info as GET on the built-in type", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1486.7807481296759, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl logs all pod logs the Deployment has 2 replicas and each pod has 2 containers should get logs from all pods based on default container", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5339.7695261845365, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl logs all pod logs the Deployment has 2 replicas and each pod has 2 containers should get logs from each pod and each container in Deployment", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5826.002044887781, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Events API should delete a collection of events [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1236.327630922693, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1891.7129426433917, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Events should delete a collection of events [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1234.5686284289277, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-instrumentation] MetricsGrabber should grab all metrics slis from API server.", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 976.5054364089774, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] Networking should provide Internet connection for containers [Feature:Networking-IPv4]", "TotalTestCount": 2755, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.003266787658802178, "AvgDurationMs": 5540.774932735428, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] ConfigMap should update ConfigMap successfully", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1025.2708229426435, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 10960.537481296762, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Runtime blackbox test when running a container with a new image should be able to pull image [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5296.24523690773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide host IP and pod IP as an env var if pod uses host network [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5294.230897755612, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Ephemeral Containers [NodeConformance] will start an ephemeral container in an existing pod [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 7626.599401496262, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase CPU limits only", "TotalTestCount": 4199, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.0021433674684448677, "AvgDurationMs": 12873.925871794872, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PodTemplates should replace a pod template [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 1046.3036408977553, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pod Container lifecycle should not create extra sandbox if all containers are done", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 8956.337231920199, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance]", "TotalTestCount": 3741, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.0024057738572574178, "AvgDurationMs": 3743.7395611285265, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PreStop graceful pod terminated should wait until preStop hook completes the process", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 24378.45645885287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 53655.05189526185, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a container with runAsNonRoot should run with an explicit non-root user ID [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5637.796658354115, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context When creating a pod with HostUsers should mount all volumes with proper permissions with hostUsers=false [LinuxOnly] [Feature:UserNamespacesSupport] [FeatureGate:UserNamespacesSupport] [Beta]", "TotalTestCount": 4199, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.0021433674684448677, "AvgDurationMs": 6126.920179487179, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support pod.Spec.SecurityContext.SupplementalGroups [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5857.81867830424, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support seccomp unconfined on the container [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5911.3712468827935, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Security Context should support seccomp unconfined on the pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.002127156700543607, "AvgDurationMs": 5889.024887780547, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] with multiple drivers using drapbv1beta1 and drapbv1 work", "TotalTestCount": 2954, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.003046716316858497, "AvgDurationMs": 93799.75760299628, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Probing restartable init container should be ready immediately after startupProbe succeeds", "TotalTestCount": 4199, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.0021433674684448677, "AvgDurationMs": 38335.63425641025, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [FeatureGate:KubeletFineGrainedAuthz] [Beta] when calling kubelet API check /healthz enpoint is accessible via nodes/proxy RBAC", "TotalTestCount": 4199, "TotalFailureCount": 9, "TotalFlakeCount": 0, "FailureRate": 0.0021433674684448677, "AvgDurationMs": 4664.898846153846, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API priority and fairness should ensure that requests can't be drowned out (fairness)", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 902.6952119700753, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 18161.705461346628, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018726591760299626, "AvgDurationMs": 6308.997548543689, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Generated clientset should create pods, set the deletionTimestamp and deletionGracePeriodSeconds of the pod", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 4190.515087281797, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018726591760299626, "AvgDurationMs": 7630.657354368932, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Server request timeout should return HTTP status code 400 if the user specifies an invalid timeout in the request URL", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 959.4538403990025, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Server request timeout the request should be served with a default timeout if the specified timeout in the request URL exceeds maximum allowed", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 953.477481296758, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should create an applied object if it does not already exist", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1448.5443640897759, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ServerSideApply should not remove a field if an owner unsets the field but other managers still have ownership of the field", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1441.1792768079804, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ValidatingAdmissionPolicy [Privileged:ClusterAdmin] should support ValidatingAdmissionPolicy API operations [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 2055.94770573566, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] client-go should negotiate watch and report errors with accept \"application/vnd.kubernetes.protobuf\"", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1076.226259351621, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] Job should run a job to completion when tasks succeed", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 11313.05770573566, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][OCPFeatureGate:Example] should only run FeatureGated test when enabled [Suite:openshift/conformance/parallel]", "TotalTestCount": 471, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.016985138004246284, "AvgDurationMs": 1.9631250000000005, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-auth] NodeAuthenticator The kubelet's main port 10250 should reject requests with no credentials", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 3756.010897755612, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth] [Feature:NodeAuthorizer] A node shouldn't be able to create another node", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1134.6436408977552, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 9780.366608478804, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl validation should detect unknown metadata fields of a typed object", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1254.5028179551125, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018726591760299626, "AvgDurationMs": 1084.3676213592232, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl delete interactive based on user confirmation input", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 7616.226284289278, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressIP][apigroup:operator.openshift.io] [external-targets][apigroup:user.openshift.io][apigroup:security.openshift.io] only pods matched by the pod selector should have the EgressIPs [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 683, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.01171303074670571, "AvgDurationMs": 35460.46459183674, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressIP][apigroup:operator.openshift.io] [external-targets][apigroup:user.openshift.io][apigroup:security.openshift.io] pods should have the assigned EgressIPs and EgressIPs can be deleted and recreated [Skipped:azure][apigroup:route.openshift.io] [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 616, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.012987012987012988, "AvgDurationMs": 42581.27140350876, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressIP][apigroup:operator.openshift.io] [external-targets][apigroup:user.openshift.io][apigroup:security.openshift.io] pods should have the assigned EgressIPs and EgressIPs can be updated [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 681, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.011747430249632892, "AvgDurationMs": 37496.14556122449, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][Feature:EgressIP][apigroup:operator.openshift.io] [external-targets][apigroup:user.openshift.io][apigroup:security.openshift.io] pods should keep the assigned EgressIPs when being rescheduled to another node [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 683, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.01171303074670571, "AvgDurationMs": 34233.42908163265, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] AppArmor load AppArmor profiles should enforce an AppArmor profile specified on the container", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 925.7363092269328, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 10293.789576059851, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 3948.549276807981, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Kubelet with pods in a privileged namespace when scheduling an agnhost Pod with hostAliases and hostNetwork should write entries to /etc/hosts when hostNetwork is enabled [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 4776.84234413965, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, mixed containers - add limits", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 17494.76335897436, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests and limits", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 13174.498102564105, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - decrease CPU requests only", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 17354.93384615385, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pod InPlace Resize Container [FeatureGate:InPlacePodVerticalScaling] [Beta] Burstable QoS pod, one container with cpu \u0026 memory requests + limits - increase memory requests and decrease CPU limits", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 13087.713769230768, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] PodTemplates should delete a collection of pod templates [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1302.3888778054863, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods Extended Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 1020.4158104738159, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 8430.035660847876, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container should be restarted startup probe fails", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 73519.87820448876, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0018908059560387616, "AvgDurationMs": 60998.95224438902, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane must be possible for the driver to update the ResourceClaim.Status.Devices once allocated [FeatureGate:DRAResourceClaimDeviceStatus] [Beta]", "TotalTestCount": 2954, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.002708192281651997, "AvgDurationMs": 12056.691011235956, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 11813.926666666663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [Feature:SidecarContainers] [FeatureGate:SidecarContainers] Restartable Init Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly", "TotalTestCount": 4199, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0019052155275065491, "AvgDurationMs": 11632.484076923081, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not unstage RWO volume when starting a second pod with the same SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.016985138004246284, "AvgDurationMs": 76744.17097222223, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "verify operator conditions ingress", "TotalTestCount": 6807, "TotalFailureCount": 8, "TotalFlakeCount": 0, "FailureRate": 0.0011752607609813426, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should reject validating webhook configurations with invalid match conditions [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 6533.886957605984, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CustomResourceValidationRules [Privileged:ClusterAdmin] MUST fail create of a custom resource definition that contains an x-kubernetes-validations rule that contains a syntax error", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 953.4600748129676, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016385767790262173, "AvgDurationMs": 11169.190800970875, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should list, patch and delete a collection of StatefulSets [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 22505.89930174564, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Early] Operators low level operators should have at least the conditions we had in 4.17 [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.001334604385128694, "AvgDurationMs": 1096.535085995086, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-architecture] Conformance Tests should have at least two untainted nodes [Conformance]", "TotalTestCount": 3781, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0018513620735255223, "AvgDurationMs": 982.6321022727269, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:SCC][Early] should not have pod creation failures during install [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.001334604385128694, "AvgDurationMs": 450.94149877149874, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]", "TotalTestCount": 4272, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016385767790262173, "AvgDurationMs": 1111.617718446602, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should return command exit codes execing into a container with a successful command", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 17402.60137157107, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support exec through kubectl proxy", "TotalTestCount": 3741, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0018711574445335472, "AvgDurationMs": 17313.614545454544, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 10935.3585286783, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][Feature:Machines][Early] Managed cluster should have same number of Machines and Nodes [apigroup:machine.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 4768, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0014681208053691276, "AvgDurationMs": 485.4947246376811, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Network Policy Audit logging] when using openshift ovn-kubernetes should ensure acl logs are created and correct [apigroup:project.openshift.io][apigroup:network.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 486, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.01440329218106996, "AvgDurationMs": 13629.782848101264, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:DNSNameResolver][Feature:EgressFirewall] when using openshift ovn-kubernetes should ensure egressfirewall with wildcard dns rules is created [Suite:openshift/conformance/parallel]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 78, "FailureRate": 0.014861995753715499, "AvgDurationMs": 23269.506250000013, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop https hook properly [MinimumKubeletVersion:1.23] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 10972.901571072322, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Containers should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 5800.484413965087, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 5913.032668329178, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Sysctls [LinuxOnly] [NodeConformance] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 963.5493516209474, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.0016544552115339162, "AvgDurationMs": 5891.285112219452, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane supports count/resourceclaims.resource.k8s.io ResourceQuota [ConformanceCandidate]", "TotalTestCount": 2954, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.002369668246445498, "AvgDurationMs": 8749.574906367037, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] [FeatureGate:DRAPrioritizedList] [Beta] filters config correctly for multiple devices", "TotalTestCount": 2954, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.002369668246445498, "AvgDurationMs": 86788.20142322096, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMWebhookProviderOpenshiftServiceCA] OLMv1 operator with webhooks should clean up webhooks when the extension is uninstalled [Serial]", "TotalTestCount": 699, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.010014306151645207, "AvgDurationMs": 47259.8923857868, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The HAProxy router pods [apigroup:route.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 7, "TotalFlakeCount": 1, "FailureRate": 0.0015283842794759825, "AvgDurationMs": 3079.736604361371, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should not pass SELinux mount option for RWO volume with SELinuxMount disabled and Recursive policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 70959.58756944441, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should pass SELinux mount option for RWO volume with SELinuxMount enabled and nil policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 71846.64256944442, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWO volume when starting a second pod with different policy (MountOption -\u003e Recursive) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 90817.71847222222, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] MutableCSINodeAllocatableCount [FeatureGate:MutableCSINodeAllocatableCount] [Beta] [Feature:OffByDefault] Attach Limit Exceeded should transition pod to failed state when attachment limit exceeded", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 195811.2624305555, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] MutableCSINodeAllocatableCount [FeatureGate:MutableCSINodeAllocatableCount] [Beta] [Feature:OffByDefault] Dynamic Allocatable Count should observe dynamic changes in CSINode allocatable count", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 65988.45833333334, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 30141.19736111111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 50970.96423611111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 81052.75902777778, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 1376.176666666666, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 96901.17152777777, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 52749.06861111112, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 1390.1097916666663, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 63295.946527777785, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes", "TotalTestCount": 378, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.018518518518518517, "AvgDurationMs": 50966.391832061076, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 48881.45027777778, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 40401.71145833334, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 40844.54826388889, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 97534.9167361111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 1389.600486111111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 25538.345347222225, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 101812.7707638889, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 73346.4146527778, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 38432.48951388888, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 49811.16541666668, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 1375.2878472222224, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 37729.20451388888, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 38406.43194444444, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 99898.19374999998, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 7, "TotalFlakeCount": 0, "FailureRate": 0.014861995753715499, "AvgDurationMs": 100919.2015972222, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-arch][Early] APIs for openshift.io must have stable versions [Suite:openshift/conformance/parallel]", "TotalTestCount": 5181, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0011580775911986102, "AvgDurationMs": 3038.209747474747, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-arch][Early] CRDs for openshift.io should have subresource.status [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001143946615824595, "AvgDurationMs": 3062.863906633906, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-ci] [Early] prow job name should match cluster version [apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5057, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0011864741941862765, "AvgDurationMs": 2709.4146796657387, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-ci] [Early] prow job name should match security mode [Suite:openshift/conformance/parallel]", "TotalTestCount": 5181, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0011580775911986102, "AvgDurationMs": 2862.300101010101, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on 0.0.0.0 should support forwarding over websockets", "TotalTestCount": 3741, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0016038492381716118, "AvgDurationMs": 8393.984608150467, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl Port forwarding With a server listening on localhost should support forwarding over websockets", "TotalTestCount": 3741, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0016038492381716118, "AvgDurationMs": 8573.214733542321, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][Feature:Machines] Managed cluster should [sig-scheduling][Early] control plane machine set operator should not cause an early rollout [Suite:openshift/conformance/parallel]", "TotalTestCount": 5238, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001145475372279496, "AvgDurationMs": 255.4955528255528, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][Feature:Machines] Managed cluster should [sig-scheduling][Early] control plane machine set operator should not have any events [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001143946615824595, "AvgDurationMs": 174.38916461916466, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-etcd] etcd cluster has the same number of master nodes and voting members from the endpoints configmap [Early][apigroup:config.openshift.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5045, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0011892963330029733, "AvgDurationMs": 282.80142061281333, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-etcd] etcd record the start revision of the etcd-operator [Early] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001143946615824595, "AvgDurationMs": 5.97120393120393, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-kubevirt] migration when running openshift cluster on KubeVirt virtual machines and live migrate hosted control plane workers [Early] should maintain node readiness [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001143946615824595, "AvgDurationMs": 2808.39995085995, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to preserve UDP traffic when server pod cycles for a LoadBalancer service on the same nodes", "TotalTestCount": 1574, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0038119440914866584, "AvgDurationMs": 58924.53513761467, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart https hook properly [MinimumKubeletVersion:1.23] [NodeConformance]", "TotalTestCount": 4231, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0014181044670290711, "AvgDurationMs": 10358.528778054866, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Managed cluster record the number of nodes at the beginning of the tests [Early] [Suite:openshift/conformance/parallel]", "TotalTestCount": 5245, "TotalFailureCount": 6, "TotalFlakeCount": 1, "FailureRate": 0.001143946615824595, "AvgDurationMs": 2895.4239557739565, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance]", "TotalTestCount": 4231, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0014181044670290711, "AvgDurationMs": 1393.6803491271814, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] [FeatureGate:DRAPrioritizedList] [Beta] selects the first subrequest that can be satisfied", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 87044.17359550564, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must map configs and devices to the right containers", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 26464.882771535584, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must retry NodePrepareResources", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 91049.76752808987, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node retries pod scheduling after updating device class", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 49295.62842696629, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports claim and class parameters", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 47687.98082397004, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports simple pod referencing inline resource claim", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 47070.309101123596, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] rolling update [KubeletMinVersion:1.33]", "TotalTestCount": 2954, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.002031144211238998, "AvgDurationMs": 56472.79958801499, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {ClusterRoleBindings} are not specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 7943.886785714289, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {ConfigMap:resourceNames} are not all specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 11174.743071428573, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {clusterextension/finalizer} is not specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 7842.489, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {create} verb is not specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 6380.364785714285, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {escalate, bind} is not specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 5648.427714285715, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMPreflightPermissionChecks][Skipped:Disconnected] OLMv1 operator preflight checks should report error when {services} are not specified", "TotalTestCount": 377, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.015915119363395226, "AvgDurationMs": 14437.109642857142, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 openshift-certified-operators Catalog should serve FBC via the /v1/api/all endpoint", "TotalTestCount": 3611, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001661589587371919, "AvgDurationMs": 275.97142011834313, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator][Jira:OLM] OLMv0 should PolarionID:83583-[OTP][Skipped:Disconnected]olmv0 networkpolicy on hosted hypershift", "TotalTestCount": 3759, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0015961691939345571, "AvgDurationMs": 8391.086099476443, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-apiserver pods [apigroup:authorization.openshift.io][apigroup:build.openshift.io][apigroup:image.openshift.io][apigroup:project.openshift.io][apigroup:quota.openshift.io][apigroup:route.openshift.io][apigroup:security.openshift.io][apigroup:template.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4579, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001310329766324525, "AvgDurationMs": 3103.7928348909663, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-authentication pods [apigroup:oauth.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0013100436681222707, "AvgDurationMs": 3072.4659501557626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-console console pods [apigroup:console.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4579, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001310329766324525, "AvgDurationMs": 3084.1170716510906, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-console downloads pods [apigroup:console.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4579, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001310329766324525, "AvgDurationMs": 3070.1214641744546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-etcd pods [apigroup:operator.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0013100436681222707, "AvgDurationMs": 3065.627165109035, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-image-registry pods [apigroup:imageregistry.operator.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4577, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0013109023377758358, "AvgDurationMs": 3101.7827102803735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-monitoring prometheus-adapter pods [apigroup:monitoring.coreos.com] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0013100436681222707, "AvgDurationMs": 3093.211308411214, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-monitoring thanos-querier pods [apigroup:monitoring.coreos.com] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 6, "TotalFlakeCount": 1, "FailureRate": 0.0013100436681222707, "AvgDurationMs": 3125.8909034267904, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-oauth-apiserver pods [apigroup:oauth.openshift.io][apigroup:user.openshift.io] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4580, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.0013100436681222707, "AvgDurationMs": 3105.838785046729, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling][Early] The openshift-operator-lifecycle-manager pods [apigroup:packages.operators.coreos.com] should be scheduled on different nodes [Suite:openshift/conformance/parallel]", "TotalTestCount": 4579, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.001310329766324525, "AvgDurationMs": 3116.877975077881, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWO volume when starting a second pod with different SELinux context [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 91972.63166666668, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 63151.47625, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 43226.304166666676, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 1171.6706944444443, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 101695.82333333335, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 1328.6057638888892, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 81095.33131944448, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 36868.238888888874, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 79566.5207638889, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.012738853503184714, "AvgDurationMs": 39648.237430555564, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "verify operator conditions openshift-samples", "TotalTestCount": 6806, "TotalFailureCount": 6, "TotalFlakeCount": 0, "FailureRate": 0.000881575080811049, "AvgDurationMs": 0.000060728744939271244, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a ResourceClaim [FeatureGate:DynamicResourceAllocation] [DRA]", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 7720.548014981273, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should surge pods onto nodes when spec was updated and update strategy is RollingUpdate", "TotalTestCount": 740, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.006756756756756757, "AvgDurationMs": 93395.1401369863, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-ci] [Early] prow job name should match network type [Suite:openshift/conformance/parallel]", "TotalTestCount": 5199, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0009617234083477592, "AvgDurationMs": 2840.628083538083, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-ci] [Early] prow job name should match platform type [Suite:openshift/conformance/parallel]", "TotalTestCount": 5199, "TotalFailureCount": 5, "TotalFlakeCount": 1, "FailureRate": 0.0009617234083477592, "AvgDurationMs": 2827.134471744472, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network] LoadBalancers [Feature:LoadBalancer] should be able to preserve UDP traffic when server pod cycles for a LoadBalancer service on different nodes", "TotalTestCount": 1574, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0031766200762388818, "AvgDurationMs": 55847.64715596331, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] retries pod scheduling after creating device class", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 14933.533707865166, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports simple pod referencing inline resource claim", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 13048.038764044946, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] [FeatureGate:DRAPrioritizedList] [Beta] chooses the correct subrequest subject to constraints", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 85480.37636704122, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] deletes generated claims when pod is done", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 47707.743558052425, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must unprepare resources for force-deleted pod", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 18188.87943820225, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on multiple nodes with different ResourceSlices keeps pod pending because of CEL runtime errors", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 40904.7611610487, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node must deallocate after use", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 47402.55288389513, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports init containers", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 48510.95153558052, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports reusing resources", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 72708.6565917603, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports simple pod referencing external resource claim", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 47369.77176029964, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] runs pod after driver starts", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 52331.03486891386, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] with multiple drivers using only drapbv1 [KubeletMinVersion:1.34] work", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 95421.67977528088, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] with multiple drivers using only drapbv1beta1 work", "TotalTestCount": 2954, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.0016926201760324984, "AvgDurationMs": 94083.35719101124, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 4, "FailureRate": 0.010615711252653927, "AvgDurationMs": 191086.46902777776, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 970.4493749999998, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 972.4880555555555, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 973.5374305555557, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 1126.3252083333332, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 41267.6917361111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 43102.05354166666, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 81475.32736111115, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 970.8643055555556, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 65987.03812499999, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should block a second pod from using an in-use ReadWriteOncePod volume on the same node", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 47760.255208333336, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 39824.40583333334, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 40135.62083333332, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 36301.09993055555, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 77408.89805555555, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 37550.59180555554, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 5, "TotalFlakeCount": 0, "FailureRate": 0.010615711252653927, "AvgDurationMs": 963.6709027777775, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API data in etcd should be stored at the correct location and version for all resources [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 710, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.005633802816901409, "AvgDurationMs": 18094.452451923084, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Simple pod should support exec through an HTTP proxy", "TotalTestCount": 3741, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0010692328254477412, "AvgDurationMs": 17746.493291536062, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-network][Feature:Router][apigroup:route.openshift.io] when FIPS is disabled the HAProxy router should serve routes when configured with a 1024-bit RSA key [Suite:openshift/conformance/parallel]", "TotalTestCount": 193, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.02072538860103627, "AvgDurationMs": 25205.81744186047, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]", "TotalTestCount": 3741, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0010692328254477412, "AvgDurationMs": 3798.8129153605028, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] retries pod scheduling after updating device class", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 14025.353483146067, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] runs a pod without a generated resource claim", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 15695.423183520597, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports external claim referenced by multiple containers of multiple pods", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 13018.557565543073, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports external claim referenced by multiple pods", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 12201.595880149813, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports init containers", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 13208.140898876403, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports reusing resources", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 25627.415692883904, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports sharing a claim concurrently", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 13138.53835205993, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports simple pod referencing external resource claim", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 12359.721460674155, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] with different ResourceSlices keeps pod pending because of CEL runtime errors", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 11550.721797752809, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane must apply per-node permission checks [ConformanceCandidate]", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 11066.55797752809, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane truncates the name of a generated resource claim [ConformanceCandidate]", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 11771.958089887641, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane validate ResourceClaimTemplate and ResourceClaim for admin access [FeatureGate:DRAAdminAccess] [Beta]", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 7344.363782771536, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] [FeatureGate:DRAPrioritizedList] [Beta] uses the config for the selected subrequest", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 85628.74213483145, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] does not delete generated claims when pod is restarting", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 67443.14295880149, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] failed update [KubeletMinVersion:1.33]", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 57058.994681647935, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must call NodePrepareResources even if not used by any container", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 18208.865955056186, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must manage ResourceSlices", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 62247.96288389515, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports inline claim referenced by multiple containers", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 49179.55509363296, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports sharing a claim concurrently", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 48779.09018726593, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] registers plugin", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 10888.855131086148, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] supports init containers with external claims", "TotalTestCount": 2954, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0013540961408259986, "AvgDurationMs": 51271.901011235954, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should pass SELinux mount option for RWO volume with SELinuxMount enabled and MountOption policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 75445.60555555556, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount SELinuxMount [LinuxOnly] [Feature:SELinux] should unstage RWO volume when starting a second pod with different policy (Recursive -\u003e MountOption) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 91925.18694444443, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 30015.015763888892, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 30330.70631944444, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 35722.8646527778, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 52043.13840277777, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 925.0177083333335, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 65720.90270833333, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41620.85958333332, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 37451.62909722223, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 56310.37270833333, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 51953.114236111105, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 51800.00923611112, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 937.0954166666668, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41917.10194444445, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 59591.99236111111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41609.25791666667, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41839.259027777785, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41268.82118055556, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41788.45284722223, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 41266.956805555565, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 55463.339513888895, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 920.116527777778, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 911.5744444444446, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 922.1456944444445, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 952.2736111111113, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 37180.90361111111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 50030.48104166667, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 76297.59305555555, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 51303.735902777764, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 471, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.008492569002123142, "AvgDurationMs": 926.7220138888889, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "verify operator conditions control-plane-machine-set", "TotalTestCount": 6807, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0005876303804906713, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions node-tuning", "TotalTestCount": 6806, "TotalFailureCount": 4, "TotalFlakeCount": 0, "FailureRate": 0.0005877167205406994, "AvgDurationMs": 0.0008097165991902831, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] MutatingAdmissionPolicy [Privileged:ClusterAdmin] [Feature:MutatingAdmissionPolicy] [FeatureGate:MutatingAdmissionPolicy] [Beta] [Feature:OffByDefault] should support MutatingAdmissionPolicy API operations", "TotalTestCount": 471, "TotalFailureCount": 3, "TotalFlakeCount": 1, "FailureRate": 0.006369426751592357, "AvgDurationMs": 1966.6025694444445, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-arch][Early] Managed cluster should [apigroup:config.openshift.io] start all core operators [Suite:openshift/conformance/parallel]", "TotalTestCount": 4800, "TotalFailureCount": 3, "TotalFlakeCount": 2, "FailureRate": 0.000625, "AvgDurationMs": 255.36890756302526, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports claim and class parameters", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 13368.794307116103, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] supports inline claim referenced by multiple containers", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 12711.291610486891, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] blocks new pod after force-delete [KubeletMinVersion:1.34]", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 98788.93910112356, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] must not run a pod if a claim is not ready", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 32362.787265917606, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on multiple nodes with node-local resources uses all resources", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 47672.616441947575, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node retries pod scheduling after creating device class", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 48787.32348314606, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node supports external claim referenced by multiple containers of multiple pods", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 51130.53943820226, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] removes reservation from claim when pod is done", "TotalTestCount": 2954, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.001015572105619499, "AvgDurationMs": 47567.70749063671, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1] OLMv1 should pass a trivial sanity check", "TotalTestCount": 4231, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.0007090522335145356, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 openshift-redhat-operators Catalog should serve FBC via the /v1/api/all endpoint", "TotalTestCount": 3611, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.0008307947936859596, "AvgDurationMs": 259.83331360946755, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions cloud-credential", "TotalTestCount": 6807, "TotalFailureCount": 3, "TotalFlakeCount": 0, "FailureRate": 0.00044072278536800354, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-api-machinery] MutatingAdmissionPolicy [Privileged:ClusterAdmin] [Feature:MutatingAdmissionPolicy] [FeatureGate:MutatingAdmissionPolicy] [Beta] [Feature:OffByDefault] should mutate a Deployment", "TotalTestCount": 471, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.004246284501061571, "AvgDurationMs": 31280.996388888892, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] MutatingAdmissionPolicy [Privileged:ClusterAdmin] [Feature:MutatingAdmissionPolicy] [FeatureGate:MutatingAdmissionPolicy] [Beta] [Feature:OffByDefault] should support MutatingAdmissionPolicyBinding API operations", "TotalTestCount": 471, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.004246284501061571, "AvgDurationMs": 1876.4959027777775, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:LDAP][Serial] ldap group sync can sync groups from ldap [apigroup:user.openshift.io][apigroup:authorization.openshift.io][apigroup:security.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0028208744710860366, "AvgDurationMs": 59347.605748792295, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream admission TestImageStreamTagsAdmission [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0028208744710860366, "AvgDurationMs": 17529.291400966187, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Serial] Image signature workflow can push a signed image to openshift registry and verify it [apigroup:user.openshift.io][apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 629, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.003179650238473768, "AvgDurationMs": 47576.30025906736, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] [Serial]Should have MCN properties matching associated node properties for nodes in custom MCPs [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.002828854314002829, "AvgDurationMs": 67883.572705314, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should be able to handle large requests: http", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 24797.697631578947, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should be able to handle large requests: udp", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 24529.35473684211, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for client IP based session affinity: http [LinuxOnly]", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 42168.06052631578, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for client IP based session affinity: udp [LinuxOnly]", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 43875.85157894737, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for endpoint-Service: http", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 35938.98263157895, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for endpoint-Service: udp", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 34952.37052631579, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for node-Service: http", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 36171.19236842105, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for node-Service: udp", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 44526.347368421026, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for pod-Service: http", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 37932.915789473685, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for pod-Service: udp", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 37392.50710526317, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should function for service endpoints using hostNetwork", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 73492.51236842106, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should update endpoints: http", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 93401.46210526317, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] Granular Checks: Services Secondary IP Family [LinuxOnly] should update endpoints: udp", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 97653.8847368421, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should be able to reach pod on ipv4 and ipv6 ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 129007.15315789476, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create a single stack service with cluster ip from primary service range", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 7654.783157894736, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create pod, add ipv6 and ipv4 ip to host ips", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 6398.956052631578, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create pod, add ipv6 and ipv4 ip to pod ips", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 6492.058157894736, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create service with ipv4 cluster ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 7235.10342105263, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create service with ipv4,v6 cluster ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 7334.142368421052, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create service with ipv6 cluster ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 6710.630789473683, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should create service with ipv6,v4 cluster ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 7116.415789473685, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] [Feature:IPv6DualStack] should have ipv4 and ipv6 internal node ip", "TotalTestCount": 117, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.017094017094017096, "AvgDurationMs": 1262.432105263158, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] must deallocate after use", "TotalTestCount": 2954, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0006770480704129993, "AvgDurationMs": 11864.617602996257, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] control plane [ConformanceCandidate] with node-local resources uses all resources", "TotalTestCount": 2954, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0006770480704129993, "AvgDurationMs": 14977.94617977528, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] DaemonSet with admin access [FeatureGate:DRAAdminAccess] [Beta]", "TotalTestCount": 2954, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0006770480704129993, "AvgDurationMs": 18319.760337078656, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [DRA] kubelet [Feature:DynamicResourceAllocation] on single node runs a pod without a generated resource claim", "TotalTestCount": 2954, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0006770480704129993, "AvgDurationMs": 45963.78374531835, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 openshift-community-operators Catalog should serve FBC via the /v1/api/all endpoint", "TotalTestCount": 3611, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.0005538631957906397, "AvgDurationMs": 282.6452662721893, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions openshift-controller-manager", "TotalTestCount": 6807, "TotalFailureCount": 2, "TotalFlakeCount": 0, "FailureRate": 0.00029381519024533566, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Conformance][sig-sno][Serial] Cluster should allow a fast rollout of kube-apiserver with no pods restarts during API disruption [apigroup:config.openshift.io][apigroup:operator.openshift.io] [Suite:openshift/conformance/serial/minimal]", "TotalTestCount": 707, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014144271570014145, "AvgDurationMs": 9705.138309178748, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Serial] [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013513513513513514, "AvgDurationMs": 4192.411780821917, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Serial] [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013477088948787063, "AvgDurationMs": 12050.821826484018, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Serial] [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013477088948787063, "AvgDurationMs": 42357.606301369844, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should always delete fast (ALL of 100 namespaces in 150 seconds) [Feature:ComprehensiveNamespaceDraining]", "TotalTestCount": 742, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013477088948787063, "AvgDurationMs": 77352.24968036529, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should delete fast enough (90 percent of 100 namespaces in 150 seconds)", "TotalTestCount": 740, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013513513513513514, "AvgDurationMs": 74895.54849315068, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-console [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.001445086705202312, "AvgDurationMs": 3723.574305555556, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-operator-lifecycle-manager [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 690, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014492753623188406, "AvgDurationMs": 3770.1511574074057, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] oc project can switch between different projects [apigroup:authorization.openshift.io][apigroup:user.openshift.io][apigroup:project.openshift.io][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 25239.346473429956, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiDisk][platform:vsphere][Disruptive] Managed cluster should create machinesets with eagerly zeroed data disk [apigroup:machine.openshift.io][Serial][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.007518796992481203, "AvgDurationMs": 168816.4607142857, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream API TestImageStreamWithoutDockerImageConfig [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 2073.0309178743964, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ChunkSizeMiB][Serial][apigroup:imageregistry.operator.openshift.io] Image Registry Config ChunkSizeMiB should reject ChunkSizeMiB value greater than 5 GiB [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 2748.674541062802, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] NoExecuteTaintManager Single Pod [Serial] eventually evict pod with finite tolerations from tainted nodes", "TotalTestCount": 651, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0015360983102918587, "AvgDurationMs": 132103.47139784944, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMOwnSingleNamespace] OLMv1 operator installation support for ownNamespace and single namespace watch mode with operator should install cluster extensions successfully in both watch modes", "TotalTestCount": 471, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0021231422505307855, "AvgDurationMs": 255078.78895833337, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMWebhookProviderOpenshiftServiceCA] OLMv1 operator with webhooks should be tolerant to tls secret deletion [Serial]", "TotalTestCount": 699, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.001430615164520744, "AvgDurationMs": 57120.234010152286, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMWebhookProviderOpenshiftServiceCA] OLMv1 operator with webhooks should have a working conversion webhook [Serial]", "TotalTestCount": 699, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.001430615164520744, "AvgDurationMs": 63154.23441624366, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMWebhookProviderOpenshiftServiceCA] OLMv1 operator with webhooks should have a working mutating webhook [Serial]", "TotalTestCount": 699, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.001430615164520744, "AvgDurationMs": 57625.02335025382, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 Catalogs should be installed", "TotalTestCount": 3611, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00027693159789531985, "AvgDurationMs": 243.3135798816568, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLM][Skipped:Disconnected] OLMv1 openshift-redhat-marketplace Catalog should serve FBC via the /v1/api/all endpoint", "TotalTestCount": 3611, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00027693159789531985, "AvgDurationMs": 259.75269230769226, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-operator][Jira:OLM] OLMv0 should pass a trivial sanity check", "TotalTestCount": 4231, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002363507445048452, "AvgDurationMs": 0.06234413965087282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] when PVC has node-affinity to non-existent/illegal nodes, the pod should be scheduled normally if suitable nodes exist", "TotalTestCount": 740, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013513513513513514, "AvgDurationMs": 13779.06922374429, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002363507445048452, "AvgDurationMs": 7814.516009975063, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002363507445048452, "AvgDurationMs": 9514.670798004987, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002363507445048452, "AvgDurationMs": 7732.413765586036, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] baseline namespace should allow pods with inline volumes when the driver uses the restricted label [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0013477088948787063, "AvgDurationMs": 2130.5774429223743, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:DisableStorageClass][Serial][apigroup:operator.openshift.io] should not reconcile the StorageClass when StorageClassState is Unmanaged [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 10145.020531400964, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:DisableStorageClass][Serial][apigroup:operator.openshift.io] should remove the StorageClass when StorageClassState is Removed [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 11970.637487922706, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][FeatureGate:VSphereDriverConfiguration][Serial][apigroup:operator.openshift.io] vSphere CSI Driver Configuration snapshot options in clusterCSIDriver should use default when unset [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0014104372355430183, "AvgDurationMs": 10636.736666666668, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with invalid namespace labels on should fail to create namespace with invalid fsgroup label", "TotalTestCount": 4083, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002449179524859172, "AvgDurationMs": 93.25504201680671, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage][OCPFeatureGate:StoragePerformantSecurityPolicy] Storage Performant Policy with invalid namespace labels on should fail to create namespace with invalid selinux label", "TotalTestCount": 4083, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.0002449179524859172, "AvgDurationMs": 105.01263305322126, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions dns", "TotalTestCount": 6807, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00014690759512266783, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions kube-scheduler", "TotalTestCount": 6807, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00014690759512266783, "AvgDurationMs": 0.000020242914979757085, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions machine-config", "TotalTestCount": 6807, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00014690759512266783, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions operator-lifecycle-manager-packageserver", "TotalTestCount": 6806, "TotalFailureCount": 1, "TotalFlakeCount": 0, "FailureRate": 0.00014692918013517486, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "Cluster should be stable after installation is complete", "TotalTestCount": 5538, "TotalFailureCount": 0, "TotalFlakeCount": 137, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "Cluster should be stable before test is started", "TotalTestCount": 5538, "TotalFailureCount": 0, "TotalFlakeCount": 38, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:\"Cluster Version Operator\"] cluster-version-operator-tests should support passing tests", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.003466334164588529, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:Monitoring][sig-instrumentation] sanity test should always pass [Suite:openshift/cluster-monitoring-operator/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.02493765586034913, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:config-operator][sig-api-machinery] sanity test should always pass [Suite:openshift/cluster-config-operator/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.08728179551122195, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:oauth-apiserver][sig-api-machinery] sanity test should always pass [Suite:openshift/oauth-apiserver/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:openshift-apiserver][sig-api-machinery] sanity test should always pass [Suite:openshift/cluster-openshift-apiserver-operator/conformance/parallel]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.08491271820448876, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[Jira:openshift-apiserver][sig-api-machinery] sanity test should always pass [Suite:openshift/openshift-apiserver/conformance/parallel]", "TotalTestCount": 629, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-12", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Jira:openshift-controller-manager][sig-openshift-controller-manager] sanity test should always pass [Suite:openshift/cluster-openshift-controller-manager-operator/conformance/parallel]", "TotalTestCount": 2288, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-12", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Jira:openshift-controller-manager][sig-openshift-controller-manager] sanity test should always pass [Suite:openshift/openshift-controller-manager/conformance/parallel]", "TotalTestCount": 2318, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-12", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Jira:storage-version-migrator][sig-api-machinery] sanity test should always pass [Suite:openshift/cluster-kube-storage-version-migrator-operator/conformance/parallel]", "TotalTestCount": 1459, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.033898305084745776, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-12", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Serial] [sig-auth][Feature:OAuthServer] [RequestHeaders] [IdP] test RequestHeaders IdP [apigroup:config.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 156977.11985507247, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Serial][sig-cli] oc adm upgrade recommend When the update service has no recommendations runs successfully [Suite:openshift/conformance/serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 19243.05082191781, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Serial][sig-cli] oc adm upgrade recommend runs successfully with an empty channel [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1111.1924657534248, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[Serial][sig-cli] oc adm upgrade recommend runs successfully, even without upstream OpenShift Update Service customization [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 710.6288127853882, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] reflector doesn't support receiving resources as Tables", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10815.838125, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] server supports sending resources in Table format", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 971.1767307692305, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] should NOT be requested by client-go's List method when WatchListClient is enabled", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 815.3604807692307, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] should NOT be requested by dynamic client's List method when WatchListClient is enabled", "TotalTestCount": 708, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 847.5459134615386, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] should NOT be requested by metadata client's List method when WatchListClient is enabled", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 964.6697596153846, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] should be requested by informers when WatchListClient is enabled", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 979.4959134615384, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] API Streaming (aka. WatchList) [FeatureGate:WatchList] [Beta] [Serial] should be requested by metadatainformer when WatchListClient is enabled", "TotalTestCount": 708, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1025.6292788461535, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] CBOR [Feature:CBOR] clients remain compatible with the 1.17 sample-apiserver [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 25743.24849315068, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should apply a finalizer to a Namespace [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1279.119771689497, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should apply an update to a Namespace [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 719.6282191780822, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should apply changes to a namespace status [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 783.4828767123289, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 16060.856347031959, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 8869.126712328769, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1247.9331506849314, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery] OpenAPIV3 should contain OpenAPI V3 for Aggregated APIServer [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 26380.035707762563, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-cloud-credential-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 690, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3694.5803703703714, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-cluster-version [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3714.5875925925925, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-console-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3670.64050925926, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-controller-manager-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 690, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5716.755601851853, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-dns [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3693.219027777777, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-etcd [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3703.8258333333333, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-etcd-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 690, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3603.4134722222216, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-image-registry [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3671.7857407407405, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-insights [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3656.424861111111, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-kube-apiserver [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 688, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3706.2088425925917, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-kube-apiserver-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3635.6831944444434, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-kube-scheduler [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6394.217824074074, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-kube-scheduler-operator [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 692, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3676.4912962962962, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ReadOnlyRootFilesystem] Explicitly set readOnlyRootFilesystem to true - openshift-marketplace [OCP-83088][Skipped:Disconnected][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 690, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 20426.2924537037, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-api-machinery][Feature:ResourceQuota] Object count should properly count the number of persistentvolumeclaims resources [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4025.3552054794527, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] ControllerRevision [Serial] should manage the lifecycle of a ControllerRevision [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7017.633287671234, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should list and delete a collection of DaemonSets [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3756.164200913242, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should not update pod when spec was updated and update strategy is OnDelete", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6332.461004566209, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9277.15579908676, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]", "TotalTestCount": 649, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 11952.000107526886, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9982.40698630137, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should run and stop complex daemon with node affinity", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6165.657351598173, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10503.603835616439, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 13429.489315068491, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Daemon set [Serial] should verify changes to a daemon set status [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6243.5723744292245, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: maxUnavailable deny evictions, integer =\u003e should not allow an eviction [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2980.6321461187217, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] DisruptionController evictions: too few pods, replicaSet, percentage =\u003e should not allow an eviction [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3048.5273972602745, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-apps] Job should run a job to completion with CPU requests [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 45633.66484018265, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:OpenShiftAuthorization][Serial] authorization TestAuthorizationResourceAccessReview should succeed [apigroup:authorization.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10427.881545893717, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-auth][Feature:ProjectAPI][Serial] TestUnprivilegedNewProjectDenied [apigroup:authorization.openshift.io][apigroup:project.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3403.951304347827, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl taint [Serial] should remove all the taints with the same key off a node", "TotalTestCount": 651, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7158.468924731182, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] Kubectl client Kubectl taint [Serial] should update the taint on a node", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4843.916027397259, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] oc adm cluster-role-reapers [Serial][apigroup:authorization.openshift.io][apigroup:user.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 14542.709082125604, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cli] oc status can show correct status after switching between projects [apigroup:project.openshift.io][apigroup:image.openshift.io][Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 20415.689710144936, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiDisk][platform:vsphere][Disruptive] Managed cluster should create machines with data disks [apigroup:machine.openshift.io][Serial][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 135807.80821428573, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiDisk][platform:vsphere][Disruptive] Managed cluster should create machinesets with a data disk using each provisioning mode [apigroup:machine.openshift.io][Serial][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 167964.76714285713, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiDisk][platform:vsphere][Disruptive] Managed cluster should create machinesets with thick data disk [apigroup:machine.openshift.io][Serial][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 164317.63035714283, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiDisk][platform:vsphere][Disruptive] Managed cluster should create machinesets with thin data disk [apigroup:machine.openshift.io][Serial][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 159876.23285714284, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiNetworks][platform:vsphere] Managed cluster should machine network should be correlated with node networking [apigroup:machine.openshift.io][Suite:openshift/conformance/parallel]", "TotalTestCount": 243, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 164.8392857142857, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiNetworks][platform:vsphere] Managed cluster should machines should have all specified portgroup associated with their failure domain [apigroup:machine.openshift.io][Suite:openshift/conformance/parallel]", "TotalTestCount": 243, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 147.12761904761908, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiNetworks][platform:vsphere] Managed cluster should new machines should pass multi network tests [Serial][apigroup:machine.openshift.io][Suite:openshift/conformance/serial]", "TotalTestCount": 133, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 238518.4525, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiNetworks][platform:vsphere] Managed cluster should node VMs should have all specified portgroups attached which are associated with their failure domain [apigroup:machine.openshift.io][Suite:openshift/conformance/parallel]", "TotalTestCount": 243, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2394.7071428571426, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-cluster-lifecycle][OCPFeatureGate:VSphereMultiNetworks][platform:vsphere] Managed cluster should node addresses should be correlated with the machine network [apigroup:machine.openshift.io][Suite:openshift/conformance/parallel]", "TotalTestCount": 243, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 167.83142857142855, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-etcd][OCPFeatureGate:HardwareSpeed][Serial] etcd is able to set the hardware speed to \"\" [Timeout:30m][apigroup:machine.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 699, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 236949.64568527916, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-etcd][OCPFeatureGate:HardwareSpeed][Serial] etcd is able to set the hardware speed to Slower [Timeout:30m][apigroup:machine.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 699, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 777765.1861928933, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-etcd][OCPFeatureGate:HardwareSpeed][Serial] etcd is able to set the hardware speed to Standard [Timeout:30m][apigroup:machine.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 699, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 331364.69284263963, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial] clusterimagepolicy signature validation tests fail with PKI email does not match [Suite:openshift/conformance/serial]", "TotalTestCount": 189, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 246639.61304347825, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial] clusterimagepolicy signature validation tests fail with PKI root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 189, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 246610.1447826087, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial] clusterimagepolicy signature validation tests pass with valid PKI [Suite:openshift/conformance/serial]", "TotalTestCount": 189, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 246967.39101449278, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial] imagepolicy signature validation tests fail with PKI root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 191, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 247700.24130434782, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial] imagepolicy signature validation tests pass with valid PKI [Suite:openshift/conformance/serial]", "TotalTestCount": 191, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 248914.25115942024, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial][Skipped:Disconnected] clusterimagepolicy signature validation tests fail with PKI email does not match [Suite:openshift/conformance/serial]", "TotalTestCount": 244, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113612.02102941177, "PeriodStart": "2025-12-09", "PeriodEnd": "2025-12-13", "DaysWithData": 5 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial][Skipped:Disconnected] clusterimagepolicy signature validation tests fail with PKI root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 244, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 114037.63352941177, "PeriodStart": "2025-12-09", "PeriodEnd": "2025-12-13", "DaysWithData": 5 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial][Skipped:Disconnected] clusterimagepolicy signature validation tests pass with valid PKI [Suite:openshift/conformance/serial]", "TotalTestCount": 241, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 114367.81838235294, "PeriodStart": "2025-12-09", "PeriodEnd": "2025-12-13", "DaysWithData": 5 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial][Skipped:Disconnected] imagepolicy signature validation tests fail with PKI root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 244, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 116000.98029411765, "PeriodStart": "2025-12-09", "PeriodEnd": "2025-12-13", "DaysWithData": 5 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerificationPKI][Serial][Skipped:Disconnected] imagepolicy signature validation tests pass with valid PKI [Suite:openshift/conformance/serial]", "TotalTestCount": 244, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113924.87735294117, "PeriodStart": "2025-12-09", "PeriodEnd": "2025-12-13", "DaysWithData": 5 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerification][Serial] Should fail clusterimagepolicy signature validation root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113488.19260869561, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerification][Serial] Should fail clusterimagepolicy signature validation when scope in allowedRegistries list does not skip signature verification [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 221992.69579710145, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerification][Serial] Should fail imagepolicy signature validation in different namespaces root of trust does not match the identity in the signature [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113368.62231884057, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerification][Serial] Should pass clusterimagepolicy signature validation with signed image [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113734.10106280191, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imagepolicy][OCPFeatureGate:SigstoreImageVerification][Serial] Should pass imagepolicy signature validation with signed image in namespaces [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 113410.62483091788, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream API TestImageStreamMappingCreate [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3709.2522705314013, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream API TestImageStreamTagLifecycleHook [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2247.123768115941, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream admission TestImageStreamAdmitSpecUpdate [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17411.292270531412, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][Feature:ImageTriggers][Serial] ImageStream admission TestImageStreamAdmitStatusUpdate [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 12387.359903381643, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ChunkSizeMiB][Serial][apigroup:imageregistry.operator.openshift.io] Image Registry Config ChunkSizeMiB should not accept invalid ChunkSizeMiB value [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1989.926956521739, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ChunkSizeMiB][Serial][apigroup:imageregistry.operator.openshift.io] Image Registry Config ChunkSizeMiB should set ChunkSizeMiB value [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1982.81806763285, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ChunkSizeMiB][Serial][apigroup:imageregistry.operator.openshift.io] Image Registry Config ChunkSizeMiB should set maximum valid ChunkSizeMiB value [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1981.525410628019, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ChunkSizeMiB][Serial][apigroup:imageregistry.operator.openshift.io] Image Registry Config ChunkSizeMiB should set minimum valid ChunkSizeMiB value [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2009.256231884058, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-imageregistry][OCPFeatureGate:ImageStreamImportMode][Serial] ImageStream API import mode should be PreserveOriginal or Legacy depending on desired.architecture field in the CV [apigroup:image.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2032.0521739130431, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation] Prometheus [apigroup:image.openshift.io] when installed on the cluster should report telemetry [Serial] [Late] [Suite:openshift/conformance/serial]", "TotalTestCount": 882, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2325.3600980392157, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][Late] Platform Prometheus targets should not be accessible without auth [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 977, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 278796.4281278538, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][OCPFeatureGate:MetricsCollectionProfiles][Serial] The collection profiles feature-set in a heterogeneous environment, should expose information about the applied collection profile using meta-metrics [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 21374.11415458938, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][OCPFeatureGate:MetricsCollectionProfiles][Serial] The collection profiles feature-set in a heterogeneous environment, should have at least one implementation for each collection profile [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1322.9714009661839, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][OCPFeatureGate:MetricsCollectionProfiles][Serial] The collection profiles feature-set in a heterogeneous environment, should revert to default collection profile when an empty collection profile value is specified [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4630.136714975846, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][OCPFeatureGate:MetricsCollectionProfiles][Serial] The collection profiles feature-set in a homogeneous minimal environment, should hide default metrics [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 86185.96666666667, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-instrumentation][OCPFeatureGate:MetricsCollectionProfiles][Serial] The collection profiles feature-set initially, in a homogeneous default environment, should expose default metrics [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4507.360579710146, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] [Serial]Should properly transition through MCN conditions on rebootless node update [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 47869.918840579696, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:MachineConfigNodes] [Serial]Should properly update the MCN from the associated MCD [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1522.542222222222, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImagesAWS][Serial] Should degrade on a MachineSet with an OwnerReference [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 12361.710144927534, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImagesAWS][Serial] Should not update boot images on any MachineSet when not configured [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 21414.870821256038, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImagesAWS][Serial] Should stamp coreos-bootimages configmap with current MCO hash and release version [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 13670.656376811594, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImagesAWS][Serial] Should update boot images on all MachineSets when configured [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 23945.130048309187, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImagesAWS][Serial] Should update boot images only on MachineSets that are opted in [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 23736.96294685991, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImages][Serial] Should degrade on a MachineSet with an OwnerReference [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 11801.907198067634, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImages][Serial] Should not update boot images on any MachineSet when not configured [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17305.309468599033, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImages][Serial] Should stamp coreos-bootimages configmap with current MCO hash and release version [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 13807.507777777777, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImages][Serial] Should update boot images on all MachineSets when configured [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17770.186570048307, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-mco][OCPFeatureGate:ManagedBootImages][Serial] Should update boot images only on MachineSets that are opted in [apigroup:machineconfiguration.openshift.io] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 18184.79492753623, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling [apigroup:apps.openshift.io][apigroup:route.openshift.io] should handle many TCP connections by possibly dropping those over a certain bound [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 23400.577294685998, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling [apigroup:apps.openshift.io][apigroup:route.openshift.io] should handle many UDP senders (by continuing to drop all packets on the floor) [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 35751.616231884065, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling with Deployments [apigroup:route.openshift.io] should handle many TCP connections by possibly dropping those over a certain bound [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 21276.8520289855, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network-edge][Feature:Idling] Unidling with Deployments [apigroup:route.openshift.io] should handle many UDP senders (by continuing to drop all packets on the floor) [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33365.67695652174, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] IngressClass [Feature:Ingress] should allow IngressClass to have Namespace-scoped parameters [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 752.6487671232878, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] IngressClass [Feature:Ingress] should choose the one with the later CreationTimestamp, if equal the one with the lower name when two ingressClasses are marked as default [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1840.4705022831047, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] IngressClass [Feature:Ingress] should not set default value if no default IngressClass [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1819.9171689497716, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] IngressClass [Feature:Ingress] should set default value on new IngressClass [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1790.7075342465753, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] Networking should allow creating a Pod with an SCTP HostPort [LinuxOnly] [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7840.719589041097, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] Service endpoints latency should not be very high [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5995.059908675801, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] Services should allow creating a basic SCTP service with pod and endpoints [LinuxOnly] [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 12601.650684931506, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] services when running openshift ipv4 cluster ensures external ip policy is configured correctly on the cluster [apigroup:config.openshift.io] [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 993599.6641545892, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network] services when running openshift ipv4 cluster on bare metal [apigroup:config.openshift.io] ensures external auto assign cidr is configured correctly on the cluster [apigroup:config.openshift.io] [Serial] [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 103598.33449275364, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][Feature:MultiNetworkPolicy][Serial][apigroup:operator.openshift.io] should enforce a network policies on secondary network IPv4 [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3470.9184541062805, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][Feature:MultiNetworkPolicy][Serial][apigroup:operator.openshift.io] should enforce a network policies on secondary network IPv6 [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6066.6548309178725, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should be enabled by default [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 14124.433816425124, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should function without any target pods [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 34111.9320289855, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should move the source diagnostics pods based on the new selector and tolerations [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 30532.36787439614, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should move the target diagnostics pods based on the new selector and tolerations [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32095.44869565217, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should remove all network diagnostics pods when disabled [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 36813.37381642511, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-network][OCPFeatureGate:NetworkDiagnosticsConfig][Serial] Should set the condition to false if there are no nodes able to host the source pods [Suite:openshift/conformance/serial]", "TotalTestCount": 707, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 30058.523768115938, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] NoExecuteTaintManager Multiple Pods [Serial] only evicts pods without tolerations from tainted nodes", "TotalTestCount": 651, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 126229.68000000004, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] NoExecuteTaintManager Single Pod [Serial] doesn't evict pod with tolerations from tainted nodes", "TotalTestCount": 651, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 126243.79715053762, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] NoExecuteTaintManager Single Pod [Serial] evicts pods from tainted nodes", "TotalTestCount": 649, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 64128.74043010753, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] NoExecuteTaintManager Single Pod [Serial] pods evicted from tainted nodes have pod disruption condition", "TotalTestCount": 651, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 63355.39258064515, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Burstable QoS pod with container resources", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17631.653990384617, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Burstable QoS pod, 1 container with resources", "TotalTestCount": 708, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17493.97615384616, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Burstable QoS pod, no container resources", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17643.9973076923, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Guaranteed QoS pod with container resources", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 20662.032884615397, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Guaranteed QoS pod, 1 container with resources", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17839.407019230766, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] Pod Level Resources [Serial] [Feature:PodLevelResources] [FeatureGate:PodLevelResources] [Beta] Guaranteed QoS pod, no container resources", "TotalTestCount": 708, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17762.655625000003, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] [Serial] Pod InPlace Resize Container (scheduler-focused) [FeatureGate:InPlacePodVerticalScaling] [Beta] pod-resize-scheduler-tests", "TotalTestCount": 708, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 13921.426971153847, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node] kubelet Clean up pods on node kubelet should be able to delete 10 pods per node in 1m0s.", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 16183.60598173516, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-node][DRA][OCPFeatureGate:DynamicResourceAllocation] Dynamic Resource Allocation should verify beta and alpha DRA APIs are disabled [apigroup:resource.k8s.io] [Suite:openshift/conformance/parallel]", "TotalTestCount": 427, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2914.968148148148, "PeriodStart": "2025-12-12", "PeriodEnd": "2025-12-13", "DaysWithData": 2 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMCatalogdAPIV1Metas][Skipped:Disconnected] OLMv1 openshift-certified-operators Catalog should serve FBC via the /v1/api/metas endpoint", "TotalTestCount": 377, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 256.47535714285715, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMCatalogdAPIV1Metas][Skipped:Disconnected] OLMv1 openshift-community-operators Catalog should serve FBC via the /v1/api/metas endpoint", "TotalTestCount": 377, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 275.0167857142858, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMCatalogdAPIV1Metas][Skipped:Disconnected] OLMv1 openshift-redhat-marketplace Catalog should serve FBC via the /v1/api/metas endpoint", "TotalTestCount": 377, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 259.7510000000001, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMCatalogdAPIV1Metas][Skipped:Disconnected] OLMv1 openshift-redhat-operators Catalog should serve FBC via the /v1/api/metas endpoint", "TotalTestCount": 377, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 274.5704285714286, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMOwnSingleNamespace] OLMv1 operator installation should reject invalid watch namespace configuration and update the status conditions accordingly should fail to install the ClusterExtension when watch namespace is invalid", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 67441.88937500001, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMOwnSingleNamespace] OLMv1 operator installation support for ownNamespace watch mode with operator should install a cluster extension successfully", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 77435.48958333331, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMOwnSingleNamespace] OLMv1 operator installation support for singleNamespace watch mode with operator should install a cluster extension successfully", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73646.70562500002, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-olmv1][OCPFeatureGate:NewOLMOwnSingleNamespace][Serial] OLMv1 operator installation support for ownNamespace watch mode with an operator that does not support ownNamespace installation mode should fail to install a cluster extension successfully", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32856.74106382979, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] Multi-AZ Clusters should spread the pods of a service across zones [Serial]", "TotalTestCount": 716, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32435.501634615386, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] PodTopologySpread Filtering validates 4 pods with MaxSkew=1 are evenly distributed into 2 nodes", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7185.223652968037, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates Pods with non-empty schedulingGates are blocked on scheduling", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4911.524337899545, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates local ephemeral storage resource limits of pods that are allowed to run", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 11203.831461187217, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates pod overhead is considered along with resource limits of pods that are allowed to run verify pod overhead is accounted for", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 13374.375433789957, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7718.990547945207, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeAffinity is respected if not matching", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 11168.07196347032, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5766.591232876712, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2352.7856164383547, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that required NodeAffinity setting is respected if matching", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6236.071780821917, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if matching", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5857.953287671231, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that taints-tolerations is respected if not matching", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5792.252191780822, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPredicates [Serial] validates that there is no conflict between pods with same hostPort but different hostIP and protocol", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10225.529360730592, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] PodTopologySpread Preemption validates proper pods are preempted", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 76519.97433789955, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 80123.11050228309, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints verify PriorityClass endpoints can be operated with different HTTP methods [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 62109.08552511417, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 67927.83315068489, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 67968.2925114155, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPreemption [Serial] validates pod disruption condition is added to the preempted pod [Conformance]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 65142.13826484019, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPriorities [Serial] Pod should be preferably scheduled to nodes pod can tolerate", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 76314.28712328766, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPriorities [Serial] Pod should be scheduled to node that don't match the PodAntiAffinity terms", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70058.67515981737, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-scheduling] SchedulerPriorities [Serial] PodTopologySpread Scoring validates pod should be preferably scheduled to node which makes the matching pods more evenly distributed", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 75591.24077625568, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on a privileged and unprivileged Pod with given SELinux with MountOption policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70383.84755319146, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on an unprivileged and privileged Pod with given SELinux with MountOption policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 71091.38308510643, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with MountOption policy and a different context on RWOP volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 69393.59925531918, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWO volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70019.61000000002, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWOP volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMountReadWriteOncePod] [Beta]", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 74891.38336538464, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different context on RWX volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 69600.7269148936, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different policy on RWO volume and SELinuxMount enabled (Recursive + MountOption) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 71303.35138297873, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different policy on RWO volume and SELinuxMount enabled (Recursive + nil) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 272, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73675.46085106385, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is bumped on two Pods with a different policy on RWO volume and SELinuxMount enabled (nil + Recursive) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70957.00351063827, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on a privileged and unprivileged Pod with given SELinux context and recursive policy [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 272, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70192.76574468079, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two Pods with Recursive policy and a different context on RWX volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 70017.62414893616, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two Pods with the same context on RWO volume and SELinuxMount enabled [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73432.67723404255, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two Pods with the same policy RWX volume (MountOption + MountOption) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 71746.48946808514, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two Pods with the same policy RWX volume (nil + MountOption) [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 273, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73140.10638297873, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two privileged Pods with mount policy RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73910.62074468087, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] error is not bumped on two privileged Pods with recursive policy RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxMount] [Beta] [Feature:OffByDefault]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 71032.48244680848, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is bumped on two Pods with a different context on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly]", "TotalTestCount": 436, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 75306.73649122805, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is bumped on two Pods with different policies on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly]", "TotalTestCount": 436, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 75739.63026315787, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is not bumped on two Pods with Recursive policy and a different context on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly]", "TotalTestCount": 435, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 73057.45114035088, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Mock selinux on mount metrics and SELinuxWarningController SELinuxMount metrics [LinuxOnly] [Feature:SELinux] [Serial] warning is not bumped on two Pods with the same context on RWO volume [FeatureGate:SELinuxMountReadWriteOncePod] [Beta] [FeatureGate:SELinuxChangePolicy] [Beta] [Feature:SELinuxMountReadWriteOncePodOnly]", "TotalTestCount": 435, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 75215.7350877193, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 664.6776712328767, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 64727.029543379, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes [Serial]", "TotalTestCount": 165, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 47823.85652173912, "PeriodStart": "2025-12-11", "PeriodEnd": "2025-12-13", "DaysWithData": 3 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 45527.919269406375, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: csi-hostpath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: CSI Ephemeral-volume (default fs)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 600, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 600, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should block a second pod from using an in-use ReadWriteOncePod volume on the same node", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes", "TotalTestCount": 518, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-12", "DaysWithData": 13 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] read-write-once-pod [MinimumKubeletVersion:1.27] should preempt lower priority pods using ReadWriteOncePod volumes [Serial]", "TotalTestCount": 157, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-12-11", "PeriodEnd": "2025-12-13", "DaysWithData": 3 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 600, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Dynamic Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Ephemeral Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (delete policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works after modifying source data, check deletion (persistent)", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] CSI Volumes [Driver: pd.csi.storage.gke.io] [Serial] [Testpattern: Pre-provisioned Snapshot (retain policy)] snapshottable [Feature:VolumeSnapshotDataSource] volume snapshot controller should check snapshot fields, check restore correctly works, check deletion (ephemeral)", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 23981.889589041093, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 100.89773869346733, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 98.93856410256413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 79.82497716894976, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 97.9863341645885, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 96.60506234413963, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 96.45466334164588, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 101.3735641025641, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 101.82007692307694, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5879.791695760599, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10511.888528678302, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9361.523266832914, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9754.245985037405, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9780.930249376557, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9676.816533665835, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9497.96221945137, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9687.189002493767, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9791.565561097255, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9870.79876884422, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 98.89292307692307, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 79.76410958904108, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 101.6130174563591, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 98.20254364089776, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 95.01658354114711, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 100.41961538461535, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 101.8196153846154, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7661.4130922693275, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 9552.614738154614, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7797.9049376558605, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7477.18029925187, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10109.780498753122, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7442.919950124689, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7685.475635910224, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10372.497655860348, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 163.33224438902744, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 3460, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7478.0308977556115, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 10308.318802992522, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6251.117356608478, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 184.63104738154615, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 3460, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 94.70139650872822, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 391.3883042394015, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 8563.091471321693, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 14341.830124688277, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 376.44977556109745, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 367.11147132169566, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 366.52665835411466, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 94.47002493765585, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 365.0774563591022, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7990.510374064836, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7846.4465835411465, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 385.90895261845395, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 7757.270972568579, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 14688.821670822945, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-disk] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 95.53635910224438, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1058.236708229426, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1059.5490274314218, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1059.0614713216958, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3522.1428643216072, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 99.34348717948717, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 79.15525114155253, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 96.55680798004987, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 99.08957605985036, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 103.20039900249375, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 95.95592307692309, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 100.52248717948716, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1959.000523690773, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2407.6288528678306, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1840.2459850374066, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1731.670224438903, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4332.382219451372, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1750.8561097256854, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1844.6764089775559, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1894.4243142144637, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1913.8642643391522, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1633.198428927681, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3705.8618952618467, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2038.474738154613, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1418.6208229426427, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1433.64174563591, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 393.76608478802996, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 97.6599501246883, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2308.9442643391526, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1569.2396758104735, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1463.5373067331673, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 375.86733167082303, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1556.183940149626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2446.759900249376, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: azure-file] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: cinder] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.003466334164588526, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: emptydir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPathSymlink] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.002643391521197007, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.0034663341645885283, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.002643391521197007, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: hostPath] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: block] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.005935162094763091, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.002967581047381546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: blockfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.0029675810473815457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.006234413965087282, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link-bindmounted] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.005935162094763092, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir-link] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.005935162094763091, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.0029675810473815457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: dir] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 600, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 642, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 598, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 600, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 644, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 697.7228717948718, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3047.255076923077, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3052.942051282052, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3057.1544615384614, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3054.5274358974357, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3040.7278461538467, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3056.852769230769, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3044.842923076924, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3051.0944615384624, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3034.0290256410253, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3057.513897435898, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3059.2896410256417, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: gce-localssd-scsi-fs] [Serial] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: local] [LocalVolumeType: tmpfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.0029675810473815457, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 670.8060273972603, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs3] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 658.2426484018263, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4110, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4199, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: nfs] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: (delete policy)] volumegroupsnapshottable [Feature:volumegroupsnapshot] VolumeGroupSnapshottable should create snapshots for multiple volumes in a pod", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4081, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 31.779246753246742, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.886684350132626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 51.966438356164375, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with mount options", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 34.054278350515446, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.423634020618564, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.670721649484534, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 35.464217506631286, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.871538461538464, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1442.8029896907215, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (block volmode)] volumes should store data", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3118.526056701031, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)(allowExpansion)] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2909.866159793814, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2984.5804123711337, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2979.1672680412366, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2957.34675257732, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2929.2898195876287, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3007.545335051546, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2895.438350515463, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4081, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32.18181818181818, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.14899204244031, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 51.418036529680364, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with mount options", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.20141752577319, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 31.962216494845364, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32.902242268041235, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.12063660477454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 35.8369230769231, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2025.1089175257734, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2994.455180412371, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing directory", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2119.0008247422684, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2003.7548453608256, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3750.7655927835053, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2099.489871134022, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2148.190412371134, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2173.236056701032, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2257.996649484536, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (default fs)] volumes should store data", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3121.1673969072162, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2224.602422680413, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3255.9701288659803, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 1474.1687371134017, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 4081, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 4170, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32.342010309278365, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2330.09100515464, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 190.50201030927832, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2459.9354639175262, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (block volmode) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4093.115541237114, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 189.32262886597937, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 189.88427835051544, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 189.78577319587626, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33.091159793814434, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (immediate-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 187.8897164948454, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read-only inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2272.485902061856, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should create read/write inline ephemeral volume", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2267.9111340206196, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support expansion of pvcs created for ephemeral pvcs", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 191.8304639175258, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support multiple inline ephemeral volumes", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2519.1619329896903, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs) (late-binding)] ephemeral should support two pods which have the same volume definition", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 4304.059948453609, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 4202, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] In-tree Volumes [Driver: vsphere] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 4168, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (block volmode)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 274, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 719.4855319148938, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] capacity provides storage capacity information", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup applied to the volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (Always)[LinuxOnly], rwop pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, new pod fsgroup applied to volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with different fsgroup applied to the volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] fsgroupchangepolicy (OnRootMismatch)[LinuxOnly], pod created with an initial fsgroup, volume contents ownership changed via chgrp in first pod, new pod with same fsgroup skips ownership changes to the volume contents", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (default fs)] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 273, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 63941.56, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (delayed binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (delayed binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ext3)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ext4)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (filesystem volmode)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 272, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 44547.269361702114, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (immediate binding)] topology should fail to schedule a pod which has topologies that conflict with AllowedTopologies", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (immediate binding)] topology should provision a volume and schedule a pod with AllowedTopologies", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand Verify if offline PVC expansion works", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited and the pod is re-created on the same node after controller resize is finished", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)(allowExpansion)] [Feature:Windows] volume-expand should resize volume when PVC is edited while pod is using it", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should mount multiple PV pointing to the same storage on the same node", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision correct filesystem size when restoring snapshot to larger size pvc [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with any volume data source [Serial]", "TotalTestCount": 272, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with mount options", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with pvc data source (ROX mode)", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source (ROX mode) [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] provisioning should provision storage with snapshot data source [Feature:VolumeSnapshotDataSource]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing directory", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support non-existent path", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-expand should not allow expansion of pvcs without AllowVolumeExpansion property", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should be protected by vac-protection finalizer", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should create a volume with VAC", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume that already has a VAC", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volume-modify [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass] should modify volume with no VAC", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Dynamic PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should support volume limits [Serial]", "TotalTestCount": 275, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Generic Ephemeral-volume (default fs)] volumeLimits should verify that all csinodes have volume limits", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support existing directory", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support non-existent path", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (default fs)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ext3)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ext4)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Inline-volume (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (block volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (block volmode)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should be able to unmount after the subpath directory is deleted [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directories when readOnly specified in the volumeSource", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing directory", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support existing single file [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support file as subpath [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support non-existent path", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly directory specified in the volumeMount", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] subPath should support readOnly file specified in the volumeMount [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (default fs)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ext3)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ext3)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ext4)] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ext4)] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (filesystem volmode)] volumeMode should not mount / map unused volumes in a pod [LinuxOnly]", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should allow exec of files on the volume", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] OCP CSI Volumes [Driver: csi-hostpath-groupsnapshot] [OCPFeatureGate:VolumeGroupSnapshot] [Testpattern: Pre-provisioned PV (ntfs)] [Feature:Windows] volumes should store data", "TotalTestCount": 471, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local Stress with local volumes [Serial] should be able to process many pods and reuse local volumes", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 64340.29132420091, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] One pod requesting one prebound PVC should be able to mount volume and read from pod1", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3043.118307692307, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] One pod requesting one prebound PVC should be able to mount volume and write from pod1", "TotalTestCount": 674, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3037.7644615384615, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume at the same time should be able to write from pod1 and read from pod2", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3053.552871794871, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] PersistentVolumes-local [Volume type: gce-localssd-scsi-fs] [Serial] Two pods mounting a local volume one after the other should be able to write from pod1 and read from pod2", "TotalTestCount": 676, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 3051.569794871795, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create metrics for total number of volumes in A/D Controller", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 707.4521004566209, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create metrics for total time taken in volume operations in P/V Controller", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 6857.858675799087, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create prometheus metrics for volume provisioning and attach/detach", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 708.1776712328767, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics in Volume Manager", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 8513.18579908676, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics with the correct BlockMode PVC ref", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 33550.68566210046, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics Ephemeral should create volume metrics with the correct FilesystemMode PVC ref", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 31795.464474885848, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create metrics for total number of volumes in A/D Controller", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 749.847214611872, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create metrics for total time taken in volume operations in P/V Controller", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 163503.40305936075, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create prometheus metrics for volume provisioning and attach/detach", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 708.3780365296802, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics in Volume Manager", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 17319.017579908672, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics with the correct BlockMode PVC ref", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 32204.58799086758, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVC should create volume metrics with the correct FilesystemMode PVC ref", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 41350.293561643826, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create bound pv/pvc count metrics for pvc controller after creating both pv and pvc", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 718.5043378995437, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create bound pv/pvc count metrics for pvc controller with volume attributes class dimension after creating both pv and pvc [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass]", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 715.078173076923, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create none metrics for pvc controller before creating any PV or PVC", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 704.9527397260274, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create total pv count metrics for with plugin and volume mode labels after creating pv", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 709.3829680365294, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create unbound pv count metrics for pvc controller after creating pv only", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 706.9210045662102, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create unbound pvc count metrics for pvc controller after creating pvc only", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 707.8831506849316, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage] [Serial] Volume metrics PVController should create unbound pvc count metrics for pvc controller with volume attributes class dimension after creating pvc only [FeatureGate:VolumeAttributesClass] [Feature:VolumeAttributesClass]", "TotalTestCount": 710, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 714.5561538461541, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] baseline namespace should allow pods with inline volumes when the driver uses the baseline label [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2133.568767123287, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] baseline namespace should deny pods with inline volumes when the driver uses the privileged label [Suite:openshift/conformance/serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2024.84899543379, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] privileged namespace should allow pods with inline volumes when the driver uses the privileged label [Suite:openshift/conformance/serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2100.6159360730585, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] privileged namespace should allow pods with inline volumes when the driver uses the restricted label [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2130.9482191780826, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] restricted namespace should allow pods with inline volumes when the driver uses the restricted label [Suite:openshift/conformance/serial]", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2078.930456621005, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] restricted namespace should deny pods with inline volumes when the driver uses the baseline label [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2083.071369863014, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:CSIInlineVolumeAdmission][Serial] restricted namespace should deny pods with inline volumes when the driver uses the privileged label [Suite:openshift/conformance/serial]", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 2088.3289041095895, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-storage][Feature:DisableStorageClass][Serial][apigroup:operator.openshift.io] should reconcile the StorageClass when StorageClassState is Managed [Suite:openshift/conformance/serial]", "TotalTestCount": 709, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 5921.728647342996, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-windows] Hybrid cluster network for all supported CNIs should have stable networking for Linux and Windows pods", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] Hybrid cluster network for all supported CNIs should provide Internet connection and DNS for Windows containers [Feature:Networking-IPv4] [Feature:Networking-DNS]", "TotalTestCount": 3890, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.028409090909090908, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] Hybrid cluster network for all supported CNIs should provide Internet connection for Linux containers [Feature:Networking-IPv4]", "TotalTestCount": 3890, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.04261363636363637, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] Services should be able to create a functioning NodePort service for Windows", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers container command path validation", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers container stats validation", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers metrics should report count of started and failed to start HostProcess containers", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should run as a process on the host/node", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should run as localgroup accounts", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.01855361596009975, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support init containers", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support querying api-server using in-cluster config", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHostProcessContainers] [MinimumKubeletVersion:1.22] HostProcess containers should support various volume mount types", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:WindowsHyperVContainers] HyperV containers should start a hyperv isolated container", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Cpu Resources [Serial] Container limits should not be exceeded after waiting 2 minutes", "TotalTestCount": 740, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] DNS should support configurable pod DNS servers", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Kubelet-Stats Kubelet stats collection for Windows nodes when running 3 pods should return within 10 seconds", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Kubelet-Stats Kubelet stats collection for Windows nodes when windows is booted should return bootid within 10 seconds", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Kubelet-Stats [Serial] Kubelet stats collection for Windows nodes when running 10 pods should return within 10 seconds", "TotalTestCount": 742, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should be able create pods and run containers with a given username", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should be able to create pod and run containers", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should ignore Linux Specific SecurityContext if set", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with containers running as CONTAINERADMINISTRATOR when runAsNonRoot is true", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with containers running as ContainerAdministrator when runAsNonRoot is true", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with unknown usernames at Container level", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should not be able to create pods with unknown usernames at Pod level", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] SecurityContext should override SecurityContext username if set", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Windows volume mounts check volume mount permissions container should have readOnly permissions on emptyDir", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "[sig-windows] [Feature:Windows] Windows volume mounts check volume mount permissions container should have readOnly permissions on hostMapPath", "TotalTestCount": 4231, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions baremetal", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.00006072874493927126, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions cloud-controller-manager", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.00008097165991902834, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions cluster-api", "TotalTestCount": 943, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-30", "PeriodEnd": "2025-12-13", "DaysWithData": 14 }, { "Release": "4.22", "TestName": "verify operator conditions config-operator", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.00010121457489878545, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions csi-snapshot-controller", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions insights", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.00002024291497975708, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions kube-storage-version-migrator", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.000020242914979757088, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions machine-approver", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions marketplace", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions olm", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions operator-lifecycle-manager", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.000020242914979757088, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions operator-lifecycle-manager-catalog", "TotalTestCount": 6806, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions service-ca", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 }, { "Release": "4.22", "TestName": "verify operator conditions storage", "TotalTestCount": 6807, "TotalFailureCount": 0, "TotalFlakeCount": 0, "FailureRate": 0, "AvgDurationMs": 0.00016194331983805673, "PeriodStart": "2025-11-29", "PeriodEnd": "2025-12-13", "DaysWithData": 15 } ]