api: generated files for PodSchedulingContext
This commit is contained in:
parent
fec5233668
commit
2b8a4e8097
172
api/openapi-spec/swagger.json
generated
172
api/openapi-spec/swagger.json
generated
@ -13035,8 +13035,8 @@
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodScheduling": {
|
||||
"description": "PodScheduling objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContext": {
|
||||
"description": "PodSchedulingContext objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
"properties": {
|
||||
"apiVersion": {
|
||||
"description": "APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources",
|
||||
@ -13051,11 +13051,11 @@
|
||||
"description": "Standard object metadata"
|
||||
},
|
||||
"spec": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingSpec",
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec",
|
||||
"description": "Spec describes where resources for the Pod are needed."
|
||||
},
|
||||
"status": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingStatus",
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus",
|
||||
"description": "Status describes where resources for the Pod can be allocated."
|
||||
}
|
||||
},
|
||||
@ -13066,22 +13066,22 @@
|
||||
"x-kubernetes-group-version-kind": [
|
||||
{
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
]
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingList": {
|
||||
"description": "PodSchedulingList is a collection of Pod scheduling objects.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextList": {
|
||||
"description": "PodSchedulingContextList is a collection of Pod scheduling objects.",
|
||||
"properties": {
|
||||
"apiVersion": {
|
||||
"description": "APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources",
|
||||
"type": "string"
|
||||
},
|
||||
"items": {
|
||||
"description": "Items is the list of PodScheduling objects.",
|
||||
"description": "Items is the list of PodSchedulingContext objects.",
|
||||
"items": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
},
|
||||
"type": "array"
|
||||
},
|
||||
@ -13101,13 +13101,13 @@
|
||||
"x-kubernetes-group-version-kind": [
|
||||
{
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodSchedulingList",
|
||||
"kind": "PodSchedulingContextList",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
]
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingSpec": {
|
||||
"description": "PodSchedulingSpec describes where resources for the Pod are needed.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec": {
|
||||
"description": "PodSchedulingContextSpec describes where resources for the Pod are needed.",
|
||||
"properties": {
|
||||
"potentialNodes": {
|
||||
"description": "PotentialNodes lists nodes where the Pod might be able to run.\n\nThe size of this field is limited to 128. This is large enough for many clusters. Larger clusters may need more attempts to find a node that suits all pending resources. This may get increased in the future, but not reduced.",
|
||||
@ -13124,8 +13124,8 @@
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingStatus": {
|
||||
"description": "PodSchedulingStatus describes where resources for the Pod can be allocated.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus": {
|
||||
"description": "PodSchedulingContextStatus describes where resources for the Pod can be allocated.",
|
||||
"properties": {
|
||||
"resourceClaims": {
|
||||
"description": "ResourceClaims describes resource availability for each pod.spec.resourceClaim entry where the corresponding ResourceClaim uses \"WaitForFirstConsumer\" allocation mode.",
|
||||
@ -78180,13 +78180,13 @@
|
||||
]
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts": {
|
||||
"delete": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "delete collection of PodScheduling",
|
||||
"operationId": "deleteResourceV1alpha2CollectionNamespacedPodScheduling",
|
||||
"description": "delete collection of PodSchedulingContext",
|
||||
"operationId": "deleteResourceV1alpha2CollectionNamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
@ -78305,7 +78305,7 @@
|
||||
"x-kubernetes-action": "deletecollection",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -78313,8 +78313,8 @@
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "list or watch objects of kind PodScheduling",
|
||||
"operationId": "listResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "list or watch objects of kind PodSchedulingContext",
|
||||
"operationId": "listResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "allowWatchBookmarks requests watch events with type \"BOOKMARK\". Servers that do not implement bookmarks may ignore this flag and bookmarks are sent at the server's discretion. Clients should not assume bookmarks are returned at any specific interval, nor may they assume the server will send any BOOKMARK event during a session. If this is not a watch, this field is ignored.",
|
||||
@ -78398,7 +78398,7 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78414,7 +78414,7 @@
|
||||
"x-kubernetes-action": "list",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -78439,15 +78439,15 @@
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "create a PodScheduling",
|
||||
"operationId": "createResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "create a PodSchedulingContext",
|
||||
"operationId": "createResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
"name": "body",
|
||||
"required": true,
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
{
|
||||
@ -78481,19 +78481,19 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"201": {
|
||||
"description": "Created",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"202": {
|
||||
"description": "Accepted",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78509,18 +78509,18 @@
|
||||
"x-kubernetes-action": "post",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings/{name}": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts/{name}": {
|
||||
"delete": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "delete a PodScheduling",
|
||||
"operationId": "deleteResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "delete a PodSchedulingContext",
|
||||
"operationId": "deleteResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
@ -78567,13 +78567,13 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"202": {
|
||||
"description": "Accepted",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78589,7 +78589,7 @@
|
||||
"x-kubernetes-action": "delete",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -78597,8 +78597,8 @@
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "read the specified PodScheduling",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "read the specified PodSchedulingContext",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -78608,7 +78608,7 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78624,13 +78624,13 @@
|
||||
"x-kubernetes-action": "get",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"parameters": [
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -78660,8 +78660,8 @@
|
||||
"application/strategic-merge-patch+json",
|
||||
"application/apply-patch+yaml"
|
||||
],
|
||||
"description": "partially update the specified PodScheduling",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "partially update the specified PodSchedulingContext",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
@ -78709,13 +78709,13 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"201": {
|
||||
"description": "Created",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78731,7 +78731,7 @@
|
||||
"x-kubernetes-action": "patch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -78739,15 +78739,15 @@
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "replace the specified PodScheduling",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "replace the specified PodSchedulingContext",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
"name": "body",
|
||||
"required": true,
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
{
|
||||
@ -78781,13 +78781,13 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"201": {
|
||||
"description": "Created",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78803,18 +78803,18 @@
|
||||
"x-kubernetes-action": "put",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings/{name}/status": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts/{name}/status": {
|
||||
"get": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "read status of the specified PodScheduling",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "read status of the specified PodSchedulingContext",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -78824,7 +78824,7 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78840,13 +78840,13 @@
|
||||
"x-kubernetes-action": "get",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"parameters": [
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -78876,8 +78876,8 @@
|
||||
"application/strategic-merge-patch+json",
|
||||
"application/apply-patch+yaml"
|
||||
],
|
||||
"description": "partially update status of the specified PodScheduling",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "partially update status of the specified PodSchedulingContext",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
@ -78925,13 +78925,13 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"201": {
|
||||
"description": "Created",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -78947,7 +78947,7 @@
|
||||
"x-kubernetes-action": "patch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -78955,15 +78955,15 @@
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "replace status of the specified PodScheduling",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "replace status of the specified PodSchedulingContext",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"parameters": [
|
||||
{
|
||||
"in": "body",
|
||||
"name": "body",
|
||||
"required": true,
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
{
|
||||
@ -78997,13 +78997,13 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"201": {
|
||||
"description": "Created",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -79019,7 +79019,7 @@
|
||||
"x-kubernetes-action": "put",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
@ -80496,13 +80496,13 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/podschedulingcontexts": {
|
||||
"get": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "list or watch objects of kind PodScheduling",
|
||||
"operationId": "listResourceV1alpha2PodSchedulingForAllNamespaces",
|
||||
"description": "list or watch objects of kind PodSchedulingContext",
|
||||
"operationId": "listResourceV1alpha2PodSchedulingContextForAllNamespaces",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -80514,7 +80514,7 @@
|
||||
"200": {
|
||||
"description": "OK",
|
||||
"schema": {
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/definitions/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"401": {
|
||||
@ -80530,7 +80530,7 @@
|
||||
"x-kubernetes-action": "list",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -81462,13 +81462,13 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulingcontexts": {
|
||||
"get": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "watch individual changes to a list of PodScheduling. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingList",
|
||||
"description": "watch individual changes to a list of PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingContextList",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -81496,7 +81496,7 @@
|
||||
"x-kubernetes-action": "watchlist",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -81588,13 +81588,13 @@
|
||||
}
|
||||
]
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulings/{name}": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulingcontexts/{name}": {
|
||||
"get": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "watch changes to an object of kind PodScheduling. deprecated: use the 'watch' parameter with a list operation instead, filtered to a single item with the 'fieldSelector' parameter.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "watch changes to an object of kind PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead, filtered to a single item with the 'fieldSelector' parameter.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -81622,7 +81622,7 @@
|
||||
"x-kubernetes-action": "watch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -81663,7 +81663,7 @@
|
||||
"uniqueItems": true
|
||||
},
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -82242,13 +82242,13 @@
|
||||
}
|
||||
]
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/podschedulingcontexts": {
|
||||
"get": {
|
||||
"consumes": [
|
||||
"*/*"
|
||||
],
|
||||
"description": "watch individual changes to a list of PodScheduling. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2PodSchedulingListForAllNamespaces",
|
||||
"description": "watch individual changes to a list of PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2PodSchedulingContextListForAllNamespaces",
|
||||
"produces": [
|
||||
"application/json",
|
||||
"application/yaml",
|
||||
@ -82276,7 +82276,7 @@
|
||||
"x-kubernetes-action": "watchlist",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
|
@ -104,8 +104,8 @@
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodScheduling": {
|
||||
"description": "PodScheduling objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContext": {
|
||||
"description": "PodSchedulingContext objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
"properties": {
|
||||
"apiVersion": {
|
||||
"description": "APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources",
|
||||
@ -127,7 +127,7 @@
|
||||
"spec": {
|
||||
"allOf": [
|
||||
{
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingSpec"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec"
|
||||
}
|
||||
],
|
||||
"default": {},
|
||||
@ -136,7 +136,7 @@
|
||||
"status": {
|
||||
"allOf": [
|
||||
{
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingStatus"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus"
|
||||
}
|
||||
],
|
||||
"default": {},
|
||||
@ -150,24 +150,24 @@
|
||||
"x-kubernetes-group-version-kind": [
|
||||
{
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
]
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingList": {
|
||||
"description": "PodSchedulingList is a collection of Pod scheduling objects.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextList": {
|
||||
"description": "PodSchedulingContextList is a collection of Pod scheduling objects.",
|
||||
"properties": {
|
||||
"apiVersion": {
|
||||
"description": "APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources",
|
||||
"type": "string"
|
||||
},
|
||||
"items": {
|
||||
"description": "Items is the list of PodScheduling objects.",
|
||||
"description": "Items is the list of PodSchedulingContext objects.",
|
||||
"items": {
|
||||
"allOf": [
|
||||
{
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
],
|
||||
"default": {}
|
||||
@ -195,13 +195,13 @@
|
||||
"x-kubernetes-group-version-kind": [
|
||||
{
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodSchedulingList",
|
||||
"kind": "PodSchedulingContextList",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
]
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingSpec": {
|
||||
"description": "PodSchedulingSpec describes where resources for the Pod are needed.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec": {
|
||||
"description": "PodSchedulingContextSpec describes where resources for the Pod are needed.",
|
||||
"properties": {
|
||||
"potentialNodes": {
|
||||
"description": "PotentialNodes lists nodes where the Pod might be able to run.\n\nThe size of this field is limited to 128. This is large enough for many clusters. Larger clusters may need more attempts to find a node that suits all pending resources. This may get increased in the future, but not reduced.",
|
||||
@ -219,8 +219,8 @@
|
||||
},
|
||||
"type": "object"
|
||||
},
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingStatus": {
|
||||
"description": "PodSchedulingStatus describes where resources for the Pod can be allocated.",
|
||||
"io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus": {
|
||||
"description": "PodSchedulingContextStatus describes where resources for the Pod can be allocated.",
|
||||
"properties": {
|
||||
"resourceClaims": {
|
||||
"description": "ResourceClaims describes resource availability for each pod.spec.resourceClaim entry where the corresponding ResourceClaim uses \"WaitForFirstConsumer\" allocation mode.",
|
||||
@ -1878,10 +1878,10 @@
|
||||
]
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts": {
|
||||
"delete": {
|
||||
"description": "delete collection of PodScheduling",
|
||||
"operationId": "deleteResourceV1alpha2CollectionNamespacedPodScheduling",
|
||||
"description": "delete collection of PodSchedulingContext",
|
||||
"operationId": "deleteResourceV1alpha2CollectionNamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "The continue option should be set when retrieving more results from the server. Since this value is server defined, clients may only use the continue value from a previous query result with identical query parameters (except for the value of continue) and the server may reject a continue value it does not recognize. If the specified continue value is no longer valid whether due to expiration (generally five to fifteen minutes) or a configuration change on the server, the server will respond with a 410 ResourceExpired error together with a continue token. If the client needs a consistent list, it must restart their list without the continue field. Otherwise, the client may send another list request with the token received with the 410 error, the server will respond with a list starting from the next key, but from the latest snapshot, which is inconsistent from the previous list results - objects that are created, modified, or deleted after the first list request will be included in the response, as long as their keys are after the \"next key\".\n\nThis field is not supported when watch is true. Clients may start a watch from the last resourceVersion value returned by the server and not miss any modifications.",
|
||||
@ -2032,13 +2032,13 @@
|
||||
"x-kubernetes-action": "deletecollection",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"get": {
|
||||
"description": "list or watch objects of kind PodScheduling",
|
||||
"operationId": "listResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "list or watch objects of kind PodSchedulingContext",
|
||||
"operationId": "listResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "allowWatchBookmarks requests watch events with type \"BOOKMARK\". Servers that do not implement bookmarks may ignore this flag and bookmarks are sent at the server's discretion. Clients should not assume bookmarks are returned at any specific interval, nor may they assume the server will send any BOOKMARK event during a session. If this is not a watch, this field is ignored.",
|
||||
@ -2136,27 +2136,27 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/json;stream=watch": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf;stream=watch": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2172,7 +2172,7 @@
|
||||
"x-kubernetes-action": "list",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -2198,8 +2198,8 @@
|
||||
}
|
||||
],
|
||||
"post": {
|
||||
"description": "create a PodScheduling",
|
||||
"operationId": "createResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "create a PodSchedulingContext",
|
||||
"operationId": "createResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2233,7 +2233,7 @@
|
||||
"content": {
|
||||
"*/*": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -2243,17 +2243,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2263,17 +2263,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2283,17 +2283,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2309,15 +2309,15 @@
|
||||
"x-kubernetes-action": "post",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings/{name}": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts/{name}": {
|
||||
"delete": {
|
||||
"description": "delete a PodScheduling",
|
||||
"operationId": "deleteResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "delete a PodSchedulingContext",
|
||||
"operationId": "deleteResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2370,17 +2370,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2390,17 +2390,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2416,29 +2416,29 @@
|
||||
"x-kubernetes-action": "delete",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"get": {
|
||||
"description": "read the specified PodScheduling",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "read the specified PodSchedulingContext",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2454,13 +2454,13 @@
|
||||
"x-kubernetes-action": "get",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"parameters": [
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -2490,8 +2490,8 @@
|
||||
}
|
||||
],
|
||||
"patch": {
|
||||
"description": "partially update the specified PodScheduling",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "partially update the specified PodSchedulingContext",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2559,17 +2559,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2579,17 +2579,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2605,13 +2605,13 @@
|
||||
"x-kubernetes-action": "patch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"put": {
|
||||
"description": "replace the specified PodScheduling",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "replace the specified PodSchedulingContext",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2645,7 +2645,7 @@
|
||||
"content": {
|
||||
"*/*": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -2655,17 +2655,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2675,17 +2675,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2701,31 +2701,31 @@
|
||||
"x-kubernetes-action": "put",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulings/{name}/status": {
|
||||
"/apis/resource.k8s.io/v1alpha2/namespaces/{namespace}/podschedulingcontexts/{name}/status": {
|
||||
"get": {
|
||||
"description": "read status of the specified PodScheduling",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "read status of the specified PodSchedulingContext",
|
||||
"operationId": "readResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2741,13 +2741,13 @@
|
||||
"x-kubernetes-action": "get",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"parameters": [
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -2777,8 +2777,8 @@
|
||||
}
|
||||
],
|
||||
"patch": {
|
||||
"description": "partially update status of the specified PodScheduling",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "partially update status of the specified PodSchedulingContext",
|
||||
"operationId": "patchResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2846,17 +2846,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2866,17 +2866,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2892,13 +2892,13 @@
|
||||
"x-kubernetes-action": "patch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
"put": {
|
||||
"description": "replace status of the specified PodScheduling",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingStatus",
|
||||
"description": "replace status of the specified PodSchedulingContext",
|
||||
"operationId": "replaceResourceV1alpha2NamespacedPodSchedulingContextStatus",
|
||||
"parameters": [
|
||||
{
|
||||
"description": "When present, indicates that modifications should not be persisted. An invalid or unrecognized dryRun directive will result in an error response and no further processing of the request. Valid values are: - All: all dry run stages will be processed",
|
||||
@ -2932,7 +2932,7 @@
|
||||
"content": {
|
||||
"*/*": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
}
|
||||
@ -2942,17 +2942,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2962,17 +2962,17 @@
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodScheduling"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContext"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -2988,7 +2988,7 @@
|
||||
"x-kubernetes-action": "put",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
}
|
||||
@ -4936,36 +4936,36 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/podschedulingcontexts": {
|
||||
"get": {
|
||||
"description": "list or watch objects of kind PodScheduling",
|
||||
"operationId": "listResourceV1alpha2PodSchedulingForAllNamespaces",
|
||||
"description": "list or watch objects of kind PodSchedulingContext",
|
||||
"operationId": "listResourceV1alpha2PodSchedulingContextForAllNamespaces",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
"application/json": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/json;stream=watch": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/vnd.kubernetes.protobuf;stream=watch": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
},
|
||||
"application/yaml": {
|
||||
"schema": {
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingList"
|
||||
"$ref": "#/components/schemas/io.k8s.api.resource.v1alpha2.PodSchedulingContextList"
|
||||
}
|
||||
}
|
||||
},
|
||||
@ -4981,7 +4981,7 @@
|
||||
"x-kubernetes-action": "list",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -6197,10 +6197,10 @@
|
||||
}
|
||||
}
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulingcontexts": {
|
||||
"get": {
|
||||
"description": "watch individual changes to a list of PodScheduling. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingList",
|
||||
"description": "watch individual changes to a list of PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingContextList",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
@ -6242,7 +6242,7 @@
|
||||
"x-kubernetes-action": "watchlist",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -6358,10 +6358,10 @@
|
||||
}
|
||||
]
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulings/{name}": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/namespaces/{namespace}/podschedulingcontexts/{name}": {
|
||||
"get": {
|
||||
"description": "watch changes to an object of kind PodScheduling. deprecated: use the 'watch' parameter with a list operation instead, filtered to a single item with the 'fieldSelector' parameter.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodScheduling",
|
||||
"description": "watch changes to an object of kind PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead, filtered to a single item with the 'fieldSelector' parameter.",
|
||||
"operationId": "watchResourceV1alpha2NamespacedPodSchedulingContext",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
@ -6403,7 +6403,7 @@
|
||||
"x-kubernetes-action": "watch",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
@ -6454,7 +6454,7 @@
|
||||
}
|
||||
},
|
||||
{
|
||||
"description": "name of the PodScheduling",
|
||||
"description": "name of the PodSchedulingContext",
|
||||
"in": "path",
|
||||
"name": "name",
|
||||
"required": true,
|
||||
@ -7193,10 +7193,10 @@
|
||||
}
|
||||
]
|
||||
},
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/podschedulings": {
|
||||
"/apis/resource.k8s.io/v1alpha2/watch/podschedulingcontexts": {
|
||||
"get": {
|
||||
"description": "watch individual changes to a list of PodScheduling. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2PodSchedulingListForAllNamespaces",
|
||||
"description": "watch individual changes to a list of PodSchedulingContext. deprecated: use the 'watch' parameter with a list operation instead.",
|
||||
"operationId": "watchResourceV1alpha2PodSchedulingContextListForAllNamespaces",
|
||||
"responses": {
|
||||
"200": {
|
||||
"content": {
|
||||
@ -7238,7 +7238,7 @@
|
||||
"x-kubernetes-action": "watchlist",
|
||||
"x-kubernetes-group-version-kind": {
|
||||
"group": "resource.k8s.io",
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"version": "v1alpha2"
|
||||
}
|
||||
},
|
||||
|
108
pkg/apis/resource/v1alpha2/zz_generated.conversion.go
generated
108
pkg/apis/resource/v1alpha2/zz_generated.conversion.go
generated
@ -50,43 +50,43 @@ func RegisterConversions(s *runtime.Scheme) error {
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodScheduling)(nil), (*resource.PodScheduling)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodScheduling_To_resource_PodScheduling(a.(*v1alpha2.PodScheduling), b.(*resource.PodScheduling), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingContext)(nil), (*resource.PodSchedulingContext)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingContext_To_resource_PodSchedulingContext(a.(*v1alpha2.PodSchedulingContext), b.(*resource.PodSchedulingContext), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodScheduling)(nil), (*v1alpha2.PodScheduling)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodScheduling_To_v1alpha2_PodScheduling(a.(*resource.PodScheduling), b.(*v1alpha2.PodScheduling), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingContext)(nil), (*v1alpha2.PodSchedulingContext)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingContext_To_v1alpha2_PodSchedulingContext(a.(*resource.PodSchedulingContext), b.(*v1alpha2.PodSchedulingContext), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingList)(nil), (*resource.PodSchedulingList)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingList_To_resource_PodSchedulingList(a.(*v1alpha2.PodSchedulingList), b.(*resource.PodSchedulingList), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingContextList)(nil), (*resource.PodSchedulingContextList)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingContextList_To_resource_PodSchedulingContextList(a.(*v1alpha2.PodSchedulingContextList), b.(*resource.PodSchedulingContextList), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingList)(nil), (*v1alpha2.PodSchedulingList)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingList_To_v1alpha2_PodSchedulingList(a.(*resource.PodSchedulingList), b.(*v1alpha2.PodSchedulingList), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingContextList)(nil), (*v1alpha2.PodSchedulingContextList)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingContextList_To_v1alpha2_PodSchedulingContextList(a.(*resource.PodSchedulingContextList), b.(*v1alpha2.PodSchedulingContextList), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingSpec)(nil), (*resource.PodSchedulingSpec)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec(a.(*v1alpha2.PodSchedulingSpec), b.(*resource.PodSchedulingSpec), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingContextSpec)(nil), (*resource.PodSchedulingContextSpec)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec(a.(*v1alpha2.PodSchedulingContextSpec), b.(*resource.PodSchedulingContextSpec), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingSpec)(nil), (*v1alpha2.PodSchedulingSpec)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec(a.(*resource.PodSchedulingSpec), b.(*v1alpha2.PodSchedulingSpec), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingContextSpec)(nil), (*v1alpha2.PodSchedulingContextSpec)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec(a.(*resource.PodSchedulingContextSpec), b.(*v1alpha2.PodSchedulingContextSpec), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingStatus)(nil), (*resource.PodSchedulingStatus)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus(a.(*v1alpha2.PodSchedulingStatus), b.(*resource.PodSchedulingStatus), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*v1alpha2.PodSchedulingContextStatus)(nil), (*resource.PodSchedulingContextStatus)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus(a.(*v1alpha2.PodSchedulingContextStatus), b.(*resource.PodSchedulingContextStatus), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingStatus)(nil), (*v1alpha2.PodSchedulingStatus)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus(a.(*resource.PodSchedulingStatus), b.(*v1alpha2.PodSchedulingStatus), scope)
|
||||
if err := s.AddGeneratedConversionFunc((*resource.PodSchedulingContextStatus)(nil), (*v1alpha2.PodSchedulingContextStatus)(nil), func(a, b interface{}, scope conversion.Scope) error {
|
||||
return Convert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus(a.(*resource.PodSchedulingContextStatus), b.(*v1alpha2.PodSchedulingContextStatus), scope)
|
||||
}); err != nil {
|
||||
return err
|
||||
}
|
||||
@ -247,100 +247,100 @@ func Convert_resource_AllocationResult_To_v1alpha2_AllocationResult(in *resource
|
||||
return autoConvert_resource_AllocationResult_To_v1alpha2_AllocationResult(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_v1alpha2_PodScheduling_To_resource_PodScheduling(in *v1alpha2.PodScheduling, out *resource.PodScheduling, s conversion.Scope) error {
|
||||
func autoConvert_v1alpha2_PodSchedulingContext_To_resource_PodSchedulingContext(in *v1alpha2.PodSchedulingContext, out *resource.PodSchedulingContext, s conversion.Scope) error {
|
||||
out.ObjectMeta = in.ObjectMeta
|
||||
if err := Convert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec(&in.Spec, &out.Spec, s); err != nil {
|
||||
if err := Convert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec(&in.Spec, &out.Spec, s); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := Convert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus(&in.Status, &out.Status, s); err != nil {
|
||||
if err := Convert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus(&in.Status, &out.Status, s); err != nil {
|
||||
return err
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_v1alpha2_PodScheduling_To_resource_PodScheduling is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodScheduling_To_resource_PodScheduling(in *v1alpha2.PodScheduling, out *resource.PodScheduling, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodScheduling_To_resource_PodScheduling(in, out, s)
|
||||
// Convert_v1alpha2_PodSchedulingContext_To_resource_PodSchedulingContext is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingContext_To_resource_PodSchedulingContext(in *v1alpha2.PodSchedulingContext, out *resource.PodSchedulingContext, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingContext_To_resource_PodSchedulingContext(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_resource_PodScheduling_To_v1alpha2_PodScheduling(in *resource.PodScheduling, out *v1alpha2.PodScheduling, s conversion.Scope) error {
|
||||
func autoConvert_resource_PodSchedulingContext_To_v1alpha2_PodSchedulingContext(in *resource.PodSchedulingContext, out *v1alpha2.PodSchedulingContext, s conversion.Scope) error {
|
||||
out.ObjectMeta = in.ObjectMeta
|
||||
if err := Convert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec(&in.Spec, &out.Spec, s); err != nil {
|
||||
if err := Convert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec(&in.Spec, &out.Spec, s); err != nil {
|
||||
return err
|
||||
}
|
||||
if err := Convert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus(&in.Status, &out.Status, s); err != nil {
|
||||
if err := Convert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus(&in.Status, &out.Status, s); err != nil {
|
||||
return err
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_resource_PodScheduling_To_v1alpha2_PodScheduling is an autogenerated conversion function.
|
||||
func Convert_resource_PodScheduling_To_v1alpha2_PodScheduling(in *resource.PodScheduling, out *v1alpha2.PodScheduling, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodScheduling_To_v1alpha2_PodScheduling(in, out, s)
|
||||
// Convert_resource_PodSchedulingContext_To_v1alpha2_PodSchedulingContext is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingContext_To_v1alpha2_PodSchedulingContext(in *resource.PodSchedulingContext, out *v1alpha2.PodSchedulingContext, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingContext_To_v1alpha2_PodSchedulingContext(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_v1alpha2_PodSchedulingList_To_resource_PodSchedulingList(in *v1alpha2.PodSchedulingList, out *resource.PodSchedulingList, s conversion.Scope) error {
|
||||
func autoConvert_v1alpha2_PodSchedulingContextList_To_resource_PodSchedulingContextList(in *v1alpha2.PodSchedulingContextList, out *resource.PodSchedulingContextList, s conversion.Scope) error {
|
||||
out.ListMeta = in.ListMeta
|
||||
out.Items = *(*[]resource.PodScheduling)(unsafe.Pointer(&in.Items))
|
||||
out.Items = *(*[]resource.PodSchedulingContext)(unsafe.Pointer(&in.Items))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_v1alpha2_PodSchedulingList_To_resource_PodSchedulingList is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingList_To_resource_PodSchedulingList(in *v1alpha2.PodSchedulingList, out *resource.PodSchedulingList, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingList_To_resource_PodSchedulingList(in, out, s)
|
||||
// Convert_v1alpha2_PodSchedulingContextList_To_resource_PodSchedulingContextList is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingContextList_To_resource_PodSchedulingContextList(in *v1alpha2.PodSchedulingContextList, out *resource.PodSchedulingContextList, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingContextList_To_resource_PodSchedulingContextList(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_resource_PodSchedulingList_To_v1alpha2_PodSchedulingList(in *resource.PodSchedulingList, out *v1alpha2.PodSchedulingList, s conversion.Scope) error {
|
||||
func autoConvert_resource_PodSchedulingContextList_To_v1alpha2_PodSchedulingContextList(in *resource.PodSchedulingContextList, out *v1alpha2.PodSchedulingContextList, s conversion.Scope) error {
|
||||
out.ListMeta = in.ListMeta
|
||||
out.Items = *(*[]v1alpha2.PodScheduling)(unsafe.Pointer(&in.Items))
|
||||
out.Items = *(*[]v1alpha2.PodSchedulingContext)(unsafe.Pointer(&in.Items))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_resource_PodSchedulingList_To_v1alpha2_PodSchedulingList is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingList_To_v1alpha2_PodSchedulingList(in *resource.PodSchedulingList, out *v1alpha2.PodSchedulingList, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingList_To_v1alpha2_PodSchedulingList(in, out, s)
|
||||
// Convert_resource_PodSchedulingContextList_To_v1alpha2_PodSchedulingContextList is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingContextList_To_v1alpha2_PodSchedulingContextList(in *resource.PodSchedulingContextList, out *v1alpha2.PodSchedulingContextList, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingContextList_To_v1alpha2_PodSchedulingContextList(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec(in *v1alpha2.PodSchedulingSpec, out *resource.PodSchedulingSpec, s conversion.Scope) error {
|
||||
func autoConvert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec(in *v1alpha2.PodSchedulingContextSpec, out *resource.PodSchedulingContextSpec, s conversion.Scope) error {
|
||||
out.SelectedNode = in.SelectedNode
|
||||
out.PotentialNodes = *(*[]string)(unsafe.Pointer(&in.PotentialNodes))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec(in *v1alpha2.PodSchedulingSpec, out *resource.PodSchedulingSpec, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingSpec_To_resource_PodSchedulingSpec(in, out, s)
|
||||
// Convert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec(in *v1alpha2.PodSchedulingContextSpec, out *resource.PodSchedulingContextSpec, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingContextSpec_To_resource_PodSchedulingContextSpec(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec(in *resource.PodSchedulingSpec, out *v1alpha2.PodSchedulingSpec, s conversion.Scope) error {
|
||||
func autoConvert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec(in *resource.PodSchedulingContextSpec, out *v1alpha2.PodSchedulingContextSpec, s conversion.Scope) error {
|
||||
out.SelectedNode = in.SelectedNode
|
||||
out.PotentialNodes = *(*[]string)(unsafe.Pointer(&in.PotentialNodes))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec(in *resource.PodSchedulingSpec, out *v1alpha2.PodSchedulingSpec, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingSpec_To_v1alpha2_PodSchedulingSpec(in, out, s)
|
||||
// Convert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec(in *resource.PodSchedulingContextSpec, out *v1alpha2.PodSchedulingContextSpec, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingContextSpec_To_v1alpha2_PodSchedulingContextSpec(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus(in *v1alpha2.PodSchedulingStatus, out *resource.PodSchedulingStatus, s conversion.Scope) error {
|
||||
func autoConvert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus(in *v1alpha2.PodSchedulingContextStatus, out *resource.PodSchedulingContextStatus, s conversion.Scope) error {
|
||||
out.ResourceClaims = *(*[]resource.ResourceClaimSchedulingStatus)(unsafe.Pointer(&in.ResourceClaims))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus(in *v1alpha2.PodSchedulingStatus, out *resource.PodSchedulingStatus, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingStatus_To_resource_PodSchedulingStatus(in, out, s)
|
||||
// Convert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus is an autogenerated conversion function.
|
||||
func Convert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus(in *v1alpha2.PodSchedulingContextStatus, out *resource.PodSchedulingContextStatus, s conversion.Scope) error {
|
||||
return autoConvert_v1alpha2_PodSchedulingContextStatus_To_resource_PodSchedulingContextStatus(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus(in *resource.PodSchedulingStatus, out *v1alpha2.PodSchedulingStatus, s conversion.Scope) error {
|
||||
func autoConvert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus(in *resource.PodSchedulingContextStatus, out *v1alpha2.PodSchedulingContextStatus, s conversion.Scope) error {
|
||||
out.ResourceClaims = *(*[]v1alpha2.ResourceClaimSchedulingStatus)(unsafe.Pointer(&in.ResourceClaims))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Convert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus(in *resource.PodSchedulingStatus, out *v1alpha2.PodSchedulingStatus, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingStatus_To_v1alpha2_PodSchedulingStatus(in, out, s)
|
||||
// Convert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus is an autogenerated conversion function.
|
||||
func Convert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus(in *resource.PodSchedulingContextStatus, out *v1alpha2.PodSchedulingContextStatus, s conversion.Scope) error {
|
||||
return autoConvert_resource_PodSchedulingContextStatus_To_v1alpha2_PodSchedulingContextStatus(in, out, s)
|
||||
}
|
||||
|
||||
func autoConvert_v1alpha2_ResourceClaim_To_resource_ResourceClaim(in *v1alpha2.ResourceClaim, out *resource.ResourceClaim, s conversion.Scope) error {
|
||||
|
38
pkg/apis/resource/zz_generated.deepcopy.go
generated
38
pkg/apis/resource/zz_generated.deepcopy.go
generated
@ -48,7 +48,7 @@ func (in *AllocationResult) DeepCopy() *AllocationResult {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodScheduling) DeepCopyInto(out *PodScheduling) {
|
||||
func (in *PodSchedulingContext) DeepCopyInto(out *PodSchedulingContext) {
|
||||
*out = *in
|
||||
out.TypeMeta = in.TypeMeta
|
||||
in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
|
||||
@ -57,18 +57,18 @@ func (in *PodScheduling) DeepCopyInto(out *PodScheduling) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodScheduling.
|
||||
func (in *PodScheduling) DeepCopy() *PodScheduling {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContext.
|
||||
func (in *PodSchedulingContext) DeepCopy() *PodSchedulingContext {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodScheduling)
|
||||
out := new(PodSchedulingContext)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
||||
func (in *PodScheduling) DeepCopyObject() runtime.Object {
|
||||
func (in *PodSchedulingContext) DeepCopyObject() runtime.Object {
|
||||
if c := in.DeepCopy(); c != nil {
|
||||
return c
|
||||
}
|
||||
@ -76,13 +76,13 @@ func (in *PodScheduling) DeepCopyObject() runtime.Object {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingList) DeepCopyInto(out *PodSchedulingList) {
|
||||
func (in *PodSchedulingContextList) DeepCopyInto(out *PodSchedulingContextList) {
|
||||
*out = *in
|
||||
out.TypeMeta = in.TypeMeta
|
||||
in.ListMeta.DeepCopyInto(&out.ListMeta)
|
||||
if in.Items != nil {
|
||||
in, out := &in.Items, &out.Items
|
||||
*out = make([]PodScheduling, len(*in))
|
||||
*out = make([]PodSchedulingContext, len(*in))
|
||||
for i := range *in {
|
||||
(*in)[i].DeepCopyInto(&(*out)[i])
|
||||
}
|
||||
@ -90,18 +90,18 @@ func (in *PodSchedulingList) DeepCopyInto(out *PodSchedulingList) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingList.
|
||||
func (in *PodSchedulingList) DeepCopy() *PodSchedulingList {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextList.
|
||||
func (in *PodSchedulingContextList) DeepCopy() *PodSchedulingContextList {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingList)
|
||||
out := new(PodSchedulingContextList)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
||||
func (in *PodSchedulingList) DeepCopyObject() runtime.Object {
|
||||
func (in *PodSchedulingContextList) DeepCopyObject() runtime.Object {
|
||||
if c := in.DeepCopy(); c != nil {
|
||||
return c
|
||||
}
|
||||
@ -109,7 +109,7 @@ func (in *PodSchedulingList) DeepCopyObject() runtime.Object {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingSpec) DeepCopyInto(out *PodSchedulingSpec) {
|
||||
func (in *PodSchedulingContextSpec) DeepCopyInto(out *PodSchedulingContextSpec) {
|
||||
*out = *in
|
||||
if in.PotentialNodes != nil {
|
||||
in, out := &in.PotentialNodes, &out.PotentialNodes
|
||||
@ -119,18 +119,18 @@ func (in *PodSchedulingSpec) DeepCopyInto(out *PodSchedulingSpec) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingSpec.
|
||||
func (in *PodSchedulingSpec) DeepCopy() *PodSchedulingSpec {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextSpec.
|
||||
func (in *PodSchedulingContextSpec) DeepCopy() *PodSchedulingContextSpec {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingSpec)
|
||||
out := new(PodSchedulingContextSpec)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingStatus) DeepCopyInto(out *PodSchedulingStatus) {
|
||||
func (in *PodSchedulingContextStatus) DeepCopyInto(out *PodSchedulingContextStatus) {
|
||||
*out = *in
|
||||
if in.ResourceClaims != nil {
|
||||
in, out := &in.ResourceClaims, &out.ResourceClaims
|
||||
@ -142,12 +142,12 @@ func (in *PodSchedulingStatus) DeepCopyInto(out *PodSchedulingStatus) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingStatus.
|
||||
func (in *PodSchedulingStatus) DeepCopy() *PodSchedulingStatus {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextStatus.
|
||||
func (in *PodSchedulingContextStatus) DeepCopy() *PodSchedulingContextStatus {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingStatus)
|
||||
out := new(PodSchedulingContextStatus)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
36
pkg/generated/openapi/zz_generated.openapi.go
generated
36
pkg/generated/openapi/zz_generated.openapi.go
generated
@ -825,10 +825,10 @@ func GetOpenAPIDefinitions(ref common.ReferenceCallback) map[string]common.OpenA
|
||||
"k8s.io/api/rbac/v1beta1.RoleRef": schema_k8sio_api_rbac_v1beta1_RoleRef(ref),
|
||||
"k8s.io/api/rbac/v1beta1.Subject": schema_k8sio_api_rbac_v1beta1_Subject(ref),
|
||||
"k8s.io/api/resource/v1alpha2.AllocationResult": schema_k8sio_api_resource_v1alpha2_AllocationResult(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodScheduling": schema_k8sio_api_resource_v1alpha2_PodScheduling(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingList": schema_k8sio_api_resource_v1alpha2_PodSchedulingList(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingSpec": schema_k8sio_api_resource_v1alpha2_PodSchedulingSpec(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingStatus": schema_k8sio_api_resource_v1alpha2_PodSchedulingStatus(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContext": schema_k8sio_api_resource_v1alpha2_PodSchedulingContext(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContextList": schema_k8sio_api_resource_v1alpha2_PodSchedulingContextList(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContextSpec": schema_k8sio_api_resource_v1alpha2_PodSchedulingContextSpec(ref),
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContextStatus": schema_k8sio_api_resource_v1alpha2_PodSchedulingContextStatus(ref),
|
||||
"k8s.io/api/resource/v1alpha2.ResourceClaim": schema_k8sio_api_resource_v1alpha2_ResourceClaim(ref),
|
||||
"k8s.io/api/resource/v1alpha2.ResourceClaimConsumerReference": schema_k8sio_api_resource_v1alpha2_ResourceClaimConsumerReference(ref),
|
||||
"k8s.io/api/resource/v1alpha2.ResourceClaimList": schema_k8sio_api_resource_v1alpha2_ResourceClaimList(ref),
|
||||
@ -41156,11 +41156,11 @@ func schema_k8sio_api_resource_v1alpha2_AllocationResult(ref common.ReferenceCal
|
||||
}
|
||||
}
|
||||
|
||||
func schema_k8sio_api_resource_v1alpha2_PodScheduling(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingContext(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
return common.OpenAPIDefinition{
|
||||
Schema: spec.Schema{
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "PodScheduling objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
Description: "PodSchedulingContext objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
Type: []string{"object"},
|
||||
Properties: map[string]spec.Schema{
|
||||
"kind": {
|
||||
@ -41188,14 +41188,14 @@ func schema_k8sio_api_resource_v1alpha2_PodScheduling(ref common.ReferenceCallba
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "Spec describes where resources for the Pod are needed.",
|
||||
Default: map[string]interface{}{},
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodSchedulingSpec"),
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodSchedulingContextSpec"),
|
||||
},
|
||||
},
|
||||
"status": {
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "Status describes where resources for the Pod can be allocated.",
|
||||
Default: map[string]interface{}{},
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodSchedulingStatus"),
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodSchedulingContextStatus"),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -41203,15 +41203,15 @@ func schema_k8sio_api_resource_v1alpha2_PodScheduling(ref common.ReferenceCallba
|
||||
},
|
||||
},
|
||||
Dependencies: []string{
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingSpec", "k8s.io/api/resource/v1alpha2.PodSchedulingStatus", "k8s.io/apimachinery/pkg/apis/meta/v1.ObjectMeta"},
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContextSpec", "k8s.io/api/resource/v1alpha2.PodSchedulingContextStatus", "k8s.io/apimachinery/pkg/apis/meta/v1.ObjectMeta"},
|
||||
}
|
||||
}
|
||||
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingList(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingContextList(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
return common.OpenAPIDefinition{
|
||||
Schema: spec.Schema{
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "PodSchedulingList is a collection of Pod scheduling objects.",
|
||||
Description: "PodSchedulingContextList is a collection of Pod scheduling objects.",
|
||||
Type: []string{"object"},
|
||||
Properties: map[string]spec.Schema{
|
||||
"kind": {
|
||||
@ -41237,13 +41237,13 @@ func schema_k8sio_api_resource_v1alpha2_PodSchedulingList(ref common.ReferenceCa
|
||||
},
|
||||
"items": {
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "Items is the list of PodScheduling objects.",
|
||||
Description: "Items is the list of PodSchedulingContext objects.",
|
||||
Type: []string{"array"},
|
||||
Items: &spec.SchemaOrArray{
|
||||
Schema: &spec.Schema{
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Default: map[string]interface{}{},
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodScheduling"),
|
||||
Ref: ref("k8s.io/api/resource/v1alpha2.PodSchedulingContext"),
|
||||
},
|
||||
},
|
||||
},
|
||||
@ -41254,15 +41254,15 @@ func schema_k8sio_api_resource_v1alpha2_PodSchedulingList(ref common.ReferenceCa
|
||||
},
|
||||
},
|
||||
Dependencies: []string{
|
||||
"k8s.io/api/resource/v1alpha2.PodScheduling", "k8s.io/apimachinery/pkg/apis/meta/v1.ListMeta"},
|
||||
"k8s.io/api/resource/v1alpha2.PodSchedulingContext", "k8s.io/apimachinery/pkg/apis/meta/v1.ListMeta"},
|
||||
}
|
||||
}
|
||||
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingSpec(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingContextSpec(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
return common.OpenAPIDefinition{
|
||||
Schema: spec.Schema{
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "PodSchedulingSpec describes where resources for the Pod are needed.",
|
||||
Description: "PodSchedulingContextSpec describes where resources for the Pod are needed.",
|
||||
Type: []string{"object"},
|
||||
Properties: map[string]spec.Schema{
|
||||
"selectedNode": {
|
||||
@ -41298,11 +41298,11 @@ func schema_k8sio_api_resource_v1alpha2_PodSchedulingSpec(ref common.ReferenceCa
|
||||
}
|
||||
}
|
||||
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingStatus(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
func schema_k8sio_api_resource_v1alpha2_PodSchedulingContextStatus(ref common.ReferenceCallback) common.OpenAPIDefinition {
|
||||
return common.OpenAPIDefinition{
|
||||
Schema: spec.Schema{
|
||||
SchemaProps: spec.SchemaProps{
|
||||
Description: "PodSchedulingStatus describes where resources for the Pod can be allocated.",
|
||||
Description: "PodSchedulingContextStatus describes where resources for the Pod can be allocated.",
|
||||
Type: []string{"object"},
|
||||
Properties: map[string]spec.Schema{
|
||||
"resourceClaims": {
|
||||
|
330
staging/src/k8s.io/api/resource/v1alpha2/generated.pb.go
generated
330
staging/src/k8s.io/api/resource/v1alpha2/generated.pb.go
generated
@ -74,15 +74,15 @@ func (m *AllocationResult) XXX_DiscardUnknown() {
|
||||
|
||||
var xxx_messageInfo_AllocationResult proto.InternalMessageInfo
|
||||
|
||||
func (m *PodScheduling) Reset() { *m = PodScheduling{} }
|
||||
func (*PodScheduling) ProtoMessage() {}
|
||||
func (*PodScheduling) Descriptor() ([]byte, []int) {
|
||||
func (m *PodSchedulingContext) Reset() { *m = PodSchedulingContext{} }
|
||||
func (*PodSchedulingContext) ProtoMessage() {}
|
||||
func (*PodSchedulingContext) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_3add37bbd52889e0, []int{1}
|
||||
}
|
||||
func (m *PodScheduling) XXX_Unmarshal(b []byte) error {
|
||||
func (m *PodSchedulingContext) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *PodScheduling) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
func (m *PodSchedulingContext) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
@ -90,27 +90,27 @@ func (m *PodScheduling) XXX_Marshal(b []byte, deterministic bool) ([]byte, error
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
func (m *PodScheduling) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodScheduling.Merge(m, src)
|
||||
func (m *PodSchedulingContext) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingContext.Merge(m, src)
|
||||
}
|
||||
func (m *PodScheduling) XXX_Size() int {
|
||||
func (m *PodSchedulingContext) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *PodScheduling) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodScheduling.DiscardUnknown(m)
|
||||
func (m *PodSchedulingContext) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingContext.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_PodScheduling proto.InternalMessageInfo
|
||||
var xxx_messageInfo_PodSchedulingContext proto.InternalMessageInfo
|
||||
|
||||
func (m *PodSchedulingList) Reset() { *m = PodSchedulingList{} }
|
||||
func (*PodSchedulingList) ProtoMessage() {}
|
||||
func (*PodSchedulingList) Descriptor() ([]byte, []int) {
|
||||
func (m *PodSchedulingContextList) Reset() { *m = PodSchedulingContextList{} }
|
||||
func (*PodSchedulingContextList) ProtoMessage() {}
|
||||
func (*PodSchedulingContextList) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_3add37bbd52889e0, []int{2}
|
||||
}
|
||||
func (m *PodSchedulingList) XXX_Unmarshal(b []byte) error {
|
||||
func (m *PodSchedulingContextList) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *PodSchedulingList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
func (m *PodSchedulingContextList) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
@ -118,27 +118,27 @@ func (m *PodSchedulingList) XXX_Marshal(b []byte, deterministic bool) ([]byte, e
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
func (m *PodSchedulingList) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingList.Merge(m, src)
|
||||
func (m *PodSchedulingContextList) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingContextList.Merge(m, src)
|
||||
}
|
||||
func (m *PodSchedulingList) XXX_Size() int {
|
||||
func (m *PodSchedulingContextList) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *PodSchedulingList) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingList.DiscardUnknown(m)
|
||||
func (m *PodSchedulingContextList) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingContextList.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_PodSchedulingList proto.InternalMessageInfo
|
||||
var xxx_messageInfo_PodSchedulingContextList proto.InternalMessageInfo
|
||||
|
||||
func (m *PodSchedulingSpec) Reset() { *m = PodSchedulingSpec{} }
|
||||
func (*PodSchedulingSpec) ProtoMessage() {}
|
||||
func (*PodSchedulingSpec) Descriptor() ([]byte, []int) {
|
||||
func (m *PodSchedulingContextSpec) Reset() { *m = PodSchedulingContextSpec{} }
|
||||
func (*PodSchedulingContextSpec) ProtoMessage() {}
|
||||
func (*PodSchedulingContextSpec) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_3add37bbd52889e0, []int{3}
|
||||
}
|
||||
func (m *PodSchedulingSpec) XXX_Unmarshal(b []byte) error {
|
||||
func (m *PodSchedulingContextSpec) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *PodSchedulingSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
func (m *PodSchedulingContextSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
@ -146,27 +146,27 @@ func (m *PodSchedulingSpec) XXX_Marshal(b []byte, deterministic bool) ([]byte, e
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
func (m *PodSchedulingSpec) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingSpec.Merge(m, src)
|
||||
func (m *PodSchedulingContextSpec) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingContextSpec.Merge(m, src)
|
||||
}
|
||||
func (m *PodSchedulingSpec) XXX_Size() int {
|
||||
func (m *PodSchedulingContextSpec) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *PodSchedulingSpec) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingSpec.DiscardUnknown(m)
|
||||
func (m *PodSchedulingContextSpec) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingContextSpec.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_PodSchedulingSpec proto.InternalMessageInfo
|
||||
var xxx_messageInfo_PodSchedulingContextSpec proto.InternalMessageInfo
|
||||
|
||||
func (m *PodSchedulingStatus) Reset() { *m = PodSchedulingStatus{} }
|
||||
func (*PodSchedulingStatus) ProtoMessage() {}
|
||||
func (*PodSchedulingStatus) Descriptor() ([]byte, []int) {
|
||||
func (m *PodSchedulingContextStatus) Reset() { *m = PodSchedulingContextStatus{} }
|
||||
func (*PodSchedulingContextStatus) ProtoMessage() {}
|
||||
func (*PodSchedulingContextStatus) Descriptor() ([]byte, []int) {
|
||||
return fileDescriptor_3add37bbd52889e0, []int{4}
|
||||
}
|
||||
func (m *PodSchedulingStatus) XXX_Unmarshal(b []byte) error {
|
||||
func (m *PodSchedulingContextStatus) XXX_Unmarshal(b []byte) error {
|
||||
return m.Unmarshal(b)
|
||||
}
|
||||
func (m *PodSchedulingStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
func (m *PodSchedulingContextStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) {
|
||||
b = b[:cap(b)]
|
||||
n, err := m.MarshalToSizedBuffer(b)
|
||||
if err != nil {
|
||||
@ -174,17 +174,17 @@ func (m *PodSchedulingStatus) XXX_Marshal(b []byte, deterministic bool) ([]byte,
|
||||
}
|
||||
return b[:n], nil
|
||||
}
|
||||
func (m *PodSchedulingStatus) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingStatus.Merge(m, src)
|
||||
func (m *PodSchedulingContextStatus) XXX_Merge(src proto.Message) {
|
||||
xxx_messageInfo_PodSchedulingContextStatus.Merge(m, src)
|
||||
}
|
||||
func (m *PodSchedulingStatus) XXX_Size() int {
|
||||
func (m *PodSchedulingContextStatus) XXX_Size() int {
|
||||
return m.Size()
|
||||
}
|
||||
func (m *PodSchedulingStatus) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingStatus.DiscardUnknown(m)
|
||||
func (m *PodSchedulingContextStatus) XXX_DiscardUnknown() {
|
||||
xxx_messageInfo_PodSchedulingContextStatus.DiscardUnknown(m)
|
||||
}
|
||||
|
||||
var xxx_messageInfo_PodSchedulingStatus proto.InternalMessageInfo
|
||||
var xxx_messageInfo_PodSchedulingContextStatus proto.InternalMessageInfo
|
||||
|
||||
func (m *ResourceClaim) Reset() { *m = ResourceClaim{} }
|
||||
func (*ResourceClaim) ProtoMessage() {}
|
||||
@ -552,10 +552,10 @@ var xxx_messageInfo_ResourceClassParametersReference proto.InternalMessageInfo
|
||||
|
||||
func init() {
|
||||
proto.RegisterType((*AllocationResult)(nil), "k8s.io.api.resource.v1alpha2.AllocationResult")
|
||||
proto.RegisterType((*PodScheduling)(nil), "k8s.io.api.resource.v1alpha2.PodScheduling")
|
||||
proto.RegisterType((*PodSchedulingList)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingList")
|
||||
proto.RegisterType((*PodSchedulingSpec)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingSpec")
|
||||
proto.RegisterType((*PodSchedulingStatus)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingStatus")
|
||||
proto.RegisterType((*PodSchedulingContext)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingContext")
|
||||
proto.RegisterType((*PodSchedulingContextList)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingContextList")
|
||||
proto.RegisterType((*PodSchedulingContextSpec)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingContextSpec")
|
||||
proto.RegisterType((*PodSchedulingContextStatus)(nil), "k8s.io.api.resource.v1alpha2.PodSchedulingContextStatus")
|
||||
proto.RegisterType((*ResourceClaim)(nil), "k8s.io.api.resource.v1alpha2.ResourceClaim")
|
||||
proto.RegisterType((*ResourceClaimConsumerReference)(nil), "k8s.io.api.resource.v1alpha2.ResourceClaimConsumerReference")
|
||||
proto.RegisterType((*ResourceClaimList)(nil), "k8s.io.api.resource.v1alpha2.ResourceClaimList")
|
||||
@ -576,81 +576,83 @@ func init() {
|
||||
}
|
||||
|
||||
var fileDescriptor_3add37bbd52889e0 = []byte{
|
||||
// 1174 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x58, 0xcd, 0x6f, 0x1b, 0x45,
|
||||
0x14, 0xcf, 0xc6, 0x6e, 0x95, 0x8c, 0x1b, 0x37, 0xd9, 0xb4, 0xc8, 0xad, 0x5a, 0xdb, 0xec, 0xc9,
|
||||
0x12, 0xb0, 0xdb, 0x18, 0x04, 0x15, 0x1f, 0x95, 0xb2, 0x0d, 0x94, 0x08, 0x9a, 0x9a, 0x31, 0x91,
|
||||
0x08, 0x42, 0x88, 0xf1, 0xee, 0xab, 0xbd, 0x64, 0xbf, 0xd8, 0xd9, 0x35, 0xaa, 0xb8, 0xf4, 0xca,
|
||||
0x0d, 0x21, 0xee, 0x1c, 0xf9, 0x43, 0x10, 0x52, 0x8e, 0x91, 0xe0, 0xd0, 0x93, 0x45, 0xcc, 0x81,
|
||||
0x3f, 0x80, 0x13, 0x3d, 0xa1, 0x19, 0xef, 0xae, 0x77, 0xd6, 0x1f, 0xc4, 0x11, 0x8a, 0xc2, 0x29,
|
||||
0x99, 0x79, 0xbf, 0xf7, 0x9b, 0xf7, 0x31, 0xef, 0xcd, 0x5b, 0xa3, 0x77, 0x0f, 0xef, 0x52, 0xd5,
|
||||
0xf2, 0xb4, 0xc3, 0xa8, 0x03, 0x81, 0x0b, 0x21, 0x50, 0xad, 0x0f, 0xae, 0xe9, 0x05, 0x5a, 0x2c,
|
||||
0x20, 0xbe, 0xa5, 0x05, 0x40, 0xbd, 0x28, 0x30, 0x40, 0xeb, 0x6f, 0x11, 0xdb, 0xef, 0x91, 0xa6,
|
||||
0xd6, 0x05, 0x17, 0x02, 0x12, 0x82, 0xa9, 0xfa, 0x81, 0x17, 0x7a, 0xf2, 0xad, 0x11, 0x5a, 0x25,
|
||||
0xbe, 0xa5, 0x26, 0x68, 0x35, 0x41, 0xdf, 0x7c, 0xa5, 0x6b, 0x85, 0xbd, 0xa8, 0xa3, 0x1a, 0x9e,
|
||||
0xa3, 0x75, 0xbd, 0xae, 0xa7, 0x71, 0xa5, 0x4e, 0xf4, 0x98, 0xaf, 0xf8, 0x82, 0xff, 0x37, 0x22,
|
||||
0xbb, 0xa9, 0x64, 0x8e, 0x36, 0xbc, 0x80, 0x1d, 0x9b, 0x3f, 0xf0, 0xe6, 0x6b, 0x63, 0x8c, 0x43,
|
||||
0x8c, 0x9e, 0xe5, 0x42, 0xf0, 0x44, 0xf3, 0x0f, 0xbb, 0x6c, 0x83, 0x6a, 0x0e, 0x84, 0x64, 0x9a,
|
||||
0x96, 0x36, 0x4b, 0x2b, 0x88, 0xdc, 0xd0, 0x72, 0x60, 0x42, 0xe1, 0xf5, 0x7f, 0x53, 0xa0, 0x46,
|
||||
0x0f, 0x1c, 0x92, 0xd7, 0x53, 0xfe, 0x94, 0xd0, 0xfa, 0xb6, 0x6d, 0x7b, 0x06, 0x09, 0x2d, 0xcf,
|
||||
0xc5, 0x40, 0x23, 0x3b, 0x94, 0xef, 0xa1, 0x72, 0x12, 0x9b, 0xf7, 0x89, 0x6b, 0xda, 0x50, 0x91,
|
||||
0xea, 0x52, 0x63, 0x55, 0x7f, 0xe1, 0x68, 0x50, 0x5b, 0x1a, 0x0e, 0x6a, 0x65, 0x2c, 0x48, 0x71,
|
||||
0x0e, 0x2d, 0x77, 0xd0, 0x3a, 0xe9, 0x13, 0xcb, 0x26, 0x1d, 0x1b, 0x1e, 0xb9, 0x7b, 0x9e, 0x09,
|
||||
0xb4, 0xb2, 0x5c, 0x97, 0x1a, 0xa5, 0x66, 0x5d, 0xcd, 0xc4, 0x9f, 0x85, 0x4c, 0xed, 0x6f, 0xa9,
|
||||
0x0c, 0xd0, 0x06, 0x1b, 0x8c, 0xd0, 0x0b, 0xf4, 0x6b, 0xc3, 0x41, 0x6d, 0x7d, 0x3b, 0xa7, 0x8d,
|
||||
0x27, 0xf8, 0x64, 0x0d, 0xad, 0xd2, 0x1e, 0x09, 0x80, 0xed, 0x55, 0x0a, 0x75, 0xa9, 0xb1, 0xa2,
|
||||
0x6f, 0xc4, 0xe6, 0xad, 0xb6, 0x13, 0x01, 0x1e, 0x63, 0x94, 0x1f, 0x97, 0xd1, 0x5a, 0xcb, 0x33,
|
||||
0xdb, 0x46, 0x0f, 0xcc, 0xc8, 0xb6, 0xdc, 0xae, 0xfc, 0x05, 0x5a, 0x61, 0xf1, 0x37, 0x49, 0x48,
|
||||
0xb8, 0x83, 0xa5, 0xe6, 0x9d, 0x8c, 0x79, 0x69, 0x18, 0x55, 0xff, 0xb0, 0xcb, 0x36, 0xa8, 0xca,
|
||||
0xd0, 0xcc, 0xe0, 0x47, 0x9d, 0x2f, 0xc1, 0x08, 0x1f, 0x42, 0x48, 0x74, 0x39, 0x3e, 0x13, 0x8d,
|
||||
0xf7, 0x70, 0xca, 0x2a, 0x7f, 0x84, 0x8a, 0xd4, 0x07, 0x23, 0x76, 0x5e, 0x53, 0xe7, 0x5d, 0x3e,
|
||||
0x55, 0x30, 0xae, 0xed, 0x83, 0xa1, 0x5f, 0x89, 0xc9, 0x8b, 0x6c, 0x85, 0x39, 0x95, 0x7c, 0x80,
|
||||
0x2e, 0xd3, 0x90, 0x84, 0x11, 0xe5, 0x4e, 0x97, 0x9a, 0x5b, 0x8b, 0x90, 0x72, 0x45, 0xbd, 0x1c,
|
||||
0xd3, 0x5e, 0x1e, 0xad, 0x71, 0x4c, 0xa8, 0xfc, 0x2c, 0xa1, 0x0d, 0x01, 0xff, 0xa1, 0x45, 0x43,
|
||||
0xf9, 0xb3, 0x89, 0x28, 0xa9, 0xa7, 0x8b, 0x12, 0xd3, 0xe6, 0x31, 0x5a, 0x8f, 0xcf, 0x5b, 0x49,
|
||||
0x76, 0x32, 0x11, 0x6a, 0xa1, 0x4b, 0x56, 0x08, 0x0e, 0xbb, 0x1f, 0x85, 0x46, 0xa9, 0xf9, 0xd2,
|
||||
0x02, 0xde, 0xe8, 0x6b, 0x31, 0xef, 0xa5, 0x5d, 0xc6, 0x80, 0x47, 0x44, 0xca, 0xb7, 0x79, 0x2f,
|
||||
0x58, 0xf0, 0xe4, 0xbb, 0xe8, 0x0a, 0xe5, 0x57, 0x0c, 0x4c, 0x76, 0x7f, 0xe2, 0x0b, 0x7d, 0x2d,
|
||||
0x66, 0xb8, 0xd2, 0xce, 0xc8, 0xb0, 0x80, 0x94, 0xdf, 0x44, 0x65, 0xdf, 0x0b, 0xc1, 0x0d, 0x2d,
|
||||
0x62, 0x27, 0x57, 0xb9, 0xd0, 0x58, 0xd5, 0x65, 0x56, 0x08, 0x2d, 0x41, 0x82, 0x73, 0x48, 0xe5,
|
||||
0x7b, 0x09, 0x6d, 0x4e, 0xc9, 0x80, 0xfc, 0xcd, 0xb8, 0xc0, 0xee, 0xdb, 0xc4, 0x72, 0x68, 0x45,
|
||||
0xe2, 0xee, 0xbf, 0x35, 0xdf, 0x7d, 0x9c, 0xd5, 0x99, 0x48, 0xeb, 0x44, 0x75, 0x8e, 0xa8, 0x71,
|
||||
0xee, 0x28, 0x5e, 0x08, 0x02, 0xe4, 0xa2, 0x15, 0x82, 0xe8, 0xe6, 0x7f, 0x54, 0x08, 0x22, 0xe9,
|
||||
0xfc, 0x42, 0x18, 0x4a, 0xa8, 0x2a, 0xe0, 0xef, 0x7b, 0x2e, 0x8d, 0x1c, 0x08, 0x30, 0x3c, 0x86,
|
||||
0x00, 0x5c, 0x03, 0xe4, 0x97, 0xd1, 0x0a, 0xf1, 0xad, 0x07, 0x81, 0x17, 0xf9, 0xf1, 0x5d, 0x4a,
|
||||
0x6f, 0xf9, 0x76, 0x6b, 0x97, 0xef, 0xe3, 0x14, 0xc1, 0xd0, 0x89, 0x45, 0xdc, 0xda, 0x0c, 0x3a,
|
||||
0x39, 0x07, 0xa7, 0x08, 0xb9, 0x8e, 0x8a, 0x2e, 0x71, 0xa0, 0x52, 0xe4, 0xc8, 0xd4, 0xf7, 0x3d,
|
||||
0xe2, 0x00, 0xe6, 0x12, 0x59, 0x47, 0x85, 0xc8, 0x32, 0x2b, 0x97, 0x38, 0xe0, 0x4e, 0x0c, 0x28,
|
||||
0xec, 0xef, 0xee, 0x3c, 0x1f, 0xd4, 0x5e, 0x9c, 0xf5, 0x12, 0x84, 0x4f, 0x7c, 0xa0, 0xea, 0xfe,
|
||||
0xee, 0x0e, 0x66, 0xca, 0xbc, 0xda, 0x05, 0x27, 0x2f, 0x5c, 0xb5, 0x0b, 0xd6, 0xcd, 0xa8, 0xf6,
|
||||
0x1f, 0x24, 0x54, 0x17, 0x70, 0x2d, 0x12, 0x10, 0x07, 0x42, 0x08, 0xe8, 0x59, 0x93, 0x55, 0x47,
|
||||
0xc5, 0x43, 0xcb, 0x35, 0xf9, 0x5d, 0xcd, 0x84, 0xff, 0x03, 0xcb, 0x35, 0x31, 0x97, 0xa4, 0x09,
|
||||
0x2a, 0xcc, 0x4a, 0x90, 0xf2, 0x54, 0x42, 0xb7, 0xe7, 0x56, 0x6b, 0xca, 0x21, 0xcd, 0x4c, 0xf2,
|
||||
0x3b, 0xe8, 0x6a, 0xe4, 0xd2, 0xc8, 0x0a, 0xd9, 0xf3, 0x95, 0xed, 0x3c, 0x9b, 0xc3, 0x41, 0xed,
|
||||
0xea, 0xbe, 0x28, 0xc2, 0x79, 0xac, 0xf2, 0xd3, 0x72, 0x2e, 0xbf, 0xbc, 0x0f, 0x3e, 0x40, 0x1b,
|
||||
0x99, 0x76, 0x40, 0xe9, 0xde, 0xd8, 0x86, 0x1b, 0xb1, 0x0d, 0x59, 0xad, 0x11, 0x00, 0x4f, 0xea,
|
||||
0xc8, 0x5f, 0xa3, 0x35, 0x3f, 0x1b, 0xea, 0xb8, 0xb4, 0xef, 0x2d, 0x90, 0xd2, 0x29, 0xa9, 0xd2,
|
||||
0x37, 0x86, 0x83, 0xda, 0x9a, 0x20, 0xc0, 0xe2, 0x39, 0x72, 0x0b, 0x95, 0x49, 0x3a, 0xb0, 0x3c,
|
||||
0x64, 0xbd, 0x7c, 0x94, 0x86, 0x46, 0xd2, 0xfe, 0xb6, 0x05, 0xe9, 0xf3, 0x89, 0x1d, 0x9c, 0xd3,
|
||||
0x57, 0xfe, 0x5a, 0x46, 0x9b, 0x53, 0xda, 0x83, 0xdc, 0x44, 0xc8, 0x0c, 0xac, 0x3e, 0x04, 0x99,
|
||||
0x20, 0xa5, 0x6d, 0x6e, 0x27, 0x95, 0xe0, 0x0c, 0x4a, 0xfe, 0x1c, 0xa1, 0x31, 0x7b, 0x1c, 0x13,
|
||||
0x75, 0x7e, 0x4c, 0xf2, 0xe3, 0x97, 0x5e, 0x66, 0xfc, 0x99, 0xdd, 0x0c, 0xa3, 0x4c, 0x51, 0x29,
|
||||
0x00, 0x0a, 0x41, 0x1f, 0xcc, 0xf7, 0xbc, 0xa0, 0x52, 0xe0, 0x75, 0xf4, 0xf6, 0x02, 0x41, 0x9f,
|
||||
0x68, 0x65, 0xfa, 0x66, 0xec, 0x52, 0x09, 0x8f, 0x89, 0x71, 0xf6, 0x14, 0xb9, 0x8d, 0xae, 0x9b,
|
||||
0x40, 0x32, 0x66, 0x7e, 0x15, 0x01, 0x0d, 0xc1, 0xe4, 0x1d, 0x6a, 0x45, 0xbf, 0x1d, 0x13, 0x5c,
|
||||
0xdf, 0x99, 0x06, 0xc2, 0xd3, 0x75, 0x95, 0xdf, 0x24, 0x74, 0x5d, 0xb0, 0xec, 0x63, 0x70, 0x7c,
|
||||
0x9b, 0x84, 0x70, 0x0e, 0xcf, 0xd1, 0x81, 0xf0, 0x1c, 0xbd, 0xb1, 0x40, 0xf8, 0x12, 0x23, 0x67,
|
||||
0x3d, 0x4b, 0xca, 0xaf, 0x12, 0xba, 0x31, 0x55, 0xe3, 0x1c, 0xda, 0xeb, 0x27, 0x62, 0x7b, 0x7d,
|
||||
0xf5, 0x0c, 0x7e, 0xcd, 0x68, 0xb3, 0xc7, 0xb3, 0xbc, 0xe2, 0x4d, 0xe5, 0xff, 0x38, 0x3f, 0x28,
|
||||
0x7f, 0x8b, 0x63, 0x10, 0xa5, 0xe7, 0xe0, 0x86, 0xd8, 0x51, 0x96, 0x4f, 0xd5, 0x51, 0x26, 0x1a,
|
||||
0x6d, 0x61, 0xc1, 0x46, 0x4b, 0xe9, 0xd9, 0x1a, 0xed, 0x01, 0x5a, 0x13, 0x5f, 0x9f, 0xe2, 0x29,
|
||||
0x3f, 0xe1, 0x38, 0x75, 0x5b, 0x78, 0x9d, 0x44, 0xa6, 0xfc, 0xec, 0x41, 0xe9, 0x45, 0x9e, 0x3d,
|
||||
0x28, 0x9d, 0x51, 0x14, 0xbf, 0x88, 0xb3, 0xc7, 0xd4, 0x38, 0x9f, 0xff, 0xec, 0xc1, 0xbe, 0x8c,
|
||||
0xd9, 0x5f, 0xea, 0x13, 0x23, 0x99, 0x21, 0xd3, 0x2f, 0xe3, 0xbd, 0x44, 0x80, 0xc7, 0x18, 0x5d,
|
||||
0x3f, 0x3a, 0xa9, 0x2e, 0x1d, 0x9f, 0x54, 0x97, 0x9e, 0x9d, 0x54, 0x97, 0x9e, 0x0e, 0xab, 0xd2,
|
||||
0xd1, 0xb0, 0x2a, 0x1d, 0x0f, 0xab, 0xd2, 0xb3, 0x61, 0x55, 0xfa, 0x7d, 0x58, 0x95, 0xbe, 0xfb,
|
||||
0xa3, 0xba, 0xf4, 0xe9, 0xad, 0x79, 0xbf, 0xb3, 0xfc, 0x13, 0x00, 0x00, 0xff, 0xff, 0x93, 0xfa,
|
||||
0x8c, 0x28, 0x9f, 0x11, 0x00, 0x00,
|
||||
// 1209 bytes of a gzipped FileDescriptorProto
|
||||
0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xd4, 0x58, 0xcf, 0x6e, 0xe3, 0xd4,
|
||||
0x17, 0xae, 0x93, 0x74, 0xd4, 0xde, 0xb4, 0x99, 0xd6, 0x6d, 0x7f, 0xca, 0x54, 0x33, 0x49, 0x7e,
|
||||
0x5e, 0x45, 0x02, 0xec, 0x69, 0x40, 0x43, 0xc5, 0x9f, 0x91, 0xea, 0x16, 0x86, 0x0a, 0xa6, 0x13,
|
||||
0x6e, 0xa8, 0x98, 0x22, 0x84, 0xe6, 0xc6, 0x3e, 0x93, 0x98, 0xfa, 0x1f, 0xbe, 0xd7, 0x81, 0x11,
|
||||
0x9b, 0x79, 0x84, 0x59, 0xb0, 0x61, 0xc5, 0x92, 0x17, 0xe0, 0x0d, 0x10, 0x52, 0x97, 0x45, 0xb0,
|
||||
0x98, 0x55, 0x44, 0xc3, 0x82, 0x07, 0x60, 0xc5, 0xac, 0x90, 0x1d, 0xdb, 0xb1, 0x9d, 0x38, 0x34,
|
||||
0x5d, 0x44, 0xb0, 0x9a, 0xf1, 0x3d, 0xdf, 0xf9, 0xee, 0xb9, 0xdf, 0xb9, 0xe7, 0xdc, 0x93, 0xa2,
|
||||
0x77, 0x4e, 0x77, 0xa9, 0xa8, 0x59, 0xd2, 0xa9, 0xdb, 0x06, 0xc7, 0x04, 0x06, 0x54, 0xea, 0x81,
|
||||
0xa9, 0x5a, 0x8e, 0x14, 0x18, 0x88, 0xad, 0x49, 0x0e, 0x50, 0xcb, 0x75, 0x14, 0x90, 0x7a, 0x3b,
|
||||
0x44, 0xb7, 0xbb, 0xa4, 0x21, 0x75, 0xc0, 0x04, 0x87, 0x30, 0x50, 0x45, 0xdb, 0xb1, 0x98, 0xc5,
|
||||
0xdf, 0x1c, 0xa2, 0x45, 0x62, 0x6b, 0x62, 0x88, 0x16, 0x43, 0xf4, 0xf6, 0x2b, 0x1d, 0x8d, 0x75,
|
||||
0xdd, 0xb6, 0xa8, 0x58, 0x86, 0xd4, 0xb1, 0x3a, 0x96, 0xe4, 0x3b, 0xb5, 0xdd, 0xc7, 0xfe, 0x97,
|
||||
0xff, 0xe1, 0xff, 0x6f, 0x48, 0xb6, 0x2d, 0xc4, 0xb6, 0x56, 0x2c, 0xc7, 0xdb, 0x36, 0xbd, 0xe1,
|
||||
0xf6, 0x6b, 0x23, 0x8c, 0x41, 0x94, 0xae, 0x66, 0x82, 0xf3, 0x44, 0xb2, 0x4f, 0x3b, 0xde, 0x02,
|
||||
0x95, 0x0c, 0x60, 0x64, 0x92, 0x97, 0x94, 0xe5, 0xe5, 0xb8, 0x26, 0xd3, 0x0c, 0x18, 0x73, 0xb8,
|
||||
0xf3, 0x4f, 0x0e, 0x54, 0xe9, 0x82, 0x41, 0xd2, 0x7e, 0xc2, 0x1f, 0x1c, 0x5a, 0xdb, 0xd3, 0x75,
|
||||
0x4b, 0x21, 0x4c, 0xb3, 0x4c, 0x0c, 0xd4, 0xd5, 0x19, 0x7f, 0x17, 0x95, 0x42, 0x6d, 0xde, 0x23,
|
||||
0xa6, 0xaa, 0x43, 0x99, 0xab, 0x71, 0xf5, 0x65, 0xf9, 0x7f, 0x67, 0xfd, 0xea, 0xc2, 0xa0, 0x5f,
|
||||
0x2d, 0xe1, 0x84, 0x15, 0xa7, 0xd0, 0x7c, 0x1b, 0xad, 0x91, 0x1e, 0xd1, 0x74, 0xd2, 0xd6, 0xe1,
|
||||
0x81, 0x79, 0x64, 0xa9, 0x40, 0xcb, 0xb9, 0x1a, 0x57, 0x2f, 0x36, 0x6a, 0x62, 0x4c, 0x7f, 0x4f,
|
||||
0x32, 0xb1, 0xb7, 0x23, 0x7a, 0x80, 0x16, 0xe8, 0xa0, 0x30, 0xcb, 0x91, 0x37, 0x07, 0xfd, 0xea,
|
||||
0xda, 0x5e, 0xca, 0x1b, 0x8f, 0xf1, 0xf1, 0x12, 0x5a, 0xa6, 0x5d, 0xe2, 0x80, 0xb7, 0x56, 0xce,
|
||||
0xd7, 0xb8, 0xfa, 0x92, 0xbc, 0x1e, 0x84, 0xb7, 0xdc, 0x0a, 0x0d, 0x78, 0x84, 0x11, 0x7e, 0xc8,
|
||||
0xa1, 0xcd, 0xa6, 0xa5, 0xb6, 0x94, 0x2e, 0xa8, 0xae, 0xae, 0x99, 0x9d, 0x7d, 0xcb, 0x64, 0xf0,
|
||||
0x15, 0xe3, 0x1f, 0xa1, 0x25, 0x2f, 0x0d, 0x2a, 0x61, 0xc4, 0x3f, 0x67, 0xb1, 0x71, 0x3b, 0x16,
|
||||
0x65, 0xa4, 0xa6, 0x68, 0x9f, 0x76, 0xbc, 0x05, 0x2a, 0x7a, 0x68, 0x2f, 0xee, 0x07, 0xed, 0xcf,
|
||||
0x41, 0x61, 0xf7, 0x81, 0x11, 0x99, 0x0f, 0xb6, 0x46, 0xa3, 0x35, 0x1c, 0xb1, 0xf2, 0x0f, 0x51,
|
||||
0x81, 0xda, 0xa0, 0x04, 0x1a, 0xdc, 0x11, 0xa7, 0xdd, 0x41, 0x71, 0x52, 0x8c, 0x2d, 0x1b, 0x14,
|
||||
0x79, 0x25, 0xd8, 0xa3, 0xe0, 0x7d, 0x61, 0x9f, 0x91, 0x7f, 0x84, 0xae, 0x51, 0x46, 0x98, 0x4b,
|
||||
0x7d, 0x09, 0x8a, 0x8d, 0xdd, 0x2b, 0x70, 0xfb, 0xfe, 0x72, 0x29, 0x60, 0xbf, 0x36, 0xfc, 0xc6,
|
||||
0x01, 0xaf, 0xf0, 0x33, 0x87, 0xca, 0x93, 0xdc, 0x3e, 0xd0, 0x28, 0xe3, 0x3f, 0x1d, 0x93, 0x4e,
|
||||
0xbc, 0x9c, 0x74, 0x9e, 0xb7, 0x2f, 0xdc, 0x5a, 0xb0, 0xed, 0x52, 0xb8, 0x12, 0x93, 0xed, 0x63,
|
||||
0xb4, 0xa8, 0x31, 0x30, 0xbc, 0xbb, 0x93, 0xaf, 0x17, 0x1b, 0x8d, 0xd9, 0xcf, 0x26, 0xaf, 0x06,
|
||||
0xf4, 0x8b, 0x87, 0x1e, 0x11, 0x1e, 0xf2, 0x09, 0xcf, 0x32, 0xce, 0xe4, 0x09, 0xcb, 0xef, 0xa2,
|
||||
0x15, 0xea, 0x5f, 0x46, 0x50, 0xbd, 0x9b, 0x16, 0x5c, 0xfd, 0xcd, 0x80, 0x68, 0xa5, 0x15, 0xb3,
|
||||
0xe1, 0x04, 0x92, 0x7f, 0x03, 0x95, 0x6c, 0x8b, 0x81, 0xc9, 0x34, 0xa2, 0x87, 0x97, 0x3e, 0x5f,
|
||||
0x5f, 0x96, 0x79, 0xaf, 0x64, 0x9a, 0x09, 0x0b, 0x4e, 0x21, 0x85, 0x6f, 0x39, 0xb4, 0x9d, 0x9d,
|
||||
0x1d, 0xfe, 0xeb, 0x51, 0x45, 0xee, 0xeb, 0x44, 0x33, 0x68, 0x99, 0xf3, 0x35, 0x79, 0x73, 0xba,
|
||||
0x26, 0x38, 0xee, 0x33, 0xe2, 0x0e, 0x52, 0x3e, 0x56, 0xce, 0x43, 0x6a, 0x9c, 0xda, 0x4a, 0xf8,
|
||||
0x2e, 0x87, 0x56, 0x13, 0x90, 0x39, 0x94, 0xcc, 0x87, 0x89, 0x92, 0x91, 0x66, 0x39, 0x66, 0x56,
|
||||
0xad, 0x9c, 0xa4, 0x6a, 0x65, 0x67, 0x16, 0xd2, 0xe9, 0x45, 0x32, 0xe0, 0x50, 0x25, 0x81, 0xdf,
|
||||
0xb7, 0x4c, 0xea, 0x1a, 0xe0, 0x60, 0x78, 0x0c, 0x0e, 0x98, 0x0a, 0xf0, 0x2f, 0xa3, 0x25, 0x62,
|
||||
0x6b, 0xf7, 0x1c, 0xcb, 0xb5, 0x83, 0x2b, 0x15, 0x5d, 0xfd, 0xbd, 0xe6, 0xa1, 0xbf, 0x8e, 0x23,
|
||||
0x84, 0x87, 0x0e, 0x23, 0xf2, 0xa3, 0x8d, 0xa1, 0xc3, 0x7d, 0x70, 0x84, 0xe0, 0x6b, 0xa8, 0x60,
|
||||
0x12, 0x03, 0xca, 0x05, 0x1f, 0x19, 0x9d, 0xfd, 0x88, 0x18, 0x80, 0x7d, 0x0b, 0x2f, 0xa3, 0xbc,
|
||||
0xab, 0xa9, 0xe5, 0x45, 0x1f, 0x70, 0x3b, 0x00, 0xe4, 0x8f, 0x0f, 0x0f, 0x5e, 0xf4, 0xab, 0xff,
|
||||
0xcf, 0x7a, 0x3a, 0xd8, 0x13, 0x1b, 0xa8, 0x78, 0x7c, 0x78, 0x80, 0x3d, 0x67, 0xe1, 0x47, 0x0e,
|
||||
0xad, 0x27, 0x0e, 0x39, 0x87, 0x16, 0xd0, 0x4c, 0xb6, 0x80, 0x97, 0x66, 0x48, 0x59, 0x46, 0xed,
|
||||
0x7f, 0xc3, 0xa1, 0x5a, 0x02, 0xd7, 0x24, 0x0e, 0x31, 0x80, 0x81, 0x43, 0xaf, 0x9a, 0xac, 0x1a,
|
||||
0x2a, 0x9c, 0x6a, 0xa6, 0xea, 0xdf, 0xd5, 0x98, 0xfc, 0xef, 0x6b, 0xa6, 0x8a, 0x7d, 0x4b, 0x94,
|
||||
0xa0, 0x7c, 0x56, 0x82, 0x84, 0xa7, 0x1c, 0xba, 0x35, 0xb5, 0x5a, 0x23, 0x0e, 0x2e, 0x33, 0xc9,
|
||||
0x6f, 0xa3, 0xeb, 0xae, 0x49, 0x5d, 0x8d, 0x79, 0xef, 0x5d, 0xbc, 0x01, 0x6d, 0x0c, 0xfa, 0xd5,
|
||||
0xeb, 0xc7, 0x49, 0x13, 0x4e, 0x63, 0x85, 0xef, 0x73, 0xa9, 0xfc, 0xfa, 0xed, 0xf0, 0x1e, 0x5a,
|
||||
0x8f, 0xb5, 0x03, 0x4a, 0x8f, 0x46, 0x31, 0xdc, 0x08, 0x62, 0x88, 0x7b, 0x0d, 0x01, 0x78, 0xdc,
|
||||
0x87, 0xff, 0x12, 0xad, 0xda, 0x71, 0xa9, 0x83, 0xd2, 0xbe, 0x3b, 0x43, 0x4a, 0x27, 0xa4, 0x4a,
|
||||
0x5e, 0x1f, 0xf4, 0xab, 0xab, 0x09, 0x03, 0x4e, 0xee, 0xc3, 0x37, 0x51, 0x89, 0x44, 0x13, 0xce,
|
||||
0x7d, 0xaf, 0xa5, 0x0f, 0xd3, 0x50, 0x0f, 0xdb, 0xdf, 0x5e, 0xc2, 0xfa, 0x62, 0x6c, 0x05, 0xa7,
|
||||
0xfc, 0x85, 0x3f, 0x73, 0x68, 0x63, 0x42, 0x7b, 0xe0, 0x1b, 0x08, 0xa9, 0x8e, 0xd6, 0x03, 0x27,
|
||||
0x26, 0x52, 0xd4, 0xe6, 0x0e, 0x22, 0x0b, 0x8e, 0xa1, 0xf8, 0xcf, 0x10, 0x1a, 0xb1, 0x07, 0x9a,
|
||||
0x88, 0xd3, 0x35, 0x49, 0xcf, 0x6b, 0x72, 0xc9, 0xe3, 0x8f, 0xad, 0xc6, 0x18, 0x79, 0x8a, 0x8a,
|
||||
0x0e, 0x50, 0x70, 0x7a, 0xa0, 0xbe, 0x6b, 0x39, 0xe5, 0xbc, 0x5f, 0x47, 0x6f, 0xcd, 0x20, 0xfa,
|
||||
0x58, 0x2b, 0x93, 0x37, 0x82, 0x23, 0x15, 0xf1, 0x88, 0x18, 0xc7, 0x77, 0xe1, 0x5b, 0x68, 0x4b,
|
||||
0x05, 0x12, 0x0b, 0xf3, 0x0b, 0x17, 0x28, 0x03, 0xd5, 0xef, 0x50, 0x4b, 0xf2, 0xad, 0x80, 0x60,
|
||||
0xeb, 0x60, 0x12, 0x08, 0x4f, 0xf6, 0x15, 0x7e, 0xe5, 0xd0, 0x56, 0x22, 0xb2, 0x8f, 0xc0, 0xb0,
|
||||
0x75, 0xc2, 0x60, 0x0e, 0xcf, 0xd1, 0x49, 0xe2, 0x39, 0x7a, 0x7d, 0x06, 0xf9, 0xc2, 0x20, 0xb3,
|
||||
0x9e, 0x25, 0xe1, 0x17, 0x0e, 0xdd, 0x98, 0xe8, 0x31, 0x87, 0xf6, 0xfa, 0x30, 0xd9, 0x5e, 0x5f,
|
||||
0xbd, 0xc2, 0xb9, 0x32, 0xda, 0xec, 0x79, 0xd6, 0xa9, 0x5a, 0xc3, 0xb1, 0xf5, 0xbf, 0x37, 0x3f,
|
||||
0x08, 0x7f, 0x25, 0xc7, 0x20, 0x4a, 0xe7, 0x70, 0x8c, 0x64, 0x47, 0xc9, 0x5d, 0xaa, 0xa3, 0x8c,
|
||||
0x35, 0xda, 0xfc, 0x8c, 0x8d, 0x96, 0xd2, 0xab, 0x35, 0xda, 0x13, 0xb4, 0x9a, 0x7c, 0x7d, 0x0a,
|
||||
0x97, 0xfc, 0xcd, 0xe7, 0x53, 0xb7, 0x12, 0xaf, 0x53, 0x92, 0x29, 0x3d, 0x7b, 0x50, 0xfa, 0x6f,
|
||||
0x9e, 0x3d, 0x28, 0xcd, 0x28, 0x8a, 0x9f, 0x92, 0xb3, 0xc7, 0x44, 0x9d, 0xe7, 0x3f, 0x7b, 0x78,
|
||||
0x3f, 0xa5, 0xbd, 0x7f, 0xa9, 0x4d, 0x94, 0x70, 0x86, 0x8c, 0x7e, 0x4a, 0x1f, 0x85, 0x06, 0x3c,
|
||||
0xc2, 0xc8, 0xf2, 0xd9, 0x45, 0x65, 0xe1, 0xfc, 0xa2, 0xb2, 0xf0, 0xfc, 0xa2, 0xb2, 0xf0, 0x74,
|
||||
0x50, 0xe1, 0xce, 0x06, 0x15, 0xee, 0x7c, 0x50, 0xe1, 0x9e, 0x0f, 0x2a, 0xdc, 0x6f, 0x83, 0x0a,
|
||||
0xf7, 0xec, 0xf7, 0xca, 0xc2, 0x27, 0x37, 0xa7, 0xfd, 0x61, 0xe6, 0xef, 0x00, 0x00, 0x00, 0xff,
|
||||
0xff, 0x94, 0x38, 0x0b, 0x13, 0xd0, 0x11, 0x00, 0x00,
|
||||
}
|
||||
|
||||
func (m *AllocationResult) Marshal() (dAtA []byte, err error) {
|
||||
@ -701,7 +703,7 @@ func (m *AllocationResult) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func (m *PodScheduling) Marshal() (dAtA []byte, err error) {
|
||||
func (m *PodSchedulingContext) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
@ -711,12 +713,12 @@ func (m *PodScheduling) Marshal() (dAtA []byte, err error) {
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *PodScheduling) MarshalTo(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContext) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *PodScheduling) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContext) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
@ -754,7 +756,7 @@ func (m *PodScheduling) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingList) Marshal() (dAtA []byte, err error) {
|
||||
func (m *PodSchedulingContextList) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
@ -764,12 +766,12 @@ func (m *PodSchedulingList) Marshal() (dAtA []byte, err error) {
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingList) MarshalTo(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextList) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *PodSchedulingList) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextList) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
@ -801,7 +803,7 @@ func (m *PodSchedulingList) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingSpec) Marshal() (dAtA []byte, err error) {
|
||||
func (m *PodSchedulingContextSpec) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
@ -811,12 +813,12 @@ func (m *PodSchedulingSpec) Marshal() (dAtA []byte, err error) {
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingSpec) MarshalTo(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextSpec) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *PodSchedulingSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
@ -838,7 +840,7 @@ func (m *PodSchedulingSpec) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
return len(dAtA) - i, nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingStatus) Marshal() (dAtA []byte, err error) {
|
||||
func (m *PodSchedulingContextStatus) Marshal() (dAtA []byte, err error) {
|
||||
size := m.Size()
|
||||
dAtA = make([]byte, size)
|
||||
n, err := m.MarshalToSizedBuffer(dAtA[:size])
|
||||
@ -848,12 +850,12 @@ func (m *PodSchedulingStatus) Marshal() (dAtA []byte, err error) {
|
||||
return dAtA[:n], nil
|
||||
}
|
||||
|
||||
func (m *PodSchedulingStatus) MarshalTo(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextStatus) MarshalTo(dAtA []byte) (int, error) {
|
||||
size := m.Size()
|
||||
return m.MarshalToSizedBuffer(dAtA[:size])
|
||||
}
|
||||
|
||||
func (m *PodSchedulingStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
func (m *PodSchedulingContextStatus) MarshalToSizedBuffer(dAtA []byte) (int, error) {
|
||||
i := len(dAtA)
|
||||
_ = i
|
||||
var l int
|
||||
@ -1512,7 +1514,7 @@ func (m *AllocationResult) Size() (n int) {
|
||||
return n
|
||||
}
|
||||
|
||||
func (m *PodScheduling) Size() (n int) {
|
||||
func (m *PodSchedulingContext) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
@ -1527,7 +1529,7 @@ func (m *PodScheduling) Size() (n int) {
|
||||
return n
|
||||
}
|
||||
|
||||
func (m *PodSchedulingList) Size() (n int) {
|
||||
func (m *PodSchedulingContextList) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
@ -1544,7 +1546,7 @@ func (m *PodSchedulingList) Size() (n int) {
|
||||
return n
|
||||
}
|
||||
|
||||
func (m *PodSchedulingSpec) Size() (n int) {
|
||||
func (m *PodSchedulingContextSpec) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
@ -1561,7 +1563,7 @@ func (m *PodSchedulingSpec) Size() (n int) {
|
||||
return n
|
||||
}
|
||||
|
||||
func (m *PodSchedulingStatus) Size() (n int) {
|
||||
func (m *PodSchedulingContextStatus) Size() (n int) {
|
||||
if m == nil {
|
||||
return 0
|
||||
}
|
||||
@ -1812,46 +1814,46 @@ func (this *AllocationResult) String() string {
|
||||
}, "")
|
||||
return s
|
||||
}
|
||||
func (this *PodScheduling) String() string {
|
||||
func (this *PodSchedulingContext) String() string {
|
||||
if this == nil {
|
||||
return "nil"
|
||||
}
|
||||
s := strings.Join([]string{`&PodScheduling{`,
|
||||
s := strings.Join([]string{`&PodSchedulingContext{`,
|
||||
`ObjectMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ObjectMeta), "ObjectMeta", "v11.ObjectMeta", 1), `&`, ``, 1) + `,`,
|
||||
`Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "PodSchedulingSpec", "PodSchedulingSpec", 1), `&`, ``, 1) + `,`,
|
||||
`Status:` + strings.Replace(strings.Replace(this.Status.String(), "PodSchedulingStatus", "PodSchedulingStatus", 1), `&`, ``, 1) + `,`,
|
||||
`Spec:` + strings.Replace(strings.Replace(this.Spec.String(), "PodSchedulingContextSpec", "PodSchedulingContextSpec", 1), `&`, ``, 1) + `,`,
|
||||
`Status:` + strings.Replace(strings.Replace(this.Status.String(), "PodSchedulingContextStatus", "PodSchedulingContextStatus", 1), `&`, ``, 1) + `,`,
|
||||
`}`,
|
||||
}, "")
|
||||
return s
|
||||
}
|
||||
func (this *PodSchedulingList) String() string {
|
||||
func (this *PodSchedulingContextList) String() string {
|
||||
if this == nil {
|
||||
return "nil"
|
||||
}
|
||||
repeatedStringForItems := "[]PodScheduling{"
|
||||
repeatedStringForItems := "[]PodSchedulingContext{"
|
||||
for _, f := range this.Items {
|
||||
repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "PodScheduling", "PodScheduling", 1), `&`, ``, 1) + ","
|
||||
repeatedStringForItems += strings.Replace(strings.Replace(f.String(), "PodSchedulingContext", "PodSchedulingContext", 1), `&`, ``, 1) + ","
|
||||
}
|
||||
repeatedStringForItems += "}"
|
||||
s := strings.Join([]string{`&PodSchedulingList{`,
|
||||
s := strings.Join([]string{`&PodSchedulingContextList{`,
|
||||
`ListMeta:` + strings.Replace(strings.Replace(fmt.Sprintf("%v", this.ListMeta), "ListMeta", "v11.ListMeta", 1), `&`, ``, 1) + `,`,
|
||||
`Items:` + repeatedStringForItems + `,`,
|
||||
`}`,
|
||||
}, "")
|
||||
return s
|
||||
}
|
||||
func (this *PodSchedulingSpec) String() string {
|
||||
func (this *PodSchedulingContextSpec) String() string {
|
||||
if this == nil {
|
||||
return "nil"
|
||||
}
|
||||
s := strings.Join([]string{`&PodSchedulingSpec{`,
|
||||
s := strings.Join([]string{`&PodSchedulingContextSpec{`,
|
||||
`SelectedNode:` + fmt.Sprintf("%v", this.SelectedNode) + `,`,
|
||||
`PotentialNodes:` + fmt.Sprintf("%v", this.PotentialNodes) + `,`,
|
||||
`}`,
|
||||
}, "")
|
||||
return s
|
||||
}
|
||||
func (this *PodSchedulingStatus) String() string {
|
||||
func (this *PodSchedulingContextStatus) String() string {
|
||||
if this == nil {
|
||||
return "nil"
|
||||
}
|
||||
@ -1860,7 +1862,7 @@ func (this *PodSchedulingStatus) String() string {
|
||||
repeatedStringForResourceClaims += strings.Replace(strings.Replace(f.String(), "ResourceClaimSchedulingStatus", "ResourceClaimSchedulingStatus", 1), `&`, ``, 1) + ","
|
||||
}
|
||||
repeatedStringForResourceClaims += "}"
|
||||
s := strings.Join([]string{`&PodSchedulingStatus{`,
|
||||
s := strings.Join([]string{`&PodSchedulingContextStatus{`,
|
||||
`ResourceClaims:` + repeatedStringForResourceClaims + `,`,
|
||||
`}`,
|
||||
}, "")
|
||||
@ -2186,7 +2188,7 @@ func (m *AllocationResult) Unmarshal(dAtA []byte) error {
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func (m *PodScheduling) Unmarshal(dAtA []byte) error {
|
||||
func (m *PodSchedulingContext) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
@ -2209,10 +2211,10 @@ func (m *PodScheduling) Unmarshal(dAtA []byte) error {
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: PodScheduling: wiretype end group for non-group")
|
||||
return fmt.Errorf("proto: PodSchedulingContext: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: PodScheduling: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
return fmt.Errorf("proto: PodSchedulingContext: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
case 1:
|
||||
@ -2335,7 +2337,7 @@ func (m *PodScheduling) Unmarshal(dAtA []byte) error {
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func (m *PodSchedulingList) Unmarshal(dAtA []byte) error {
|
||||
func (m *PodSchedulingContextList) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
@ -2358,10 +2360,10 @@ func (m *PodSchedulingList) Unmarshal(dAtA []byte) error {
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: PodSchedulingList: wiretype end group for non-group")
|
||||
return fmt.Errorf("proto: PodSchedulingContextList: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: PodSchedulingList: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
return fmt.Errorf("proto: PodSchedulingContextList: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
case 1:
|
||||
@ -2426,7 +2428,7 @@ func (m *PodSchedulingList) Unmarshal(dAtA []byte) error {
|
||||
if postIndex > l {
|
||||
return io.ErrUnexpectedEOF
|
||||
}
|
||||
m.Items = append(m.Items, PodScheduling{})
|
||||
m.Items = append(m.Items, PodSchedulingContext{})
|
||||
if err := m.Items[len(m.Items)-1].Unmarshal(dAtA[iNdEx:postIndex]); err != nil {
|
||||
return err
|
||||
}
|
||||
@ -2452,7 +2454,7 @@ func (m *PodSchedulingList) Unmarshal(dAtA []byte) error {
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func (m *PodSchedulingSpec) Unmarshal(dAtA []byte) error {
|
||||
func (m *PodSchedulingContextSpec) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
@ -2475,10 +2477,10 @@ func (m *PodSchedulingSpec) Unmarshal(dAtA []byte) error {
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: PodSchedulingSpec: wiretype end group for non-group")
|
||||
return fmt.Errorf("proto: PodSchedulingContextSpec: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: PodSchedulingSpec: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
return fmt.Errorf("proto: PodSchedulingContextSpec: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
case 1:
|
||||
@ -2566,7 +2568,7 @@ func (m *PodSchedulingSpec) Unmarshal(dAtA []byte) error {
|
||||
}
|
||||
return nil
|
||||
}
|
||||
func (m *PodSchedulingStatus) Unmarshal(dAtA []byte) error {
|
||||
func (m *PodSchedulingContextStatus) Unmarshal(dAtA []byte) error {
|
||||
l := len(dAtA)
|
||||
iNdEx := 0
|
||||
for iNdEx < l {
|
||||
@ -2589,10 +2591,10 @@ func (m *PodSchedulingStatus) Unmarshal(dAtA []byte) error {
|
||||
fieldNum := int32(wire >> 3)
|
||||
wireType := int(wire & 0x7)
|
||||
if wireType == 4 {
|
||||
return fmt.Errorf("proto: PodSchedulingStatus: wiretype end group for non-group")
|
||||
return fmt.Errorf("proto: PodSchedulingContextStatus: wiretype end group for non-group")
|
||||
}
|
||||
if fieldNum <= 0 {
|
||||
return fmt.Errorf("proto: PodSchedulingStatus: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
return fmt.Errorf("proto: PodSchedulingContextStatus: illegal tag %d (wire type %d)", fieldNum, wire)
|
||||
}
|
||||
switch fieldNum {
|
||||
case 1:
|
||||
|
@ -56,37 +56,37 @@ message AllocationResult {
|
||||
optional bool shareable = 3;
|
||||
}
|
||||
|
||||
// PodScheduling objects hold information that is needed to schedule
|
||||
// PodSchedulingContext objects hold information that is needed to schedule
|
||||
// a Pod with ResourceClaims that use "WaitForFirstConsumer" allocation
|
||||
// mode.
|
||||
//
|
||||
// This is an alpha type and requires enabling the DynamicResourceAllocation
|
||||
// feature gate.
|
||||
message PodScheduling {
|
||||
message PodSchedulingContext {
|
||||
// Standard object metadata
|
||||
// +optional
|
||||
optional k8s.io.apimachinery.pkg.apis.meta.v1.ObjectMeta metadata = 1;
|
||||
|
||||
// Spec describes where resources for the Pod are needed.
|
||||
optional PodSchedulingSpec spec = 2;
|
||||
optional PodSchedulingContextSpec spec = 2;
|
||||
|
||||
// Status describes where resources for the Pod can be allocated.
|
||||
// +optional
|
||||
optional PodSchedulingStatus status = 3;
|
||||
optional PodSchedulingContextStatus status = 3;
|
||||
}
|
||||
|
||||
// PodSchedulingList is a collection of Pod scheduling objects.
|
||||
message PodSchedulingList {
|
||||
// PodSchedulingContextList is a collection of Pod scheduling objects.
|
||||
message PodSchedulingContextList {
|
||||
// Standard list metadata
|
||||
// +optional
|
||||
optional k8s.io.apimachinery.pkg.apis.meta.v1.ListMeta metadata = 1;
|
||||
|
||||
// Items is the list of PodScheduling objects.
|
||||
repeated PodScheduling items = 2;
|
||||
// Items is the list of PodSchedulingContext objects.
|
||||
repeated PodSchedulingContext items = 2;
|
||||
}
|
||||
|
||||
// PodSchedulingSpec describes where resources for the Pod are needed.
|
||||
message PodSchedulingSpec {
|
||||
// PodSchedulingContextSpec describes where resources for the Pod are needed.
|
||||
message PodSchedulingContextSpec {
|
||||
// SelectedNode is the node for which allocation of ResourceClaims that
|
||||
// are referenced by the Pod and that use "WaitForFirstConsumer"
|
||||
// allocation is to be attempted.
|
||||
@ -105,8 +105,8 @@ message PodSchedulingSpec {
|
||||
repeated string potentialNodes = 2;
|
||||
}
|
||||
|
||||
// PodSchedulingStatus describes where resources for the Pod can be allocated.
|
||||
message PodSchedulingStatus {
|
||||
// PodSchedulingContextStatus describes where resources for the Pod can be allocated.
|
||||
message PodSchedulingContextStatus {
|
||||
// ResourceClaims describes resource availability for each
|
||||
// pod.spec.resourceClaim entry where the corresponding ResourceClaim
|
||||
// uses "WaitForFirstConsumer" allocation mode.
|
||||
|
@ -38,44 +38,44 @@ func (AllocationResult) SwaggerDoc() map[string]string {
|
||||
return map_AllocationResult
|
||||
}
|
||||
|
||||
var map_PodScheduling = map[string]string{
|
||||
"": "PodScheduling objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
var map_PodSchedulingContext = map[string]string{
|
||||
"": "PodSchedulingContext objects hold information that is needed to schedule a Pod with ResourceClaims that use \"WaitForFirstConsumer\" allocation mode.\n\nThis is an alpha type and requires enabling the DynamicResourceAllocation feature gate.",
|
||||
"metadata": "Standard object metadata",
|
||||
"spec": "Spec describes where resources for the Pod are needed.",
|
||||
"status": "Status describes where resources for the Pod can be allocated.",
|
||||
}
|
||||
|
||||
func (PodScheduling) SwaggerDoc() map[string]string {
|
||||
return map_PodScheduling
|
||||
func (PodSchedulingContext) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingContext
|
||||
}
|
||||
|
||||
var map_PodSchedulingList = map[string]string{
|
||||
"": "PodSchedulingList is a collection of Pod scheduling objects.",
|
||||
var map_PodSchedulingContextList = map[string]string{
|
||||
"": "PodSchedulingContextList is a collection of Pod scheduling objects.",
|
||||
"metadata": "Standard list metadata",
|
||||
"items": "Items is the list of PodScheduling objects.",
|
||||
"items": "Items is the list of PodSchedulingContext objects.",
|
||||
}
|
||||
|
||||
func (PodSchedulingList) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingList
|
||||
func (PodSchedulingContextList) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingContextList
|
||||
}
|
||||
|
||||
var map_PodSchedulingSpec = map[string]string{
|
||||
"": "PodSchedulingSpec describes where resources for the Pod are needed.",
|
||||
var map_PodSchedulingContextSpec = map[string]string{
|
||||
"": "PodSchedulingContextSpec describes where resources for the Pod are needed.",
|
||||
"selectedNode": "SelectedNode is the node for which allocation of ResourceClaims that are referenced by the Pod and that use \"WaitForFirstConsumer\" allocation is to be attempted.",
|
||||
"potentialNodes": "PotentialNodes lists nodes where the Pod might be able to run.\n\nThe size of this field is limited to 128. This is large enough for many clusters. Larger clusters may need more attempts to find a node that suits all pending resources. This may get increased in the future, but not reduced.",
|
||||
}
|
||||
|
||||
func (PodSchedulingSpec) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingSpec
|
||||
func (PodSchedulingContextSpec) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingContextSpec
|
||||
}
|
||||
|
||||
var map_PodSchedulingStatus = map[string]string{
|
||||
"": "PodSchedulingStatus describes where resources for the Pod can be allocated.",
|
||||
var map_PodSchedulingContextStatus = map[string]string{
|
||||
"": "PodSchedulingContextStatus describes where resources for the Pod can be allocated.",
|
||||
"resourceClaims": "ResourceClaims describes resource availability for each pod.spec.resourceClaim entry where the corresponding ResourceClaim uses \"WaitForFirstConsumer\" allocation mode.",
|
||||
}
|
||||
|
||||
func (PodSchedulingStatus) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingStatus
|
||||
func (PodSchedulingContextStatus) SwaggerDoc() map[string]string {
|
||||
return map_PodSchedulingContextStatus
|
||||
}
|
||||
|
||||
var map_ResourceClaim = map[string]string{
|
||||
|
@ -48,7 +48,7 @@ func (in *AllocationResult) DeepCopy() *AllocationResult {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodScheduling) DeepCopyInto(out *PodScheduling) {
|
||||
func (in *PodSchedulingContext) DeepCopyInto(out *PodSchedulingContext) {
|
||||
*out = *in
|
||||
out.TypeMeta = in.TypeMeta
|
||||
in.ObjectMeta.DeepCopyInto(&out.ObjectMeta)
|
||||
@ -57,18 +57,18 @@ func (in *PodScheduling) DeepCopyInto(out *PodScheduling) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodScheduling.
|
||||
func (in *PodScheduling) DeepCopy() *PodScheduling {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContext.
|
||||
func (in *PodSchedulingContext) DeepCopy() *PodSchedulingContext {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodScheduling)
|
||||
out := new(PodSchedulingContext)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
||||
func (in *PodScheduling) DeepCopyObject() runtime.Object {
|
||||
func (in *PodSchedulingContext) DeepCopyObject() runtime.Object {
|
||||
if c := in.DeepCopy(); c != nil {
|
||||
return c
|
||||
}
|
||||
@ -76,13 +76,13 @@ func (in *PodScheduling) DeepCopyObject() runtime.Object {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingList) DeepCopyInto(out *PodSchedulingList) {
|
||||
func (in *PodSchedulingContextList) DeepCopyInto(out *PodSchedulingContextList) {
|
||||
*out = *in
|
||||
out.TypeMeta = in.TypeMeta
|
||||
in.ListMeta.DeepCopyInto(&out.ListMeta)
|
||||
if in.Items != nil {
|
||||
in, out := &in.Items, &out.Items
|
||||
*out = make([]PodScheduling, len(*in))
|
||||
*out = make([]PodSchedulingContext, len(*in))
|
||||
for i := range *in {
|
||||
(*in)[i].DeepCopyInto(&(*out)[i])
|
||||
}
|
||||
@ -90,18 +90,18 @@ func (in *PodSchedulingList) DeepCopyInto(out *PodSchedulingList) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingList.
|
||||
func (in *PodSchedulingList) DeepCopy() *PodSchedulingList {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextList.
|
||||
func (in *PodSchedulingContextList) DeepCopy() *PodSchedulingContextList {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingList)
|
||||
out := new(PodSchedulingContextList)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyObject is an autogenerated deepcopy function, copying the receiver, creating a new runtime.Object.
|
||||
func (in *PodSchedulingList) DeepCopyObject() runtime.Object {
|
||||
func (in *PodSchedulingContextList) DeepCopyObject() runtime.Object {
|
||||
if c := in.DeepCopy(); c != nil {
|
||||
return c
|
||||
}
|
||||
@ -109,7 +109,7 @@ func (in *PodSchedulingList) DeepCopyObject() runtime.Object {
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingSpec) DeepCopyInto(out *PodSchedulingSpec) {
|
||||
func (in *PodSchedulingContextSpec) DeepCopyInto(out *PodSchedulingContextSpec) {
|
||||
*out = *in
|
||||
if in.PotentialNodes != nil {
|
||||
in, out := &in.PotentialNodes, &out.PotentialNodes
|
||||
@ -119,18 +119,18 @@ func (in *PodSchedulingSpec) DeepCopyInto(out *PodSchedulingSpec) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingSpec.
|
||||
func (in *PodSchedulingSpec) DeepCopy() *PodSchedulingSpec {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextSpec.
|
||||
func (in *PodSchedulingContextSpec) DeepCopy() *PodSchedulingContextSpec {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingSpec)
|
||||
out := new(PodSchedulingContextSpec)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
||||
// DeepCopyInto is an autogenerated deepcopy function, copying the receiver, writing into out. in must be non-nil.
|
||||
func (in *PodSchedulingStatus) DeepCopyInto(out *PodSchedulingStatus) {
|
||||
func (in *PodSchedulingContextStatus) DeepCopyInto(out *PodSchedulingContextStatus) {
|
||||
*out = *in
|
||||
if in.ResourceClaims != nil {
|
||||
in, out := &in.ResourceClaims, &out.ResourceClaims
|
||||
@ -142,12 +142,12 @@ func (in *PodSchedulingStatus) DeepCopyInto(out *PodSchedulingStatus) {
|
||||
return
|
||||
}
|
||||
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingStatus.
|
||||
func (in *PodSchedulingStatus) DeepCopy() *PodSchedulingStatus {
|
||||
// DeepCopy is an autogenerated deepcopy function, copying the receiver, creating a new PodSchedulingContextStatus.
|
||||
func (in *PodSchedulingContextStatus) DeepCopy() *PodSchedulingContextStatus {
|
||||
if in == nil {
|
||||
return nil
|
||||
}
|
||||
out := new(PodSchedulingStatus)
|
||||
out := new(PodSchedulingContextStatus)
|
||||
in.DeepCopyInto(out)
|
||||
return out
|
||||
}
|
||||
|
@ -1,5 +1,5 @@
|
||||
{
|
||||
"kind": "PodScheduling",
|
||||
"kind": "PodSchedulingContext",
|
||||
"apiVersion": "resource.k8s.io/v1alpha2",
|
||||
"metadata": {
|
||||
"name": "nameValue",
|
Binary file not shown.
@ -1,5 +1,5 @@
|
||||
apiVersion: resource.k8s.io/v1alpha2
|
||||
kind: PodScheduling
|
||||
kind: PodSchedulingContext
|
||||
metadata:
|
||||
annotations:
|
||||
annotationsKey: annotationsValue
|
@ -11376,7 +11376,7 @@ var schemaYAML = typed.YAMLObject(`types:
|
||||
- name: shareable
|
||||
type:
|
||||
scalar: boolean
|
||||
- name: io.k8s.api.resource.v1alpha2.PodScheduling
|
||||
- name: io.k8s.api.resource.v1alpha2.PodSchedulingContext
|
||||
map:
|
||||
fields:
|
||||
- name: apiVersion
|
||||
@ -11391,13 +11391,13 @@ var schemaYAML = typed.YAMLObject(`types:
|
||||
default: {}
|
||||
- name: spec
|
||||
type:
|
||||
namedType: io.k8s.api.resource.v1alpha2.PodSchedulingSpec
|
||||
namedType: io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec
|
||||
default: {}
|
||||
- name: status
|
||||
type:
|
||||
namedType: io.k8s.api.resource.v1alpha2.PodSchedulingStatus
|
||||
namedType: io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus
|
||||
default: {}
|
||||
- name: io.k8s.api.resource.v1alpha2.PodSchedulingSpec
|
||||
- name: io.k8s.api.resource.v1alpha2.PodSchedulingContextSpec
|
||||
map:
|
||||
fields:
|
||||
- name: potentialNodes
|
||||
@ -11409,7 +11409,7 @@ var schemaYAML = typed.YAMLObject(`types:
|
||||
- name: selectedNode
|
||||
type:
|
||||
scalar: string
|
||||
- name: io.k8s.api.resource.v1alpha2.PodSchedulingStatus
|
||||
- name: io.k8s.api.resource.v1alpha2.PodSchedulingContextStatus
|
||||
map:
|
||||
fields:
|
||||
- name: resourceClaims
|
||||
|
@ -27,58 +27,58 @@ import (
|
||||
v1 "k8s.io/client-go/applyconfigurations/meta/v1"
|
||||
)
|
||||
|
||||
// PodSchedulingApplyConfiguration represents an declarative configuration of the PodScheduling type for use
|
||||
// PodSchedulingContextApplyConfiguration represents an declarative configuration of the PodSchedulingContext type for use
|
||||
// with apply.
|
||||
type PodSchedulingApplyConfiguration struct {
|
||||
type PodSchedulingContextApplyConfiguration struct {
|
||||
v1.TypeMetaApplyConfiguration `json:",inline"`
|
||||
*v1.ObjectMetaApplyConfiguration `json:"metadata,omitempty"`
|
||||
Spec *PodSchedulingSpecApplyConfiguration `json:"spec,omitempty"`
|
||||
Status *PodSchedulingStatusApplyConfiguration `json:"status,omitempty"`
|
||||
Spec *PodSchedulingContextSpecApplyConfiguration `json:"spec,omitempty"`
|
||||
Status *PodSchedulingContextStatusApplyConfiguration `json:"status,omitempty"`
|
||||
}
|
||||
|
||||
// PodScheduling constructs an declarative configuration of the PodScheduling type for use with
|
||||
// PodSchedulingContext constructs an declarative configuration of the PodSchedulingContext type for use with
|
||||
// apply.
|
||||
func PodScheduling(name, namespace string) *PodSchedulingApplyConfiguration {
|
||||
b := &PodSchedulingApplyConfiguration{}
|
||||
func PodSchedulingContext(name, namespace string) *PodSchedulingContextApplyConfiguration {
|
||||
b := &PodSchedulingContextApplyConfiguration{}
|
||||
b.WithName(name)
|
||||
b.WithNamespace(namespace)
|
||||
b.WithKind("PodScheduling")
|
||||
b.WithKind("PodSchedulingContext")
|
||||
b.WithAPIVersion("resource.k8s.io/v1alpha2")
|
||||
return b
|
||||
}
|
||||
|
||||
// ExtractPodScheduling extracts the applied configuration owned by fieldManager from
|
||||
// podScheduling. If no managedFields are found in podScheduling for fieldManager, a
|
||||
// PodSchedulingApplyConfiguration is returned with only the Name, Namespace (if applicable),
|
||||
// ExtractPodSchedulingContext extracts the applied configuration owned by fieldManager from
|
||||
// podSchedulingContext. If no managedFields are found in podSchedulingContext for fieldManager, a
|
||||
// PodSchedulingContextApplyConfiguration is returned with only the Name, Namespace (if applicable),
|
||||
// APIVersion and Kind populated. It is possible that no managed fields were found for because other
|
||||
// field managers have taken ownership of all the fields previously owned by fieldManager, or because
|
||||
// the fieldManager never owned fields any fields.
|
||||
// podScheduling must be a unmodified PodScheduling API object that was retrieved from the Kubernetes API.
|
||||
// ExtractPodScheduling provides a way to perform a extract/modify-in-place/apply workflow.
|
||||
// podSchedulingContext must be a unmodified PodSchedulingContext API object that was retrieved from the Kubernetes API.
|
||||
// ExtractPodSchedulingContext provides a way to perform a extract/modify-in-place/apply workflow.
|
||||
// Note that an extracted apply configuration will contain fewer fields than what the fieldManager previously
|
||||
// applied if another fieldManager has updated or force applied any of the previously applied fields.
|
||||
// Experimental!
|
||||
func ExtractPodScheduling(podScheduling *resourcev1alpha2.PodScheduling, fieldManager string) (*PodSchedulingApplyConfiguration, error) {
|
||||
return extractPodScheduling(podScheduling, fieldManager, "")
|
||||
func ExtractPodSchedulingContext(podSchedulingContext *resourcev1alpha2.PodSchedulingContext, fieldManager string) (*PodSchedulingContextApplyConfiguration, error) {
|
||||
return extractPodSchedulingContext(podSchedulingContext, fieldManager, "")
|
||||
}
|
||||
|
||||
// ExtractPodSchedulingStatus is the same as ExtractPodScheduling except
|
||||
// ExtractPodSchedulingContextStatus is the same as ExtractPodSchedulingContext except
|
||||
// that it extracts the status subresource applied configuration.
|
||||
// Experimental!
|
||||
func ExtractPodSchedulingStatus(podScheduling *resourcev1alpha2.PodScheduling, fieldManager string) (*PodSchedulingApplyConfiguration, error) {
|
||||
return extractPodScheduling(podScheduling, fieldManager, "status")
|
||||
func ExtractPodSchedulingContextStatus(podSchedulingContext *resourcev1alpha2.PodSchedulingContext, fieldManager string) (*PodSchedulingContextApplyConfiguration, error) {
|
||||
return extractPodSchedulingContext(podSchedulingContext, fieldManager, "status")
|
||||
}
|
||||
|
||||
func extractPodScheduling(podScheduling *resourcev1alpha2.PodScheduling, fieldManager string, subresource string) (*PodSchedulingApplyConfiguration, error) {
|
||||
b := &PodSchedulingApplyConfiguration{}
|
||||
err := managedfields.ExtractInto(podScheduling, internal.Parser().Type("io.k8s.api.resource.v1alpha2.PodScheduling"), fieldManager, b, subresource)
|
||||
func extractPodSchedulingContext(podSchedulingContext *resourcev1alpha2.PodSchedulingContext, fieldManager string, subresource string) (*PodSchedulingContextApplyConfiguration, error) {
|
||||
b := &PodSchedulingContextApplyConfiguration{}
|
||||
err := managedfields.ExtractInto(podSchedulingContext, internal.Parser().Type("io.k8s.api.resource.v1alpha2.PodSchedulingContext"), fieldManager, b, subresource)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
b.WithName(podScheduling.Name)
|
||||
b.WithNamespace(podScheduling.Namespace)
|
||||
b.WithName(podSchedulingContext.Name)
|
||||
b.WithNamespace(podSchedulingContext.Namespace)
|
||||
|
||||
b.WithKind("PodScheduling")
|
||||
b.WithKind("PodSchedulingContext")
|
||||
b.WithAPIVersion("resource.k8s.io/v1alpha2")
|
||||
return b, nil
|
||||
}
|
||||
@ -86,7 +86,7 @@ func extractPodScheduling(podScheduling *resourcev1alpha2.PodScheduling, fieldMa
|
||||
// WithKind sets the Kind field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Kind field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithKind(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithKind(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.Kind = &value
|
||||
return b
|
||||
}
|
||||
@ -94,7 +94,7 @@ func (b *PodSchedulingApplyConfiguration) WithKind(value string) *PodSchedulingA
|
||||
// WithAPIVersion sets the APIVersion field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the APIVersion field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithAPIVersion(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithAPIVersion(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.APIVersion = &value
|
||||
return b
|
||||
}
|
||||
@ -102,7 +102,7 @@ func (b *PodSchedulingApplyConfiguration) WithAPIVersion(value string) *PodSched
|
||||
// WithName sets the Name field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Name field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithName(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithName(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.Name = &value
|
||||
return b
|
||||
@ -111,7 +111,7 @@ func (b *PodSchedulingApplyConfiguration) WithName(value string) *PodSchedulingA
|
||||
// WithGenerateName sets the GenerateName field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the GenerateName field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithGenerateName(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithGenerateName(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.GenerateName = &value
|
||||
return b
|
||||
@ -120,7 +120,7 @@ func (b *PodSchedulingApplyConfiguration) WithGenerateName(value string) *PodSch
|
||||
// WithNamespace sets the Namespace field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Namespace field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithNamespace(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithNamespace(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.Namespace = &value
|
||||
return b
|
||||
@ -129,7 +129,7 @@ func (b *PodSchedulingApplyConfiguration) WithNamespace(value string) *PodSchedu
|
||||
// WithUID sets the UID field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the UID field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithUID(value types.UID) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithUID(value types.UID) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.UID = &value
|
||||
return b
|
||||
@ -138,7 +138,7 @@ func (b *PodSchedulingApplyConfiguration) WithUID(value types.UID) *PodSchedulin
|
||||
// WithResourceVersion sets the ResourceVersion field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the ResourceVersion field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithResourceVersion(value string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithResourceVersion(value string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.ResourceVersion = &value
|
||||
return b
|
||||
@ -147,7 +147,7 @@ func (b *PodSchedulingApplyConfiguration) WithResourceVersion(value string) *Pod
|
||||
// WithGeneration sets the Generation field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Generation field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithGeneration(value int64) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithGeneration(value int64) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.Generation = &value
|
||||
return b
|
||||
@ -156,7 +156,7 @@ func (b *PodSchedulingApplyConfiguration) WithGeneration(value int64) *PodSchedu
|
||||
// WithCreationTimestamp sets the CreationTimestamp field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the CreationTimestamp field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithCreationTimestamp(value metav1.Time) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithCreationTimestamp(value metav1.Time) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.CreationTimestamp = &value
|
||||
return b
|
||||
@ -165,7 +165,7 @@ func (b *PodSchedulingApplyConfiguration) WithCreationTimestamp(value metav1.Tim
|
||||
// WithDeletionTimestamp sets the DeletionTimestamp field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the DeletionTimestamp field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithDeletionTimestamp(value metav1.Time) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithDeletionTimestamp(value metav1.Time) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.DeletionTimestamp = &value
|
||||
return b
|
||||
@ -174,7 +174,7 @@ func (b *PodSchedulingApplyConfiguration) WithDeletionTimestamp(value metav1.Tim
|
||||
// WithDeletionGracePeriodSeconds sets the DeletionGracePeriodSeconds field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the DeletionGracePeriodSeconds field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithDeletionGracePeriodSeconds(value int64) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithDeletionGracePeriodSeconds(value int64) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
b.DeletionGracePeriodSeconds = &value
|
||||
return b
|
||||
@ -184,7 +184,7 @@ func (b *PodSchedulingApplyConfiguration) WithDeletionGracePeriodSeconds(value i
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, the entries provided by each call will be put on the Labels field,
|
||||
// overwriting an existing map entries in Labels field with the same key.
|
||||
func (b *PodSchedulingApplyConfiguration) WithLabels(entries map[string]string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithLabels(entries map[string]string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
if b.Labels == nil && len(entries) > 0 {
|
||||
b.Labels = make(map[string]string, len(entries))
|
||||
@ -199,7 +199,7 @@ func (b *PodSchedulingApplyConfiguration) WithLabels(entries map[string]string)
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, the entries provided by each call will be put on the Annotations field,
|
||||
// overwriting an existing map entries in Annotations field with the same key.
|
||||
func (b *PodSchedulingApplyConfiguration) WithAnnotations(entries map[string]string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithAnnotations(entries map[string]string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
if b.Annotations == nil && len(entries) > 0 {
|
||||
b.Annotations = make(map[string]string, len(entries))
|
||||
@ -213,7 +213,7 @@ func (b *PodSchedulingApplyConfiguration) WithAnnotations(entries map[string]str
|
||||
// WithOwnerReferences adds the given value to the OwnerReferences field in the declarative configuration
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, values provided by each call will be appended to the OwnerReferences field.
|
||||
func (b *PodSchedulingApplyConfiguration) WithOwnerReferences(values ...*v1.OwnerReferenceApplyConfiguration) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithOwnerReferences(values ...*v1.OwnerReferenceApplyConfiguration) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
for i := range values {
|
||||
if values[i] == nil {
|
||||
@ -227,7 +227,7 @@ func (b *PodSchedulingApplyConfiguration) WithOwnerReferences(values ...*v1.Owne
|
||||
// WithFinalizers adds the given value to the Finalizers field in the declarative configuration
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, values provided by each call will be appended to the Finalizers field.
|
||||
func (b *PodSchedulingApplyConfiguration) WithFinalizers(values ...string) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithFinalizers(values ...string) *PodSchedulingContextApplyConfiguration {
|
||||
b.ensureObjectMetaApplyConfigurationExists()
|
||||
for i := range values {
|
||||
b.Finalizers = append(b.Finalizers, values[i])
|
||||
@ -235,7 +235,7 @@ func (b *PodSchedulingApplyConfiguration) WithFinalizers(values ...string) *PodS
|
||||
return b
|
||||
}
|
||||
|
||||
func (b *PodSchedulingApplyConfiguration) ensureObjectMetaApplyConfigurationExists() {
|
||||
func (b *PodSchedulingContextApplyConfiguration) ensureObjectMetaApplyConfigurationExists() {
|
||||
if b.ObjectMetaApplyConfiguration == nil {
|
||||
b.ObjectMetaApplyConfiguration = &v1.ObjectMetaApplyConfiguration{}
|
||||
}
|
||||
@ -244,7 +244,7 @@ func (b *PodSchedulingApplyConfiguration) ensureObjectMetaApplyConfigurationExis
|
||||
// WithSpec sets the Spec field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Spec field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithSpec(value *PodSchedulingSpecApplyConfiguration) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithSpec(value *PodSchedulingContextSpecApplyConfiguration) *PodSchedulingContextApplyConfiguration {
|
||||
b.Spec = value
|
||||
return b
|
||||
}
|
||||
@ -252,7 +252,7 @@ func (b *PodSchedulingApplyConfiguration) WithSpec(value *PodSchedulingSpecApply
|
||||
// WithStatus sets the Status field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the Status field is set to the value of the last call.
|
||||
func (b *PodSchedulingApplyConfiguration) WithStatus(value *PodSchedulingStatusApplyConfiguration) *PodSchedulingApplyConfiguration {
|
||||
func (b *PodSchedulingContextApplyConfiguration) WithStatus(value *PodSchedulingContextStatusApplyConfiguration) *PodSchedulingContextApplyConfiguration {
|
||||
b.Status = value
|
||||
return b
|
||||
}
|
@ -18,23 +18,23 @@ limitations under the License.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
// PodSchedulingSpecApplyConfiguration represents an declarative configuration of the PodSchedulingSpec type for use
|
||||
// PodSchedulingContextSpecApplyConfiguration represents an declarative configuration of the PodSchedulingContextSpec type for use
|
||||
// with apply.
|
||||
type PodSchedulingSpecApplyConfiguration struct {
|
||||
type PodSchedulingContextSpecApplyConfiguration struct {
|
||||
SelectedNode *string `json:"selectedNode,omitempty"`
|
||||
PotentialNodes []string `json:"potentialNodes,omitempty"`
|
||||
}
|
||||
|
||||
// PodSchedulingSpecApplyConfiguration constructs an declarative configuration of the PodSchedulingSpec type for use with
|
||||
// PodSchedulingContextSpecApplyConfiguration constructs an declarative configuration of the PodSchedulingContextSpec type for use with
|
||||
// apply.
|
||||
func PodSchedulingSpec() *PodSchedulingSpecApplyConfiguration {
|
||||
return &PodSchedulingSpecApplyConfiguration{}
|
||||
func PodSchedulingContextSpec() *PodSchedulingContextSpecApplyConfiguration {
|
||||
return &PodSchedulingContextSpecApplyConfiguration{}
|
||||
}
|
||||
|
||||
// WithSelectedNode sets the SelectedNode field in the declarative configuration to the given value
|
||||
// and returns the receiver, so that objects can be built by chaining "With" function invocations.
|
||||
// If called multiple times, the SelectedNode field is set to the value of the last call.
|
||||
func (b *PodSchedulingSpecApplyConfiguration) WithSelectedNode(value string) *PodSchedulingSpecApplyConfiguration {
|
||||
func (b *PodSchedulingContextSpecApplyConfiguration) WithSelectedNode(value string) *PodSchedulingContextSpecApplyConfiguration {
|
||||
b.SelectedNode = &value
|
||||
return b
|
||||
}
|
||||
@ -42,7 +42,7 @@ func (b *PodSchedulingSpecApplyConfiguration) WithSelectedNode(value string) *Po
|
||||
// WithPotentialNodes adds the given value to the PotentialNodes field in the declarative configuration
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, values provided by each call will be appended to the PotentialNodes field.
|
||||
func (b *PodSchedulingSpecApplyConfiguration) WithPotentialNodes(values ...string) *PodSchedulingSpecApplyConfiguration {
|
||||
func (b *PodSchedulingContextSpecApplyConfiguration) WithPotentialNodes(values ...string) *PodSchedulingContextSpecApplyConfiguration {
|
||||
for i := range values {
|
||||
b.PotentialNodes = append(b.PotentialNodes, values[i])
|
||||
}
|
@ -18,22 +18,22 @@ limitations under the License.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
// PodSchedulingStatusApplyConfiguration represents an declarative configuration of the PodSchedulingStatus type for use
|
||||
// PodSchedulingContextStatusApplyConfiguration represents an declarative configuration of the PodSchedulingContextStatus type for use
|
||||
// with apply.
|
||||
type PodSchedulingStatusApplyConfiguration struct {
|
||||
type PodSchedulingContextStatusApplyConfiguration struct {
|
||||
ResourceClaims []ResourceClaimSchedulingStatusApplyConfiguration `json:"resourceClaims,omitempty"`
|
||||
}
|
||||
|
||||
// PodSchedulingStatusApplyConfiguration constructs an declarative configuration of the PodSchedulingStatus type for use with
|
||||
// PodSchedulingContextStatusApplyConfiguration constructs an declarative configuration of the PodSchedulingContextStatus type for use with
|
||||
// apply.
|
||||
func PodSchedulingStatus() *PodSchedulingStatusApplyConfiguration {
|
||||
return &PodSchedulingStatusApplyConfiguration{}
|
||||
func PodSchedulingContextStatus() *PodSchedulingContextStatusApplyConfiguration {
|
||||
return &PodSchedulingContextStatusApplyConfiguration{}
|
||||
}
|
||||
|
||||
// WithResourceClaims adds the given value to the ResourceClaims field in the declarative configuration
|
||||
// and returns the receiver, so that objects can be build by chaining "With" function invocations.
|
||||
// If called multiple times, values provided by each call will be appended to the ResourceClaims field.
|
||||
func (b *PodSchedulingStatusApplyConfiguration) WithResourceClaims(values ...*ResourceClaimSchedulingStatusApplyConfiguration) *PodSchedulingStatusApplyConfiguration {
|
||||
func (b *PodSchedulingContextStatusApplyConfiguration) WithResourceClaims(values ...*ResourceClaimSchedulingStatusApplyConfiguration) *PodSchedulingContextStatusApplyConfiguration {
|
||||
for i := range values {
|
||||
if values[i] == nil {
|
||||
panic("nil value passed to WithResourceClaims")
|
@ -1451,12 +1451,12 @@ func ForKind(kind schema.GroupVersionKind) interface{} {
|
||||
// Group=resource.k8s.io, Version=v1alpha2
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("AllocationResult"):
|
||||
return &resourcev1alpha2.AllocationResultApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodScheduling"):
|
||||
return &resourcev1alpha2.PodSchedulingApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingSpec"):
|
||||
return &resourcev1alpha2.PodSchedulingSpecApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingStatus"):
|
||||
return &resourcev1alpha2.PodSchedulingStatusApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingContext"):
|
||||
return &resourcev1alpha2.PodSchedulingContextApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingContextSpec"):
|
||||
return &resourcev1alpha2.PodSchedulingContextSpecApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingContextStatus"):
|
||||
return &resourcev1alpha2.PodSchedulingContextStatusApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("ResourceClaim"):
|
||||
return &resourcev1alpha2.ResourceClaimApplyConfiguration{}
|
||||
case v1alpha2.SchemeGroupVersion.WithKind("ResourceClaimConsumerReference"):
|
||||
|
@ -349,8 +349,8 @@ func (f *sharedInformerFactory) ForResource(resource schema.GroupVersionResource
|
||||
return &genericInformer{resource: resource.GroupResource(), informer: f.Rbac().V1beta1().RoleBindings().Informer()}, nil
|
||||
|
||||
// Group=resource.k8s.io, Version=v1alpha2
|
||||
case v1alpha2.SchemeGroupVersion.WithResource("podschedulings"):
|
||||
return &genericInformer{resource: resource.GroupResource(), informer: f.Resource().V1alpha2().PodSchedulings().Informer()}, nil
|
||||
case v1alpha2.SchemeGroupVersion.WithResource("podschedulingcontexts"):
|
||||
return &genericInformer{resource: resource.GroupResource(), informer: f.Resource().V1alpha2().PodSchedulingContexts().Informer()}, nil
|
||||
case v1alpha2.SchemeGroupVersion.WithResource("resourceclaims"):
|
||||
return &genericInformer{resource: resource.GroupResource(), informer: f.Resource().V1alpha2().ResourceClaims().Informer()}, nil
|
||||
case v1alpha2.SchemeGroupVersion.WithResource("resourceclaimtemplates"):
|
||||
|
@ -24,8 +24,8 @@ import (
|
||||
|
||||
// Interface provides access to all the informers in this group version.
|
||||
type Interface interface {
|
||||
// PodSchedulings returns a PodSchedulingInformer.
|
||||
PodSchedulings() PodSchedulingInformer
|
||||
// PodSchedulingContexts returns a PodSchedulingContextInformer.
|
||||
PodSchedulingContexts() PodSchedulingContextInformer
|
||||
// ResourceClaims returns a ResourceClaimInformer.
|
||||
ResourceClaims() ResourceClaimInformer
|
||||
// ResourceClaimTemplates returns a ResourceClaimTemplateInformer.
|
||||
@ -45,9 +45,9 @@ func New(f internalinterfaces.SharedInformerFactory, namespace string, tweakList
|
||||
return &version{factory: f, namespace: namespace, tweakListOptions: tweakListOptions}
|
||||
}
|
||||
|
||||
// PodSchedulings returns a PodSchedulingInformer.
|
||||
func (v *version) PodSchedulings() PodSchedulingInformer {
|
||||
return &podSchedulingInformer{factory: v.factory, namespace: v.namespace, tweakListOptions: v.tweakListOptions}
|
||||
// PodSchedulingContexts returns a PodSchedulingContextInformer.
|
||||
func (v *version) PodSchedulingContexts() PodSchedulingContextInformer {
|
||||
return &podSchedulingContextInformer{factory: v.factory, namespace: v.namespace, tweakListOptions: v.tweakListOptions}
|
||||
}
|
||||
|
||||
// ResourceClaims returns a ResourceClaimInformer.
|
||||
|
@ -32,59 +32,59 @@ import (
|
||||
cache "k8s.io/client-go/tools/cache"
|
||||
)
|
||||
|
||||
// PodSchedulingInformer provides access to a shared informer and lister for
|
||||
// PodSchedulings.
|
||||
type PodSchedulingInformer interface {
|
||||
// PodSchedulingContextInformer provides access to a shared informer and lister for
|
||||
// PodSchedulingContexts.
|
||||
type PodSchedulingContextInformer interface {
|
||||
Informer() cache.SharedIndexInformer
|
||||
Lister() v1alpha2.PodSchedulingLister
|
||||
Lister() v1alpha2.PodSchedulingContextLister
|
||||
}
|
||||
|
||||
type podSchedulingInformer struct {
|
||||
type podSchedulingContextInformer struct {
|
||||
factory internalinterfaces.SharedInformerFactory
|
||||
tweakListOptions internalinterfaces.TweakListOptionsFunc
|
||||
namespace string
|
||||
}
|
||||
|
||||
// NewPodSchedulingInformer constructs a new informer for PodScheduling type.
|
||||
// NewPodSchedulingContextInformer constructs a new informer for PodSchedulingContext type.
|
||||
// Always prefer using an informer factory to get a shared informer instead of getting an independent
|
||||
// one. This reduces memory footprint and number of connections to the server.
|
||||
func NewPodSchedulingInformer(client kubernetes.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers) cache.SharedIndexInformer {
|
||||
return NewFilteredPodSchedulingInformer(client, namespace, resyncPeriod, indexers, nil)
|
||||
func NewPodSchedulingContextInformer(client kubernetes.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers) cache.SharedIndexInformer {
|
||||
return NewFilteredPodSchedulingContextInformer(client, namespace, resyncPeriod, indexers, nil)
|
||||
}
|
||||
|
||||
// NewFilteredPodSchedulingInformer constructs a new informer for PodScheduling type.
|
||||
// NewFilteredPodSchedulingContextInformer constructs a new informer for PodSchedulingContext type.
|
||||
// Always prefer using an informer factory to get a shared informer instead of getting an independent
|
||||
// one. This reduces memory footprint and number of connections to the server.
|
||||
func NewFilteredPodSchedulingInformer(client kubernetes.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers, tweakListOptions internalinterfaces.TweakListOptionsFunc) cache.SharedIndexInformer {
|
||||
func NewFilteredPodSchedulingContextInformer(client kubernetes.Interface, namespace string, resyncPeriod time.Duration, indexers cache.Indexers, tweakListOptions internalinterfaces.TweakListOptionsFunc) cache.SharedIndexInformer {
|
||||
return cache.NewSharedIndexInformer(
|
||||
&cache.ListWatch{
|
||||
ListFunc: func(options v1.ListOptions) (runtime.Object, error) {
|
||||
if tweakListOptions != nil {
|
||||
tweakListOptions(&options)
|
||||
}
|
||||
return client.ResourceV1alpha2().PodSchedulings(namespace).List(context.TODO(), options)
|
||||
return client.ResourceV1alpha2().PodSchedulingContexts(namespace).List(context.TODO(), options)
|
||||
},
|
||||
WatchFunc: func(options v1.ListOptions) (watch.Interface, error) {
|
||||
if tweakListOptions != nil {
|
||||
tweakListOptions(&options)
|
||||
}
|
||||
return client.ResourceV1alpha2().PodSchedulings(namespace).Watch(context.TODO(), options)
|
||||
return client.ResourceV1alpha2().PodSchedulingContexts(namespace).Watch(context.TODO(), options)
|
||||
},
|
||||
},
|
||||
&resourcev1alpha2.PodScheduling{},
|
||||
&resourcev1alpha2.PodSchedulingContext{},
|
||||
resyncPeriod,
|
||||
indexers,
|
||||
)
|
||||
}
|
||||
|
||||
func (f *podSchedulingInformer) defaultInformer(client kubernetes.Interface, resyncPeriod time.Duration) cache.SharedIndexInformer {
|
||||
return NewFilteredPodSchedulingInformer(client, f.namespace, resyncPeriod, cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc}, f.tweakListOptions)
|
||||
func (f *podSchedulingContextInformer) defaultInformer(client kubernetes.Interface, resyncPeriod time.Duration) cache.SharedIndexInformer {
|
||||
return NewFilteredPodSchedulingContextInformer(client, f.namespace, resyncPeriod, cache.Indexers{cache.NamespaceIndex: cache.MetaNamespaceIndexFunc}, f.tweakListOptions)
|
||||
}
|
||||
|
||||
func (f *podSchedulingInformer) Informer() cache.SharedIndexInformer {
|
||||
return f.factory.InformerFor(&resourcev1alpha2.PodScheduling{}, f.defaultInformer)
|
||||
func (f *podSchedulingContextInformer) Informer() cache.SharedIndexInformer {
|
||||
return f.factory.InformerFor(&resourcev1alpha2.PodSchedulingContext{}, f.defaultInformer)
|
||||
}
|
||||
|
||||
func (f *podSchedulingInformer) Lister() v1alpha2.PodSchedulingLister {
|
||||
return v1alpha2.NewPodSchedulingLister(f.Informer().GetIndexer())
|
||||
func (f *podSchedulingContextInformer) Lister() v1alpha2.PodSchedulingContextLister {
|
||||
return v1alpha2.NewPodSchedulingContextLister(f.Informer().GetIndexer())
|
||||
}
|
@ -1,189 +0,0 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by client-gen. DO NOT EDIT.
|
||||
|
||||
package fake
|
||||
|
||||
import (
|
||||
"context"
|
||||
json "encoding/json"
|
||||
"fmt"
|
||||
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
labels "k8s.io/apimachinery/pkg/labels"
|
||||
types "k8s.io/apimachinery/pkg/types"
|
||||
watch "k8s.io/apimachinery/pkg/watch"
|
||||
resourcev1alpha2 "k8s.io/client-go/applyconfigurations/resource/v1alpha2"
|
||||
testing "k8s.io/client-go/testing"
|
||||
)
|
||||
|
||||
// FakePodSchedulings implements PodSchedulingInterface
|
||||
type FakePodSchedulings struct {
|
||||
Fake *FakeResourceV1alpha2
|
||||
ns string
|
||||
}
|
||||
|
||||
var podschedulingsResource = v1alpha2.SchemeGroupVersion.WithResource("podschedulings")
|
||||
|
||||
var podschedulingsKind = v1alpha2.SchemeGroupVersion.WithKind("PodScheduling")
|
||||
|
||||
// Get takes name of the podScheduling, and returns the corresponding podScheduling object, and an error if there is any.
|
||||
func (c *FakePodSchedulings) Get(ctx context.Context, name string, options v1.GetOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewGetAction(podschedulingsResource, c.ns, name), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// List takes label and field selectors, and returns the list of PodSchedulings that match those selectors.
|
||||
func (c *FakePodSchedulings) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha2.PodSchedulingList, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewListAction(podschedulingsResource, podschedulingsKind, c.ns, opts), &v1alpha2.PodSchedulingList{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
label, _, _ := testing.ExtractFromListOptions(opts)
|
||||
if label == nil {
|
||||
label = labels.Everything()
|
||||
}
|
||||
list := &v1alpha2.PodSchedulingList{ListMeta: obj.(*v1alpha2.PodSchedulingList).ListMeta}
|
||||
for _, item := range obj.(*v1alpha2.PodSchedulingList).Items {
|
||||
if label.Matches(labels.Set(item.Labels)) {
|
||||
list.Items = append(list.Items, item)
|
||||
}
|
||||
}
|
||||
return list, err
|
||||
}
|
||||
|
||||
// Watch returns a watch.Interface that watches the requested podSchedulings.
|
||||
func (c *FakePodSchedulings) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {
|
||||
return c.Fake.
|
||||
InvokesWatch(testing.NewWatchAction(podschedulingsResource, c.ns, opts))
|
||||
|
||||
}
|
||||
|
||||
// Create takes the representation of a podScheduling and creates it. Returns the server's representation of the podScheduling, and an error, if there is any.
|
||||
func (c *FakePodSchedulings) Create(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.CreateOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewCreateAction(podschedulingsResource, c.ns, podScheduling), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// Update takes the representation of a podScheduling and updates it. Returns the server's representation of the podScheduling, and an error, if there is any.
|
||||
func (c *FakePodSchedulings) Update(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewUpdateAction(podschedulingsResource, c.ns, podScheduling), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// UpdateStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus().
|
||||
func (c *FakePodSchedulings) UpdateStatus(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (*v1alpha2.PodScheduling, error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewUpdateSubresourceAction(podschedulingsResource, "status", c.ns, podScheduling), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// Delete takes name of the podScheduling and deletes it. Returns an error if one occurs.
|
||||
func (c *FakePodSchedulings) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {
|
||||
_, err := c.Fake.
|
||||
Invokes(testing.NewDeleteActionWithOptions(podschedulingsResource, c.ns, name, opts), &v1alpha2.PodScheduling{})
|
||||
|
||||
return err
|
||||
}
|
||||
|
||||
// DeleteCollection deletes a collection of objects.
|
||||
func (c *FakePodSchedulings) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {
|
||||
action := testing.NewDeleteCollectionAction(podschedulingsResource, c.ns, listOpts)
|
||||
|
||||
_, err := c.Fake.Invokes(action, &v1alpha2.PodSchedulingList{})
|
||||
return err
|
||||
}
|
||||
|
||||
// Patch applies the patch and returns the patched podScheduling.
|
||||
func (c *FakePodSchedulings) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodScheduling, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingsResource, c.ns, name, pt, data, subresources...), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// Apply takes the given apply declarative configuration, applies it and returns the applied podScheduling.
|
||||
func (c *FakePodSchedulings) Apply(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
if podScheduling == nil {
|
||||
return nil, fmt.Errorf("podScheduling provided to Apply must not be nil")
|
||||
}
|
||||
data, err := json.Marshal(podScheduling)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podScheduling.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podScheduling.Name must be provided to Apply")
|
||||
}
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingsResource, c.ns, *name, types.ApplyPatchType, data), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
||||
|
||||
// ApplyStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating ApplyStatus().
|
||||
func (c *FakePodSchedulings) ApplyStatus(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
if podScheduling == nil {
|
||||
return nil, fmt.Errorf("podScheduling provided to Apply must not be nil")
|
||||
}
|
||||
data, err := json.Marshal(podScheduling)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podScheduling.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podScheduling.Name must be provided to Apply")
|
||||
}
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingsResource, c.ns, *name, types.ApplyPatchType, data, "status"), &v1alpha2.PodScheduling{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), err
|
||||
}
|
@ -0,0 +1,189 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by client-gen. DO NOT EDIT.
|
||||
|
||||
package fake
|
||||
|
||||
import (
|
||||
"context"
|
||||
json "encoding/json"
|
||||
"fmt"
|
||||
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
labels "k8s.io/apimachinery/pkg/labels"
|
||||
types "k8s.io/apimachinery/pkg/types"
|
||||
watch "k8s.io/apimachinery/pkg/watch"
|
||||
resourcev1alpha2 "k8s.io/client-go/applyconfigurations/resource/v1alpha2"
|
||||
testing "k8s.io/client-go/testing"
|
||||
)
|
||||
|
||||
// FakePodSchedulingContexts implements PodSchedulingContextInterface
|
||||
type FakePodSchedulingContexts struct {
|
||||
Fake *FakeResourceV1alpha2
|
||||
ns string
|
||||
}
|
||||
|
||||
var podschedulingcontextsResource = v1alpha2.SchemeGroupVersion.WithResource("podschedulingcontexts")
|
||||
|
||||
var podschedulingcontextsKind = v1alpha2.SchemeGroupVersion.WithKind("PodSchedulingContext")
|
||||
|
||||
// Get takes name of the podSchedulingContext, and returns the corresponding podSchedulingContext object, and an error if there is any.
|
||||
func (c *FakePodSchedulingContexts) Get(ctx context.Context, name string, options v1.GetOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewGetAction(podschedulingcontextsResource, c.ns, name), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// List takes label and field selectors, and returns the list of PodSchedulingContexts that match those selectors.
|
||||
func (c *FakePodSchedulingContexts) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha2.PodSchedulingContextList, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewListAction(podschedulingcontextsResource, podschedulingcontextsKind, c.ns, opts), &v1alpha2.PodSchedulingContextList{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
label, _, _ := testing.ExtractFromListOptions(opts)
|
||||
if label == nil {
|
||||
label = labels.Everything()
|
||||
}
|
||||
list := &v1alpha2.PodSchedulingContextList{ListMeta: obj.(*v1alpha2.PodSchedulingContextList).ListMeta}
|
||||
for _, item := range obj.(*v1alpha2.PodSchedulingContextList).Items {
|
||||
if label.Matches(labels.Set(item.Labels)) {
|
||||
list.Items = append(list.Items, item)
|
||||
}
|
||||
}
|
||||
return list, err
|
||||
}
|
||||
|
||||
// Watch returns a watch.Interface that watches the requested podSchedulingContexts.
|
||||
func (c *FakePodSchedulingContexts) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {
|
||||
return c.Fake.
|
||||
InvokesWatch(testing.NewWatchAction(podschedulingcontextsResource, c.ns, opts))
|
||||
|
||||
}
|
||||
|
||||
// Create takes the representation of a podSchedulingContext and creates it. Returns the server's representation of the podSchedulingContext, and an error, if there is any.
|
||||
func (c *FakePodSchedulingContexts) Create(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.CreateOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewCreateAction(podschedulingcontextsResource, c.ns, podSchedulingContext), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// Update takes the representation of a podSchedulingContext and updates it. Returns the server's representation of the podSchedulingContext, and an error, if there is any.
|
||||
func (c *FakePodSchedulingContexts) Update(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewUpdateAction(podschedulingcontextsResource, c.ns, podSchedulingContext), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// UpdateStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus().
|
||||
func (c *FakePodSchedulingContexts) UpdateStatus(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (*v1alpha2.PodSchedulingContext, error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewUpdateSubresourceAction(podschedulingcontextsResource, "status", c.ns, podSchedulingContext), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// Delete takes name of the podSchedulingContext and deletes it. Returns an error if one occurs.
|
||||
func (c *FakePodSchedulingContexts) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {
|
||||
_, err := c.Fake.
|
||||
Invokes(testing.NewDeleteActionWithOptions(podschedulingcontextsResource, c.ns, name, opts), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
return err
|
||||
}
|
||||
|
||||
// DeleteCollection deletes a collection of objects.
|
||||
func (c *FakePodSchedulingContexts) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {
|
||||
action := testing.NewDeleteCollectionAction(podschedulingcontextsResource, c.ns, listOpts)
|
||||
|
||||
_, err := c.Fake.Invokes(action, &v1alpha2.PodSchedulingContextList{})
|
||||
return err
|
||||
}
|
||||
|
||||
// Patch applies the patch and returns the patched podSchedulingContext.
|
||||
func (c *FakePodSchedulingContexts) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingcontextsResource, c.ns, name, pt, data, subresources...), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// Apply takes the given apply declarative configuration, applies it and returns the applied podSchedulingContext.
|
||||
func (c *FakePodSchedulingContexts) Apply(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
if podSchedulingContext == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext provided to Apply must not be nil")
|
||||
}
|
||||
data, err := json.Marshal(podSchedulingContext)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podSchedulingContext.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext.Name must be provided to Apply")
|
||||
}
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingcontextsResource, c.ns, *name, types.ApplyPatchType, data), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
||||
|
||||
// ApplyStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating ApplyStatus().
|
||||
func (c *FakePodSchedulingContexts) ApplyStatus(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
if podSchedulingContext == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext provided to Apply must not be nil")
|
||||
}
|
||||
data, err := json.Marshal(podSchedulingContext)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podSchedulingContext.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext.Name must be provided to Apply")
|
||||
}
|
||||
obj, err := c.Fake.
|
||||
Invokes(testing.NewPatchSubresourceAction(podschedulingcontextsResource, c.ns, *name, types.ApplyPatchType, data, "status"), &v1alpha2.PodSchedulingContext{})
|
||||
|
||||
if obj == nil {
|
||||
return nil, err
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), err
|
||||
}
|
@ -28,8 +28,8 @@ type FakeResourceV1alpha2 struct {
|
||||
*testing.Fake
|
||||
}
|
||||
|
||||
func (c *FakeResourceV1alpha2) PodSchedulings(namespace string) v1alpha2.PodSchedulingInterface {
|
||||
return &FakePodSchedulings{c, namespace}
|
||||
func (c *FakeResourceV1alpha2) PodSchedulingContexts(namespace string) v1alpha2.PodSchedulingContextInterface {
|
||||
return &FakePodSchedulingContexts{c, namespace}
|
||||
}
|
||||
|
||||
func (c *FakeResourceV1alpha2) ResourceClaims(namespace string) v1alpha2.ResourceClaimInterface {
|
||||
|
@ -18,7 +18,7 @@ limitations under the License.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
type PodSchedulingExpansion interface{}
|
||||
type PodSchedulingContextExpansion interface{}
|
||||
|
||||
type ResourceClaimExpansion interface{}
|
||||
|
||||
|
@ -1,256 +0,0 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by client-gen. DO NOT EDIT.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
import (
|
||||
"context"
|
||||
json "encoding/json"
|
||||
"fmt"
|
||||
"time"
|
||||
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
types "k8s.io/apimachinery/pkg/types"
|
||||
watch "k8s.io/apimachinery/pkg/watch"
|
||||
resourcev1alpha2 "k8s.io/client-go/applyconfigurations/resource/v1alpha2"
|
||||
scheme "k8s.io/client-go/kubernetes/scheme"
|
||||
rest "k8s.io/client-go/rest"
|
||||
)
|
||||
|
||||
// PodSchedulingsGetter has a method to return a PodSchedulingInterface.
|
||||
// A group's client should implement this interface.
|
||||
type PodSchedulingsGetter interface {
|
||||
PodSchedulings(namespace string) PodSchedulingInterface
|
||||
}
|
||||
|
||||
// PodSchedulingInterface has methods to work with PodScheduling resources.
|
||||
type PodSchedulingInterface interface {
|
||||
Create(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.CreateOptions) (*v1alpha2.PodScheduling, error)
|
||||
Update(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (*v1alpha2.PodScheduling, error)
|
||||
UpdateStatus(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (*v1alpha2.PodScheduling, error)
|
||||
Delete(ctx context.Context, name string, opts v1.DeleteOptions) error
|
||||
DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error
|
||||
Get(ctx context.Context, name string, opts v1.GetOptions) (*v1alpha2.PodScheduling, error)
|
||||
List(ctx context.Context, opts v1.ListOptions) (*v1alpha2.PodSchedulingList, error)
|
||||
Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error)
|
||||
Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodScheduling, err error)
|
||||
Apply(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error)
|
||||
ApplyStatus(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error)
|
||||
PodSchedulingExpansion
|
||||
}
|
||||
|
||||
// podSchedulings implements PodSchedulingInterface
|
||||
type podSchedulings struct {
|
||||
client rest.Interface
|
||||
ns string
|
||||
}
|
||||
|
||||
// newPodSchedulings returns a PodSchedulings
|
||||
func newPodSchedulings(c *ResourceV1alpha2Client, namespace string) *podSchedulings {
|
||||
return &podSchedulings{
|
||||
client: c.RESTClient(),
|
||||
ns: namespace,
|
||||
}
|
||||
}
|
||||
|
||||
// Get takes name of the podScheduling, and returns the corresponding podScheduling object, and an error if there is any.
|
||||
func (c *podSchedulings) Get(ctx context.Context, name string, options v1.GetOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(name).
|
||||
VersionedParams(&options, scheme.ParameterCodec).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// List takes label and field selectors, and returns the list of PodSchedulings that match those selectors.
|
||||
func (c *podSchedulings) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha2.PodSchedulingList, err error) {
|
||||
var timeout time.Duration
|
||||
if opts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*opts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
result = &v1alpha2.PodSchedulingList{}
|
||||
err = c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Watch returns a watch.Interface that watches the requested podSchedulings.
|
||||
func (c *podSchedulings) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {
|
||||
var timeout time.Duration
|
||||
if opts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*opts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
opts.Watch = true
|
||||
return c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Watch(ctx)
|
||||
}
|
||||
|
||||
// Create takes the representation of a podScheduling and creates it. Returns the server's representation of the podScheduling, and an error, if there is any.
|
||||
func (c *podSchedulings) Create(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.CreateOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Post().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podScheduling).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Update takes the representation of a podScheduling and updates it. Returns the server's representation of the podScheduling, and an error, if there is any.
|
||||
func (c *podSchedulings) Update(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Put().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(podScheduling.Name).
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podScheduling).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// UpdateStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus().
|
||||
func (c *podSchedulings) UpdateStatus(ctx context.Context, podScheduling *v1alpha2.PodScheduling, opts v1.UpdateOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Put().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(podScheduling.Name).
|
||||
SubResource("status").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podScheduling).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Delete takes name of the podScheduling and deletes it. Returns an error if one occurs.
|
||||
func (c *podSchedulings) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {
|
||||
return c.client.Delete().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(name).
|
||||
Body(&opts).
|
||||
Do(ctx).
|
||||
Error()
|
||||
}
|
||||
|
||||
// DeleteCollection deletes a collection of objects.
|
||||
func (c *podSchedulings) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {
|
||||
var timeout time.Duration
|
||||
if listOpts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
return c.client.Delete().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
VersionedParams(&listOpts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Body(&opts).
|
||||
Do(ctx).
|
||||
Error()
|
||||
}
|
||||
|
||||
// Patch applies the patch and returns the patched podScheduling.
|
||||
func (c *podSchedulings) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodScheduling, err error) {
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Patch(pt).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(name).
|
||||
SubResource(subresources...).
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Apply takes the given apply declarative configuration, applies it and returns the applied podScheduling.
|
||||
func (c *podSchedulings) Apply(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
if podScheduling == nil {
|
||||
return nil, fmt.Errorf("podScheduling provided to Apply must not be nil")
|
||||
}
|
||||
patchOpts := opts.ToPatchOptions()
|
||||
data, err := json.Marshal(podScheduling)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podScheduling.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podScheduling.Name must be provided to Apply")
|
||||
}
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Patch(types.ApplyPatchType).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(*name).
|
||||
VersionedParams(&patchOpts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// ApplyStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating ApplyStatus().
|
||||
func (c *podSchedulings) ApplyStatus(ctx context.Context, podScheduling *resourcev1alpha2.PodSchedulingApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodScheduling, err error) {
|
||||
if podScheduling == nil {
|
||||
return nil, fmt.Errorf("podScheduling provided to Apply must not be nil")
|
||||
}
|
||||
patchOpts := opts.ToPatchOptions()
|
||||
data, err := json.Marshal(podScheduling)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
name := podScheduling.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podScheduling.Name must be provided to Apply")
|
||||
}
|
||||
|
||||
result = &v1alpha2.PodScheduling{}
|
||||
err = c.client.Patch(types.ApplyPatchType).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulings").
|
||||
Name(*name).
|
||||
SubResource("status").
|
||||
VersionedParams(&patchOpts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
@ -0,0 +1,256 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by client-gen. DO NOT EDIT.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
import (
|
||||
"context"
|
||||
json "encoding/json"
|
||||
"fmt"
|
||||
"time"
|
||||
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
v1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
||||
types "k8s.io/apimachinery/pkg/types"
|
||||
watch "k8s.io/apimachinery/pkg/watch"
|
||||
resourcev1alpha2 "k8s.io/client-go/applyconfigurations/resource/v1alpha2"
|
||||
scheme "k8s.io/client-go/kubernetes/scheme"
|
||||
rest "k8s.io/client-go/rest"
|
||||
)
|
||||
|
||||
// PodSchedulingContextsGetter has a method to return a PodSchedulingContextInterface.
|
||||
// A group's client should implement this interface.
|
||||
type PodSchedulingContextsGetter interface {
|
||||
PodSchedulingContexts(namespace string) PodSchedulingContextInterface
|
||||
}
|
||||
|
||||
// PodSchedulingContextInterface has methods to work with PodSchedulingContext resources.
|
||||
type PodSchedulingContextInterface interface {
|
||||
Create(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.CreateOptions) (*v1alpha2.PodSchedulingContext, error)
|
||||
Update(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (*v1alpha2.PodSchedulingContext, error)
|
||||
UpdateStatus(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (*v1alpha2.PodSchedulingContext, error)
|
||||
Delete(ctx context.Context, name string, opts v1.DeleteOptions) error
|
||||
DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error
|
||||
Get(ctx context.Context, name string, opts v1.GetOptions) (*v1alpha2.PodSchedulingContext, error)
|
||||
List(ctx context.Context, opts v1.ListOptions) (*v1alpha2.PodSchedulingContextList, error)
|
||||
Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error)
|
||||
Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodSchedulingContext, err error)
|
||||
Apply(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error)
|
||||
ApplyStatus(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error)
|
||||
PodSchedulingContextExpansion
|
||||
}
|
||||
|
||||
// podSchedulingContexts implements PodSchedulingContextInterface
|
||||
type podSchedulingContexts struct {
|
||||
client rest.Interface
|
||||
ns string
|
||||
}
|
||||
|
||||
// newPodSchedulingContexts returns a PodSchedulingContexts
|
||||
func newPodSchedulingContexts(c *ResourceV1alpha2Client, namespace string) *podSchedulingContexts {
|
||||
return &podSchedulingContexts{
|
||||
client: c.RESTClient(),
|
||||
ns: namespace,
|
||||
}
|
||||
}
|
||||
|
||||
// Get takes name of the podSchedulingContext, and returns the corresponding podSchedulingContext object, and an error if there is any.
|
||||
func (c *podSchedulingContexts) Get(ctx context.Context, name string, options v1.GetOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(name).
|
||||
VersionedParams(&options, scheme.ParameterCodec).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// List takes label and field selectors, and returns the list of PodSchedulingContexts that match those selectors.
|
||||
func (c *podSchedulingContexts) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha2.PodSchedulingContextList, err error) {
|
||||
var timeout time.Duration
|
||||
if opts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*opts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
result = &v1alpha2.PodSchedulingContextList{}
|
||||
err = c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Watch returns a watch.Interface that watches the requested podSchedulingContexts.
|
||||
func (c *podSchedulingContexts) Watch(ctx context.Context, opts v1.ListOptions) (watch.Interface, error) {
|
||||
var timeout time.Duration
|
||||
if opts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*opts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
opts.Watch = true
|
||||
return c.client.Get().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Watch(ctx)
|
||||
}
|
||||
|
||||
// Create takes the representation of a podSchedulingContext and creates it. Returns the server's representation of the podSchedulingContext, and an error, if there is any.
|
||||
func (c *podSchedulingContexts) Create(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.CreateOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Post().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podSchedulingContext).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Update takes the representation of a podSchedulingContext and updates it. Returns the server's representation of the podSchedulingContext, and an error, if there is any.
|
||||
func (c *podSchedulingContexts) Update(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Put().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(podSchedulingContext.Name).
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podSchedulingContext).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// UpdateStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating UpdateStatus().
|
||||
func (c *podSchedulingContexts) UpdateStatus(ctx context.Context, podSchedulingContext *v1alpha2.PodSchedulingContext, opts v1.UpdateOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Put().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(podSchedulingContext.Name).
|
||||
SubResource("status").
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(podSchedulingContext).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Delete takes name of the podSchedulingContext and deletes it. Returns an error if one occurs.
|
||||
func (c *podSchedulingContexts) Delete(ctx context.Context, name string, opts v1.DeleteOptions) error {
|
||||
return c.client.Delete().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(name).
|
||||
Body(&opts).
|
||||
Do(ctx).
|
||||
Error()
|
||||
}
|
||||
|
||||
// DeleteCollection deletes a collection of objects.
|
||||
func (c *podSchedulingContexts) DeleteCollection(ctx context.Context, opts v1.DeleteOptions, listOpts v1.ListOptions) error {
|
||||
var timeout time.Duration
|
||||
if listOpts.TimeoutSeconds != nil {
|
||||
timeout = time.Duration(*listOpts.TimeoutSeconds) * time.Second
|
||||
}
|
||||
return c.client.Delete().
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
VersionedParams(&listOpts, scheme.ParameterCodec).
|
||||
Timeout(timeout).
|
||||
Body(&opts).
|
||||
Do(ctx).
|
||||
Error()
|
||||
}
|
||||
|
||||
// Patch applies the patch and returns the patched podSchedulingContext.
|
||||
func (c *podSchedulingContexts) Patch(ctx context.Context, name string, pt types.PatchType, data []byte, opts v1.PatchOptions, subresources ...string) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Patch(pt).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(name).
|
||||
SubResource(subresources...).
|
||||
VersionedParams(&opts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// Apply takes the given apply declarative configuration, applies it and returns the applied podSchedulingContext.
|
||||
func (c *podSchedulingContexts) Apply(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
if podSchedulingContext == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext provided to Apply must not be nil")
|
||||
}
|
||||
patchOpts := opts.ToPatchOptions()
|
||||
data, err := json.Marshal(podSchedulingContext)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
name := podSchedulingContext.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext.Name must be provided to Apply")
|
||||
}
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Patch(types.ApplyPatchType).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(*name).
|
||||
VersionedParams(&patchOpts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
||||
|
||||
// ApplyStatus was generated because the type contains a Status member.
|
||||
// Add a +genclient:noStatus comment above the type to avoid generating ApplyStatus().
|
||||
func (c *podSchedulingContexts) ApplyStatus(ctx context.Context, podSchedulingContext *resourcev1alpha2.PodSchedulingContextApplyConfiguration, opts v1.ApplyOptions) (result *v1alpha2.PodSchedulingContext, err error) {
|
||||
if podSchedulingContext == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext provided to Apply must not be nil")
|
||||
}
|
||||
patchOpts := opts.ToPatchOptions()
|
||||
data, err := json.Marshal(podSchedulingContext)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
|
||||
name := podSchedulingContext.Name
|
||||
if name == nil {
|
||||
return nil, fmt.Errorf("podSchedulingContext.Name must be provided to Apply")
|
||||
}
|
||||
|
||||
result = &v1alpha2.PodSchedulingContext{}
|
||||
err = c.client.Patch(types.ApplyPatchType).
|
||||
Namespace(c.ns).
|
||||
Resource("podschedulingcontexts").
|
||||
Name(*name).
|
||||
SubResource("status").
|
||||
VersionedParams(&patchOpts, scheme.ParameterCodec).
|
||||
Body(data).
|
||||
Do(ctx).
|
||||
Into(result)
|
||||
return
|
||||
}
|
@ -28,7 +28,7 @@ import (
|
||||
|
||||
type ResourceV1alpha2Interface interface {
|
||||
RESTClient() rest.Interface
|
||||
PodSchedulingsGetter
|
||||
PodSchedulingContextsGetter
|
||||
ResourceClaimsGetter
|
||||
ResourceClaimTemplatesGetter
|
||||
ResourceClassesGetter
|
||||
@ -39,8 +39,8 @@ type ResourceV1alpha2Client struct {
|
||||
restClient rest.Interface
|
||||
}
|
||||
|
||||
func (c *ResourceV1alpha2Client) PodSchedulings(namespace string) PodSchedulingInterface {
|
||||
return newPodSchedulings(c, namespace)
|
||||
func (c *ResourceV1alpha2Client) PodSchedulingContexts(namespace string) PodSchedulingContextInterface {
|
||||
return newPodSchedulingContexts(c, namespace)
|
||||
}
|
||||
|
||||
func (c *ResourceV1alpha2Client) ResourceClaims(namespace string) ResourceClaimInterface {
|
||||
|
@ -18,13 +18,13 @@ limitations under the License.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
// PodSchedulingListerExpansion allows custom methods to be added to
|
||||
// PodSchedulingLister.
|
||||
type PodSchedulingListerExpansion interface{}
|
||||
// PodSchedulingContextListerExpansion allows custom methods to be added to
|
||||
// PodSchedulingContextLister.
|
||||
type PodSchedulingContextListerExpansion interface{}
|
||||
|
||||
// PodSchedulingNamespaceListerExpansion allows custom methods to be added to
|
||||
// PodSchedulingNamespaceLister.
|
||||
type PodSchedulingNamespaceListerExpansion interface{}
|
||||
// PodSchedulingContextNamespaceListerExpansion allows custom methods to be added to
|
||||
// PodSchedulingContextNamespaceLister.
|
||||
type PodSchedulingContextNamespaceListerExpansion interface{}
|
||||
|
||||
// ResourceClaimListerExpansion allows custom methods to be added to
|
||||
// ResourceClaimLister.
|
||||
|
@ -1,99 +0,0 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by lister-gen. DO NOT EDIT.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
import (
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
"k8s.io/apimachinery/pkg/api/errors"
|
||||
"k8s.io/apimachinery/pkg/labels"
|
||||
"k8s.io/client-go/tools/cache"
|
||||
)
|
||||
|
||||
// PodSchedulingLister helps list PodSchedulings.
|
||||
// All objects returned here must be treated as read-only.
|
||||
type PodSchedulingLister interface {
|
||||
// List lists all PodSchedulings in the indexer.
|
||||
// Objects returned here must be treated as read-only.
|
||||
List(selector labels.Selector) (ret []*v1alpha2.PodScheduling, err error)
|
||||
// PodSchedulings returns an object that can list and get PodSchedulings.
|
||||
PodSchedulings(namespace string) PodSchedulingNamespaceLister
|
||||
PodSchedulingListerExpansion
|
||||
}
|
||||
|
||||
// podSchedulingLister implements the PodSchedulingLister interface.
|
||||
type podSchedulingLister struct {
|
||||
indexer cache.Indexer
|
||||
}
|
||||
|
||||
// NewPodSchedulingLister returns a new PodSchedulingLister.
|
||||
func NewPodSchedulingLister(indexer cache.Indexer) PodSchedulingLister {
|
||||
return &podSchedulingLister{indexer: indexer}
|
||||
}
|
||||
|
||||
// List lists all PodSchedulings in the indexer.
|
||||
func (s *podSchedulingLister) List(selector labels.Selector) (ret []*v1alpha2.PodScheduling, err error) {
|
||||
err = cache.ListAll(s.indexer, selector, func(m interface{}) {
|
||||
ret = append(ret, m.(*v1alpha2.PodScheduling))
|
||||
})
|
||||
return ret, err
|
||||
}
|
||||
|
||||
// PodSchedulings returns an object that can list and get PodSchedulings.
|
||||
func (s *podSchedulingLister) PodSchedulings(namespace string) PodSchedulingNamespaceLister {
|
||||
return podSchedulingNamespaceLister{indexer: s.indexer, namespace: namespace}
|
||||
}
|
||||
|
||||
// PodSchedulingNamespaceLister helps list and get PodSchedulings.
|
||||
// All objects returned here must be treated as read-only.
|
||||
type PodSchedulingNamespaceLister interface {
|
||||
// List lists all PodSchedulings in the indexer for a given namespace.
|
||||
// Objects returned here must be treated as read-only.
|
||||
List(selector labels.Selector) (ret []*v1alpha2.PodScheduling, err error)
|
||||
// Get retrieves the PodScheduling from the indexer for a given namespace and name.
|
||||
// Objects returned here must be treated as read-only.
|
||||
Get(name string) (*v1alpha2.PodScheduling, error)
|
||||
PodSchedulingNamespaceListerExpansion
|
||||
}
|
||||
|
||||
// podSchedulingNamespaceLister implements the PodSchedulingNamespaceLister
|
||||
// interface.
|
||||
type podSchedulingNamespaceLister struct {
|
||||
indexer cache.Indexer
|
||||
namespace string
|
||||
}
|
||||
|
||||
// List lists all PodSchedulings in the indexer for a given namespace.
|
||||
func (s podSchedulingNamespaceLister) List(selector labels.Selector) (ret []*v1alpha2.PodScheduling, err error) {
|
||||
err = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) {
|
||||
ret = append(ret, m.(*v1alpha2.PodScheduling))
|
||||
})
|
||||
return ret, err
|
||||
}
|
||||
|
||||
// Get retrieves the PodScheduling from the indexer for a given namespace and name.
|
||||
func (s podSchedulingNamespaceLister) Get(name string) (*v1alpha2.PodScheduling, error) {
|
||||
obj, exists, err := s.indexer.GetByKey(s.namespace + "/" + name)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if !exists {
|
||||
return nil, errors.NewNotFound(v1alpha2.Resource("podscheduling"), name)
|
||||
}
|
||||
return obj.(*v1alpha2.PodScheduling), nil
|
||||
}
|
@ -0,0 +1,99 @@
|
||||
/*
|
||||
Copyright The Kubernetes Authors.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
// Code generated by lister-gen. DO NOT EDIT.
|
||||
|
||||
package v1alpha2
|
||||
|
||||
import (
|
||||
v1alpha2 "k8s.io/api/resource/v1alpha2"
|
||||
"k8s.io/apimachinery/pkg/api/errors"
|
||||
"k8s.io/apimachinery/pkg/labels"
|
||||
"k8s.io/client-go/tools/cache"
|
||||
)
|
||||
|
||||
// PodSchedulingContextLister helps list PodSchedulingContexts.
|
||||
// All objects returned here must be treated as read-only.
|
||||
type PodSchedulingContextLister interface {
|
||||
// List lists all PodSchedulingContexts in the indexer.
|
||||
// Objects returned here must be treated as read-only.
|
||||
List(selector labels.Selector) (ret []*v1alpha2.PodSchedulingContext, err error)
|
||||
// PodSchedulingContexts returns an object that can list and get PodSchedulingContexts.
|
||||
PodSchedulingContexts(namespace string) PodSchedulingContextNamespaceLister
|
||||
PodSchedulingContextListerExpansion
|
||||
}
|
||||
|
||||
// podSchedulingContextLister implements the PodSchedulingContextLister interface.
|
||||
type podSchedulingContextLister struct {
|
||||
indexer cache.Indexer
|
||||
}
|
||||
|
||||
// NewPodSchedulingContextLister returns a new PodSchedulingContextLister.
|
||||
func NewPodSchedulingContextLister(indexer cache.Indexer) PodSchedulingContextLister {
|
||||
return &podSchedulingContextLister{indexer: indexer}
|
||||
}
|
||||
|
||||
// List lists all PodSchedulingContexts in the indexer.
|
||||
func (s *podSchedulingContextLister) List(selector labels.Selector) (ret []*v1alpha2.PodSchedulingContext, err error) {
|
||||
err = cache.ListAll(s.indexer, selector, func(m interface{}) {
|
||||
ret = append(ret, m.(*v1alpha2.PodSchedulingContext))
|
||||
})
|
||||
return ret, err
|
||||
}
|
||||
|
||||
// PodSchedulingContexts returns an object that can list and get PodSchedulingContexts.
|
||||
func (s *podSchedulingContextLister) PodSchedulingContexts(namespace string) PodSchedulingContextNamespaceLister {
|
||||
return podSchedulingContextNamespaceLister{indexer: s.indexer, namespace: namespace}
|
||||
}
|
||||
|
||||
// PodSchedulingContextNamespaceLister helps list and get PodSchedulingContexts.
|
||||
// All objects returned here must be treated as read-only.
|
||||
type PodSchedulingContextNamespaceLister interface {
|
||||
// List lists all PodSchedulingContexts in the indexer for a given namespace.
|
||||
// Objects returned here must be treated as read-only.
|
||||
List(selector labels.Selector) (ret []*v1alpha2.PodSchedulingContext, err error)
|
||||
// Get retrieves the PodSchedulingContext from the indexer for a given namespace and name.
|
||||
// Objects returned here must be treated as read-only.
|
||||
Get(name string) (*v1alpha2.PodSchedulingContext, error)
|
||||
PodSchedulingContextNamespaceListerExpansion
|
||||
}
|
||||
|
||||
// podSchedulingContextNamespaceLister implements the PodSchedulingContextNamespaceLister
|
||||
// interface.
|
||||
type podSchedulingContextNamespaceLister struct {
|
||||
indexer cache.Indexer
|
||||
namespace string
|
||||
}
|
||||
|
||||
// List lists all PodSchedulingContexts in the indexer for a given namespace.
|
||||
func (s podSchedulingContextNamespaceLister) List(selector labels.Selector) (ret []*v1alpha2.PodSchedulingContext, err error) {
|
||||
err = cache.ListAllByNamespace(s.indexer, s.namespace, selector, func(m interface{}) {
|
||||
ret = append(ret, m.(*v1alpha2.PodSchedulingContext))
|
||||
})
|
||||
return ret, err
|
||||
}
|
||||
|
||||
// Get retrieves the PodSchedulingContext from the indexer for a given namespace and name.
|
||||
func (s podSchedulingContextNamespaceLister) Get(name string) (*v1alpha2.PodSchedulingContext, error) {
|
||||
obj, exists, err := s.indexer.GetByKey(s.namespace + "/" + name)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if !exists {
|
||||
return nil, errors.NewNotFound(v1alpha2.Resource("podschedulingcontext"), name)
|
||||
}
|
||||
return obj.(*v1alpha2.PodSchedulingContext), nil
|
||||
}
|
Loading…
Reference in New Issue
Block a user