k8s之所有resources类型

以下组件基于1.19版本
可以使用kubectl api-resource

nameshort nameAPI groupnamespacekind
bingdstruebing
componentStatusescsfalseComponetStatus
configMapscmtrueConfigMap
endpointseptrueEndpoints
eventsevtrueEvent
limitrangeslimitstrueLimitRange
namepspacesnsfalseNamespace
nodesnofalseNode
persistentvolumeclaimspvctruePersistentVolumeClaim
persistentvolumespvfalsePersistentVolume
poddspotruePod
podtemplatestruePodTemplate
replicationcontrollersrctrueReplicationController
resourcequotasquotatrueResourceQuota
secretstrueSecret
serviceaccountssatrueServiceAccount
servicessvctrueService
mutatingwebhookconfigurationsadminssionregistration.k8s.iofalseMutatingWebHookConfiguration
validatingwebhookconfigurationsadminssionregistration.k8s.iofalseValidatingWebhookConfiguration
customresourcedefinitionsapiextensions.k8s.iofalseCustomResourceDefinition
apiservicesapiregistration.k8s.iofalseAPIService
controllerrevisionsappstrueControllerRevision
deamonsetsdsappstrueDeamonSet
deploymentsdeployappstrueDeployment
replicasetsrsappstrueReplicaSet
statefulsetsstsappstrueStatefulSet
tokenreviewsauthentication.k8s.iofalseTokenReview
localsubjectaccessreviewsauthentication.k8s.iotrueLocalSubjectAccessReview
selsubjectaccessreviewsauthentication.k8s.iofalseSelfSubjectAccessReview
subjectaccessreviewsauthentication.k8s.iofalseSubjectAccessReview
horizontalpodauscalershpa.autoscalingtrueHorizontalPodAutoscaler
cronjobscjbatchtrueCronjob
jobsbatchtrueJob
certificatessigningrequestscsrcertificates.k8s.iotrueCertificateSigningRequest
leasecoordination.k8s.iotrueLease
endpointslicesdiscovery.k8s.iotrueEndpointSlice
eventsevevents.k8s.iotrueEvent
flowschemasflowcontrol.apiserver.k8s.iofalseFlowSchema
prioritylevelconfigurationsflowcontrol.apiserver.k8s.iofalsePriorityLevelConfiguration
ingressclassesnetworking.k8s.iofalseIngressClass
ingressesingnetworking.k8s.iotrueIngress
networkpoliciesnetpolnetworking.k8s.iotrueNetworkPolicy
runtimeclassesnode.k8s.iofalseRuntimeClass
poddisruptionbugetspdbpolicytruePodDisruptionBudget
podsecuritypoliciespsppolicyfalsePodSecurityPolicy
clusterrolebindingsrbac.authorization.k8s.iofalseClusterRoleBinding
clusterrolesrbac.authorization.k8s.iofalseClusterRole
rolebindingsrbac.authorization.k8s.iotrueRoleBinding
rolesrbac.authorization.k8s.iotrueRole
priorityclassespcscheduling.k8s.iofalsePriorityClass
csidriversstorage.k8s.iofalseCSIDriver
csinodesstorage.k8s.iofalseCSINode
storageclassesscstorage.k8s.iofalseStoragetClass
volumeattachmentstorage.k8s.iofalseVolumeAttachment

我们可以看下一个刚刚安装的集群包含了哪些组件

~# kubectl get all -A
NAMESPACE              NAME                                             READY   STATUS    RESTARTS   AGE
kube-system            pod/calico-kube-controllers-8b5ff5d58-szbwh      1/1     Running   3          51d
kube-system            pod/calico-node-m2cfp                            1/1     Running   1          51d
kube-system            pod/calico-node-pwgkr                            1/1     Running   1          51d
kube-system            pod/calico-node-xbxnq                            1/1     Running   2          51d
kube-system            pod/coredns-85967d65-k2pnp                       1/1     Running   1          51d
kube-system            pod/coredns-85967d65-zkj27                       1/1     Running   1          51d
kube-system            pod/dns-autoscaler-5b7b5c9b6f-8w4vh              1/1     Running   1          51d
kube-system            pod/kube-apiserver-node1                         1/1     Running   1          51d
kube-system            pod/kube-apiserver-node2                         1/1     Running   1          51d
kube-system            pod/kube-controller-manager-node1                1/1     Running   46         51d
kube-system            pod/kube-controller-manager-node2                1/1     Running   51         51d
kube-system            pod/kube-proxy-4vv6x                             1/1     Running   1          51d
kube-system            pod/kube-proxy-dc8zm                             1/1     Running   1          51d
kube-system            pod/kube-proxy-g256z                             1/1     Running   2          51d
kube-system            pod/kube-scheduler-node1                         1/1     Running   44         51d
kube-system            pod/kube-scheduler-node2                         1/1     Running   48         51d
kube-system            pod/nginx-proxy-node3                            1/1     Running   2          51d
kube-system            pod/nodelocaldns-9469d                           1/1     Running   1          51d
kube-system            pod/nodelocaldns-kz8f2                           1/1     Running   1          51d
kube-system            pod/nodelocaldns-ld4n4                           1/1     Running   2          51d
kubernetes-dashboard   pod/dashboard-metrics-scraper-7b59f7d4df-bm742   1/1     Running   1          49d
kubernetes-dashboard   pod/kubernetes-dashboard-997f4979d-xmbfn         1/1     Running   1          49d

NAMESPACE              NAME                                TYPE        CLUSTER-IP      EXTERNAL-IP   PORT(S)                  AGE
default                service/kubernetes                  ClusterIP   10.233.0.1      <none>        443/TCP                  51d
kube-system            service/coredns                     ClusterIP   10.233.0.3      <none>        53/UDP,53/TCP,9153/TCP   51d
kubernetes-dashboard   service/dashboard-metrics-scraper   ClusterIP   10.233.54.163   <none>        8000/TCP                 49d
kubernetes-dashboard   service/kubernetes-dashboard        NodePort    10.233.36.214   <none>        443:31241/TCP            49d

NAMESPACE     NAME                          DESIRED   CURRENT   READY   UP-TO-DATE   AVAILABLE   NODE SELECTOR            AGE
kube-system   daemonset.apps/calico-node    3         3         3       3            3           <none>                   51d
kube-system   daemonset.apps/kube-proxy     3         3         3       3            3           kubernetes.io/os=linux   51d
kube-system   daemonset.apps/nodelocaldns   3         3         3       3            3           <none>                   51d

NAMESPACE              NAME                                        READY   UP-TO-DATE   AVAILABLE   AGE
kube-system            deployment.apps/calico-kube-controllers     1/1     1            1           51d
kube-system            deployment.apps/coredns                     2/2     2            2           51d
kube-system            deployment.apps/dns-autoscaler              1/1     1            1           51d
kubernetes-dashboard   deployment.apps/dashboard-metrics-scraper   1/1     1            1           49d
kubernetes-dashboard   deployment.apps/kubernetes-dashboard        1/1     1            1           49d

NAMESPACE              NAME                                                   DESIRED   CURRENT   READY   AGE
kube-system            replicaset.apps/calico-kube-controllers-8b5ff5d58      1         1         1       51d
kube-system            replicaset.apps/coredns-85967d65                       2         2         2       51d
kube-system            replicaset.apps/dns-autoscaler-5b7b5c9b6f              1         1         1       51d
kubernetes-dashboard   replicaset.apps/dashboard-metrics-scraper-7b59f7d4df   1         1         1       49d
kubernetes-dashboard   replicaset.apps/kubernetes-dashboard-74d688b6bc        0         0         0       49d
kubernetes-dashboard   replicaset.apps/kubernetes-dashboard-997f4979d         1         1         1       49d

我们发现上面有好几个组件曾经restart过,我们可以查看下详情看看什么原因呗重启了,我们使用以下命令查看上一次错误原因

kubectl descirbe pod/kube-scheduler-node1
State:          Running
      Started:      Tue, 06 Apr 2021 02:49:36 +0800
    Last State:     Terminated
      Reason:       Error
      Exit Code:    255
      Started:      Tue, 06 Apr 2021 02:04:10 +0800
      Finished:     Tue, 06 Apr 2021 02:49:35 +0800

我们发现除了 Error没其它有用信息了,没关系我们可以查看上一次镜像日志使用-p参数,执行以下命令查看上一次容器日志

kubectl logs pod/kube-scheduler-node1 -n kube-system -p |grep error

如果显示行信息不够可以增加 grep -C10 参数查看前后10行
输出如下

:~# kubectl logs pod/kube-scheduler-node1 -n kube-system -p |grep error
E0405 18:44:55.781090       1 leaderelection.go:325] error retrieving resource lock kube-system/kube-scheduler: Get "https://x.x.x.x:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-scheduler?timeout=10s": net/http: request canceled (Client.Timeout exceeded while awaiting headers)

查询了下原因stackoverflow找到答案如下

  • 1
    点赞
  • 5
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值