You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Machine ID: "ba9e6982836321b38e631faf28531fce18b41a13c8010cddc8be9802f057360b" (from machine) (pass)
Total memory: 251.8 GiB (pass)
Disk space available for /var/lib/k0s: 398.2 GiB (pass)
Name resolution: localhost: [127.0.0.1] (pass)
Operating system: Linux (pass)
Linux kernel release: 6.1.0-20-amd64 (pass)
Max. file descriptors per process: current: 1048576 / max: 1048576 (pass)
AppArmor: active (pass)
Executable in PATH: modprobe: /usr/sbin/modprobe (pass)
Executable in PATH: mount: /usr/bin/mount (pass)
Executable in PATH: umount: /usr/bin/umount (pass)
/proc file system: mounted (0x9fa0) (pass)
Control Groups: version 2 (pass)
cgroup controller "cpu": available (is a listed root controller) (pass)
cgroup controller "cpuacct": available (via cpu in version 2) (pass)
cgroup controller "cpuset": available (is a listed root controller) (pass)
cgroup controller "memory": available (is a listed root controller) (pass)
cgroup controller "devices": available (device filters attachable) (pass)
cgroup controller "freezer": available (cgroup.freeze exists) (pass)
cgroup controller "pids": available (is a listed root controller) (pass)
cgroup controller "hugetlb": available (is a listed root controller) (pass)
cgroup controller "blkio": available (via io in version 2) (pass)
CONFIG_CGROUPS: Control Group support: built-in (pass)
CONFIG_CGROUP_FREEZER: Freezer cgroup subsystem: built-in (pass)
CONFIG_CGROUP_PIDS: PIDs cgroup subsystem: built-in (pass)
CONFIG_CGROUP_DEVICE: Device controller for cgroups: built-in (pass)
CONFIG_CPUSETS: Cpuset support: built-in (pass)
CONFIG_CGROUP_CPUACCT: Simple CPU accounting cgroup subsystem: built-in (pass)
CONFIG_MEMCG: Memory Resource Controller for Control Groups: built-in (pass)
CONFIG_CGROUP_HUGETLB: HugeTLB Resource Controller for Control Groups: built-in (pass)
CONFIG_CGROUP_SCHED: Group CPU scheduler: built-in (pass)
CONFIG_FAIR_GROUP_SCHED: Group scheduling for SCHED_OTHER: built-in (pass)
CONFIG_CFS_BANDWIDTH: CPU bandwidth provisioning for FAIR_GROUP_SCHED: built-in (pass)
CONFIG_BLK_CGROUP: Block IO controller: built-in (pass)
CONFIG_NAMESPACES: Namespaces support: built-in (pass)
CONFIG_UTS_NS: UTS namespace: built-in (pass)
CONFIG_IPC_NS: IPC namespace: built-in (pass)
CONFIG_PID_NS: PID namespace: built-in (pass)
CONFIG_NET_NS: Network namespace: built-in (pass)
CONFIG_NET: Networking support: built-in (pass)
CONFIG_INET: TCP/IP networking: built-in (pass)
CONFIG_IPV6: The IPv6 protocol: built-in (pass)
CONFIG_NETFILTER: Network packet filtering framework (Netfilter): built-in (pass)
CONFIG_NETFILTER_ADVANCED: Advanced netfilter configuration: built-in (pass)
CONFIG_NF_CONNTRACK: Netfilter connection tracking support: module (pass)
CONFIG_NETFILTER_XTABLES: Netfilter Xtables support: module (pass)
CONFIG_NETFILTER_XT_TARGET_REDIRECT: REDIRECT target support: module (pass)
CONFIG_NETFILTER_XT_MATCH_COMMENT: "comment" match support: module (pass)
CONFIG_NETFILTER_XT_MARK: nfmark target and match support: module (pass)
CONFIG_NETFILTER_XT_SET: set target and match support: module (pass)
CONFIG_NETFILTER_XT_TARGET_MASQUERADE: MASQUERADE target support: module (pass)
CONFIG_NETFILTER_XT_NAT: "SNAT and DNAT" targets support: module (pass)
CONFIG_NETFILTER_XT_MATCH_ADDRTYPE: "addrtype" address type match support: module (pass)
CONFIG_NETFILTER_XT_MATCH_CONNTRACK: "conntrack" connection tracking match support: module (pass)
CONFIG_NETFILTER_XT_MATCH_MULTIPORT: "multiport" Multiple port match support: module (pass)
CONFIG_NETFILTER_XT_MATCH_RECENT: "recent" match support: module (pass)
CONFIG_NETFILTER_XT_MATCH_STATISTIC: "statistic" match support: module (pass)
CONFIG_NETFILTER_NETLINK: module (pass)
CONFIG_NF_NAT: module (pass)
CONFIG_IP_SET: IP set support: module (pass)
CONFIG_IP_SET_HASH_IP: hash:ip set support: module (pass)
CONFIG_IP_SET_HASH_NET: hash:net set support: module (pass)
CONFIG_IP_VS: IP virtual server support: module (pass)
CONFIG_IP_VS_NFCT: Netfilter connection tracking: built-in (pass)
CONFIG_IP_VS_SH: Source hashing scheduling: module (pass)
CONFIG_IP_VS_RR: Round-robin scheduling: module (pass)
CONFIG_IP_VS_WRR: Weighted round-robin scheduling: module (pass)
CONFIG_NF_CONNTRACK_IPV4: IPv4 connetion tracking support (required for NAT): unknown (warning)
CONFIG_NF_REJECT_IPV4: IPv4 packet rejection: module (pass)
CONFIG_NF_NAT_IPV4: IPv4 NAT: unknown (warning)
CONFIG_IP_NF_IPTABLES: IP tables support: module (pass)
CONFIG_IP_NF_FILTER: Packet filtering: module (pass)
CONFIG_IP_NF_TARGET_REJECT: REJECT target support: module (pass)
CONFIG_IP_NF_NAT: iptables NAT support: module (pass)
CONFIG_IP_NF_MANGLE: Packet mangling: module (pass)
CONFIG_NF_DEFRAG_IPV4: module (pass)
CONFIG_NF_CONNTRACK_IPV6: IPv6 connetion tracking support (required for NAT): unknown (warning)
CONFIG_NF_NAT_IPV6: IPv6 NAT: unknown (warning)
CONFIG_IP6_NF_IPTABLES: IP6 tables support: module (pass)
CONFIG_IP6_NF_FILTER: Packet filtering: module (pass)
CONFIG_IP6_NF_MANGLE: Packet mangling: module (pass)
CONFIG_IP6_NF_NAT: ip6tables NAT support: module (pass)
CONFIG_NF_DEFRAG_IPV6: module (pass)
CONFIG_BRIDGE: 802.1d Ethernet Bridging: module (pass)
CONFIG_LLC: module (pass)
CONFIG_STP: module (pass)
CONFIG_EXT4_FS: The Extended 4 (ext4) filesystem: module (pass)
CONFIG_PROC_FS: /proc file system support: built-in (pass)
What happened?
I tried to upgrade locally from v1.29.5+k0s.0 to v1.30.1+k0s.0 but it kept failing. I previously upgraded from v1.28.4+k0s.0 to v1.29.5+k0s.0 the same way flawlessly. I'm almost certain my problem is related to #4172, since creating a new cluster using v1.30.1+k0s.0 works. I would like to know how to make the upgrade works for my currently working cluster.
When trying to change the value for the timeout in the ClusterConfig to 10m, the change is refused since the value type accepted are integer and not string. However, the proper type to use for latest version is string instead of integer.
# * spec.extensions.helm.charts[0].timeout: Invalid value: "string": spec.extensions.helm.charts[0].timeout in body must be of type integer: "string"
Jun 05 21:14:39 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:39" level=info msg="I0605 21:14:39.019336 233273 handler.go:286] Adding GroupVersion metrics.k8s.io v1beta1 to ResourceManager" component=kube-apiserver stream=stderr
Jun 05 21:14:39 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:39" level=info msg="E0605 21:14:39.023352 233273 available_controller.go:460] v1beta1.metrics.k8s.io failed with: Operation cannot be fulfilled on apiservices.apiregistration.k8s.io \"v1beta1.metrics.k8s.io\": the object has been modified; please apply your changes to the latest version and try again" component=kube-apiserver stream=stderr
Jun 05 21:14:39 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:39" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:40 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:40" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:41 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:41" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:42 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:42" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:42 k-test1-ctl13 k0s[233185]: I0605 21:14:42.799008 233185 leaderelection.go:260] successfully acquired lease kube-node-lease/k0s-ctrl-k-test1-ctl13
Jun 05 21:14:42 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:42" level=info msg="Acquired leader lease" component=controllerlease
Jun 05 21:14:42 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:42" level=info msg="acquired leader lease" component=controllerlease
Jun 05 21:14:43 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:43" level=debug msg="Probing components" component=prober
Jun 05 21:14:43 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:43" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:44 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:44" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="counting controller lease holders" component=controllerlease
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="notifying subscribers (1) about controller count: 3" component=controllerlease
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="Sending stop message" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="Waiting for stopping is done" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="Shutting down pid 233272" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="Got event" component=Konnectivity event="{2024-06-05 21:14:45.275211949 +0000 UTC m=+13.336666017 restarting konnectivity server due to server count change map[serverCount:3]}"
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.275351 233272 server.go:167] Shutting down server." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="E0605 21:14:45.275437 233272 server.go:465] \"health server could not listen\" err=\"http: Server closed\"" component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.275454 233272 server.go:467] Health server stopped listening" component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="E0605 21:14:45.275482 233272 server.go:423] \"admin server could not listen\" err=\"http: Server closed\"" component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.275487 233272 server.go:425] Admin server stopped listening" component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="Starting to supervise" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="Started successfully, go nuts pid 233455" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="waiting for server count change" component=konnectivity
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="Got event" component=Konnectivity event="{2024-06-05 21:14:45.280654816 +0000 UTC m=+13.342108874 started konnectivity server map[serverCount:3]}"
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294889 233455 options.go:148] ServerCert set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294978 233455 options.go:149] ServerKey set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294981 233455 options.go:150] ServerCACert set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294984 233455 options.go:151] ClusterCert set to \"/var/lib/k0s/pki/server.crt\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294990 233455 options.go:152] ClusterKey set to \"/var/lib/k0s/pki/server.key\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294993 233455 options.go:153] ClusterCACert set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.294997 233455 options.go:154] Mode set to \"grpc\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295001 233455 options.go:155] UDSName set to \"/run/k0s/konnectivity-server/konnectivity-server.sock\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295005 233455 options.go:156] DeleteUDSFile set to true." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295009 233455 options.go:157] Server port set to 0." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295014 233455 options.go:158] Server bind address set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295018 233455 options.go:159] Agent port set to 8132." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295022 233455 options.go:160] Agent bind address set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295027 233455 options.go:161] Admin port set to 8133." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295032 233455 options.go:162] Admin bind address set to \"127.0.0.1\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295036 233455 options.go:163] Health port set to 8092." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295041 233455 options.go:164] Health bind address set to \"\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295046 233455 options.go:165] Keepalive time set to 1h0m0s." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295077 233455 options.go:166] Frontend keepalive time set to 1h0m0s." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295086 233455 options.go:167] EnableProfiling set to false." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295094 233455 options.go:168] EnableContentionProfiling set to false." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295103 233455 options.go:169] ServerID set to 35c14efde7c4c084741960546145d033." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295112 233455 options.go:170] ServerCount set to 3." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295122 233455 options.go:171] AgentNamespace set to \"kube-system\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295131 233455 options.go:172] AgentServiceAccount set to \"konnectivity-agent\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295145 233455 options.go:173] AuthenticationAudience set to \"system:konnectivity-server\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295155 233455 options.go:174] KubeconfigPath set to \"/var/lib/k0s/pki/konnectivity.conf\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295167 233455 options.go:175] KubeconfigQPS set to 0.000000." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295180 233455 options.go:176] KubeconfigBurst set to 0." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295193 233455 options.go:177] ProxyStrategies set to \"destHost,default\"." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.295207 233455 options.go:178] CipherSuites set to [\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256\"]." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.296735 233455 server.go:130] Starting frontend server for client connections." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.297019 233455 server.go:145] Starting agent server for tunnel connections." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.297302 233455 server.go:152] Starting admin server for debug connections." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=info msg="I0605 21:14:45.297389 233455 server.go:159] Starting health server for healthchecks." component=konnectivity stream=stderr
Jun 05 21:14:45 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:45" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:46 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:46" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:47 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:47" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:47 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:47" level=debug msg="not the leader, can't approve certificates" component=csrapprover
Jun 05 21:14:48 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:48" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:49 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:49" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:50 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:50" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:51 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:51" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:52 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:52" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:53 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:53" level=debug msg="Probing components" component=prober
Jun 05 21:14:53 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:53" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:54 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:54" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:55 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:55" level=debug msg="counting controller lease holders" component=controllerlease
Jun 05 21:14:55 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:55" level=debug msg="notifying subscribers (1) about controller count: 3" component=controllerlease
Jun 05 21:14:55 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:55" level=debug msg="waiting for server count change" component=konnectivity
Jun 05 21:14:55 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:55" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:56 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:56" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:57 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:57" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:57 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:57" level=debug msg="not the leader, can't approve certificates" component=csrapprover
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Failed to ensure the existence of the cluster configuration" component=clusterConfig-reconciler error="json: cannot unmarshal number into Go struct field Chart.spec.extensions.helm.charts.timeout of type string"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component CRD"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component CRD"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component EtcdMemberReconciler"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component Status"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component CSRApprover"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Sending stop message" component=k0s-control-api
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Waiting for stopping is done" component=k0s-control-api
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="CSR Approver context done" component=csrapprover
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="Shutting down pid 233415" component=k0s-control-api
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component K0SControlAPI"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component Manager"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="stopped component LeasePool"
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Sending stop message" component=kube-apiserver
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=debug msg="Waiting for stopping is done" component=kube-apiserver
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="Shutting down pid 233273" component=kube-apiserver
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="Lost leader lease" component=poolleaderelector
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="lost leader lease" component=poolleaderelector
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595594 233273 controller.go:128] Shutting down kubernetes service endpoint reconciler" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595628 233273 genericapiserver.go:670] \"[graceful-termination] pre-shutdown hooks completed\" name=\"PreShutdownHooksStopped\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595597 233273 genericapiserver.go:527] \"[graceful-termination] shutdown event\" name=\"ShutdownInitiated\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595640 233273 genericapiserver.go:530] \"[graceful-termination] shutdown event\" name=\"AfterShutdownDelayDuration\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595675 233273 genericapiserver.go:594] \"[graceful-termination] shutdown event\" name=\"NotAcceptingNewRequest\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595681 233273 genericapiserver.go:628] \"[graceful-termination] not going to wait for active watch request(s) to drain\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595692 233273 genericapiserver.go:619] [graceful-termination] in-flight non long-running request(s) have drained" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595725 233273 genericapiserver.go:660] \"[graceful-termination] shutdown event\" name=\"InFlightRequestsDrained\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595740 233273 dynamic_cafile_content.go:171] \"Shutting down controller\" name=\"request-header::/var/lib/k0s/pki/front-proxy-ca.crt\"" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595736 233273 controller.go:157] Shutting down quota evaluator" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595759 233273 controller.go:176] quota evaluator worker shutdown" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595789 233273 controller.go:176] quota evaluator worker shutdown" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595801 233273 controller.go:176] quota evaluator worker shutdown" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595808 233273 controller.go:176] quota evaluator worker shutdown" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595815 233273 controller.go:176] quota evaluator worker shutdown" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595824 233273 system_namespaces_controller.go:77] Shutting down system namespaces controller" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595847 233273 controller.go:84] Shutting down OpenAPI AggregationController" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595848 233273 establishing_controller.go:87] Shutting down EstablishingController" component=kube-apiserver stream=stderr
Jun 05 21:14:58 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:14:58" level=info msg="I0605 21:14:58.595870 233273 dynamic_serving_content.go:146] \"Shutting down controller\" name=\"aggregator-proxy-cert::/var/lib/k0s/pki/front-proxy-client.crt::/var/lib/k0s/pki/front-proxy-client.key\"" component=kube-apiserver stream=stderr
...
Jun 05 21:15:04 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:15:04" level=info msg="stopped component Etcd"
Jun 05 21:15:04 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:15:04" level=info msg="All node components stopped"
Jun 05 21:15:04 k-test1-ctl13 k0s[233185]: time="2024-06-05 21:15:04" level=info msg="stopping konnectivity server reconfig loop"
Jun 05 21:15:04 k-test1-ctl13 k0s[233185]: Error: failed to start cluster components: failed to ensure the existence of the cluster configuration: timed out waiting for the condition
Jun 05 21:15:04 k-test1-ctl13 systemd[1]: k0scontroller.service: Main process exited, code=exited, status=1/FAILURE
Jun 05 21:15:04 k-test1-ctl13 systemd[1]: k0scontroller.service: Failed with result 'exit-code'.
Jun 05 21:15:04 k-test1-ctl13 systemd[1]: k0scontroller.service: Consumed 34.079s CPU time.
Additional context
The challenge comes from the fact that I can't seem to be able to change the timeout value to a correct one for the new version and the upgrade fails without a correct value. Also, I noticed that the default value mentionned in the docs does not seem to apply since mine is set to 0 by default. https://docs.k0sproject.io/v1.30.1+k0s.0/helm-charts/#chart-configuration
The text was updated successfully, but these errors were encountered:
Before creating an issue, make sure you've checked the following:
Platform
Version
v1.29.5+k0s.0
Sysinfo
`k0s sysinfo`
What happened?
I tried to upgrade locally from
v1.29.5+k0s.0
tov1.30.1+k0s.0
but it kept failing. I previously upgraded fromv1.28.4+k0s.0
tov1.29.5+k0s.0
the same way flawlessly. I'm almost certain my problem is related to #4172, since creating a new cluster usingv1.30.1+k0s.0
works. I would like to know how to make the upgrade works for my currently working cluster.When trying to change the value for the timeout in the ClusterConfig to
10m
, the change is refused since the value type accepted are integer and not string. However, the proper type to use for latest version is string instead of integer.Steps to reproduce
Expected behavior
A working upgraded node.
Actual behavior
The k0s systemd service fails to run.
Screenshots and logs
k0sctl config
ClusterConfig
Journalctl logs
Additional context
The challenge comes from the fact that I can't seem to be able to change the timeout value to a correct one for the new version and the upgrade fails without a correct value. Also, I noticed that the default value mentionned in the docs does not seem to apply since mine is set to 0 by default.
https://docs.k0sproject.io/v1.30.1+k0s.0/helm-charts/#chart-configuration
The text was updated successfully, but these errors were encountered: