profile
viewpoint
If you are wondering where the data of this site comes from, please visit https://api.github.com/users/knabben/events. GitMemory does not store any data, but only uses NGINX to cache data for a period of time. The idea behind GitMemory is simply to give users a better reading experience.
Amim Knabben knabben @loadsmart New York City, NY https://thefind.live

knabben/chrn 8

Automatic GH Release Notes

knabben/k8s-labs 8

Kubernetes components and CNCF projects notes

knabben/forwarder 3

Kubernetes Automatic Port Forwarder

knabben/ggql 2

GraphQL API Depth Explorer

knabben/gql-pull 2

GraphQL pull and dump action

knabben/bov-stream 1

Realtime BVSP stock price analysis

knabben/chatops 1

A cloud native chatbot

knabben/daml-sc 1

DAML International Trading POC

knabben/kaboom 1

Traffic replicator for load test cases

knabben/argo 0

Argo Workflows: Get stuff done with Kubernetes.

issue commentkubernetes/enhancements

Provide RunAsGroup feature for Containers in a Pod

@krmayankk i believe 3/16 is the deadline to have a placeholder doc and it should also include such documentation. Let me check with the team and confirm.

krmayankk

comment created time in a few seconds

issue commentkubernetes/kubernetes

Sometime Liveness/Readiness Probes fail because of net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)

same issue on EKS 1.16

I noticed there is a potential correlation with the node network I/O. I was doing some heavy query in Thanos, and it caused a network spike, and at the same time some of the pods running on the same node as thanos pods got restarted due to liveness probe timeout. It happened three times so far which made me think it's not just a coincidence.

yuripastushenko

comment created time in a few seconds

pull request commentkubernetes/kubernetes

make unschedulableQTimeInterval configurable

@Jerry-Ge: The following tests failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
pull-kubernetes-bazel-test b74aa50fe2269d9620df785c79f7f8aeecae9c6a link /test pull-kubernetes-bazel-test
pull-kubernetes-e2e-kind 7838047b9e959f8d080e5b804e1b6890309e2c6c link /test pull-kubernetes-e2e-kind

Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR.

<details>

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. </details> <!-- test report -->

Jerry-Ge

comment created time in a few seconds

issue commentkubernetes/kubernetes

Kubelet tries to remove pod multiple times

https://github.com/containerd/containerd/blob/a72fe7da21237815731386d6b73a0e93700112f9/pkg/cri/server/instrumented_service.go#L114-L128 It doesn’t remove the pod many times, it just repeats the log

thyn

comment created time in 2 minutes

issue commentkubernetes/kubernetes

[Flaky Test][sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] should conform to Ingress spec

I agree we're relying to much on heroics of individual people that isn't sustainable. I think those are valid questions, but I would first like to hear from network folks with Google too.

sayanchowdhury

comment created time in 3 minutes

issue commentkubernetes/kubernetes

Prevent regressions after structured logging migration

I am not sure how to approach this, shouldn't we need valid key list for it?

As klog.InfoS function prototype is func(msg string, keysAndValues []interface{}) we can pass anything we like in keysAndValues and compile will not catch this. This led to misuse like this:

klog.V(2).InfoS("Keeping PVC", "PVC", klog.KRef(pvcNamespace, pvcName), " because it is still being used")

We would like to ensure that

  • Number of arguments is uneven (msg +X * (key+value) = 1 + 2X = always uneven number)
  • Argument on uneven index is always string (we should never pass a var or const as key)
ehashman

comment created time in 4 minutes

issue commentkubernetes/kubernetes

Kubelet tries to remove pod multiple times

https://github.com/CaoDonghui123/githubData/blob/6525b35d3012d0c47bdd20ee1707b7f20a035298/main.go#L27-L53 It doesn’t remove the pod many times, it just repeats the log

thyn

comment created time in 5 minutes

pull request commentkubernetes/kubernetes

Migrating pkg/kubelet/winstats to structured logging

/retest

hexxdump

comment created time in 5 minutes

pull request commentkubernetes/kubernetes

Update csiNodeIDMaxLength to 192 bytes

/assign @liggitt

For API approval

Jiawei0227

comment created time in 6 minutes

pull request commentkubernetes/kubernetes

Update csiNodeIDMaxLength to 192 bytes

/hold cancel

CSI 1.4.0 has been cut https://github.com/container-storage-interface/spec/releases/tag/v1.4.0

Jiawei0227

comment created time in 8 minutes

PR opened pandas-dev/pandas

[ArrayManager] TST: enable IO tests Internals Testing

xref https://github.com/pandas-dev/pandas/issues/39146

This basically just adds IO testing to the set of tests we run for ArrayManager. I already added skips for JSON and PyTables IO in the past, so those are still skipped now as well.

+1 -5

0 comment

2 changed files

pr created time in 9 minutes

Pull request review commentscikit-learn/scikit-learn

Fix Calibrated classifier cv predictions with pipeline

 def predict_proba(self, X):             The predicted probas.         """         check_is_fitted(self)-        X = check_array(X, accept_sparse=['csc', 'csr', 'coo'],-                        force_all_finite=False)+         # Compute the arithmetic mean of the predictions of the calibrated         # classifiers-        mean_proba = np.zeros((X.shape[0], len(self.classes_)))+        mean_proba = np.zeros((_num_samples(X), len(self.classes_)))

Maybe we could describe X as

The samples, as accepted by base_estimator.predict_proba (same below)

AlekLefebvre

comment created time in 16 minutes

Pull request review commentscikit-learn/scikit-learn

Fix Calibrated classifier cv predictions with pipeline

 def test_calibration_pipeline(text_data, text_data_pipeline):         calib_clf.n_features_in_  +def test_calibration_pipeline_prediction(text_data, text_data_pipeline):

let's just merge that with the test above? (adding the new comments etc)

AlekLefebvre

comment created time in 18 minutes

Pull request review commentscikit-learn/scikit-learn

Fix Calibrated classifier cv predictions with pipeline

 Changelog   for non-English characters. :pr:`18959` by :user:`Zero <Zeroto521>`   and :user:`wstates <wstates>`. +:mod:`sklearn.calibration`+............................++- |Fix| :class:`calibration.CalibratedClassifierCV` now delegating validation+  to base estimator to fix :func:`CalibratedClassifierCV.predict` and+  :func:`CalibratedClassifierCV.predict_proba`

I think it's more important to describe what bug the fix fixes, rather than how it fixes the bug. So I'd suggest something like:

The predict and predict_proba methods of :class:`calibration.CalibratedClassifierCV can now properly be used on prefitted pipelines.

AlekLefebvre

comment created time in 9 minutes

pull request commentkubernetes/kubernetes

reduce configmap and secret watch of kubelet

[APPROVALNOTIFIER] This PR is NOT APPROVED

This pull-request has been approved by: <a href="https://github.com/kubernetes/kubernetes/pull/99393#" title="Author self-approved">chenyw1990</a> To complete the pull request process, please assign wojtek-t after the PR has been reviewed. You can assign the PR to them by writing /assign @wojtek-t in a comment when ready.

The full list of commands accepted by this bot can be found here.

<details open> Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment Approvers can cancel approval by writing /approve cancel in a comment </details> <!-- META={"approvers":["wojtek-t"]} -->

chenyw1990

comment created time in 9 minutes

issue commentkubernetes/kubernetes

[Flaky Test][sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] should conform to Ingress spec

[To be clear - I'm not trying to push sig-network to own that, i'm just trying to understand the arguments better.]

Let me share my thoughts, I'm personally trying to make all the dashboards in sig-network green, and this week I was able to sort out most of the problems, the last flake with TCP CLOSE_WAIT seems solved, and everything is acceptable except in GCE and Azure jobs. https://testgrid.k8s.io/sig-network

The dual-stack azure jobs are pending of a new agnhost image version bump, but I know there is an Azure person that is taking responsibility and will be working on fixing it: https://testgrid.k8s.io/sig-network-dualstack-azure-e2e

For GCE jobs I use to ping @prameshj or @MrHohn, and you can see that the gce dashboard is in a good state https://testgrid.k8s.io/sig-network-gce

However, there are GCE jobs that are in very bad state, failing continuously, https://testgrid.k8s.io/sig-network-ingress-gce-e2e https://testgrid.k8s.io/sig-network-ingress-gce-e2e#ingress-gce-e2e

I'm practically doing this by myself, that is a personal decision, spending some time every day to review the dashboards and addressing the "generic" failures, but my time is limited, so if this has to go under sig-network umbrella I just want to be sure that these are really needed and, if they are GCE only features, asking for some help from GCE folks.

sayanchowdhury

comment created time in 11 minutes

Pull request review commentkubernetes/kubernetes

Migrate pkg/kubelet/kuberuntime/kuberuntime_manager.go to structured logging

 func (m *kubeGenericRuntimeManager) SyncPod(pod *v1.Pod, podStatus *kubecontaine 		isInBackOff, msg, err := m.doBackOff(pod, spec.container, podStatus, backOff) 		if isInBackOff { 			startContainerResult.Fail(err, msg)-			klog.V(4).Infof("Backing Off restarting %v %+v in pod %v", typeName, spec.container, format.Pod(pod))+			klog.V(4).InfoS("Backing Off restarting container in pod", "containerType", typeName, "container", spec.container, "pod", klog.KObj(pod))

Ok, looks like original log passed %+v with intention to dump whole container object. Passing it here also make sense.

adisky

comment created time in 13 minutes

pull request commentkubernetes/kubernetes

Promote PodDisruptionBudget to policy/v1

@liggitt I have addressed/responded to the comments, and added a separate commit to move all e2e tests to use the v1 api. I have created a PR to promote e2e tests to Conformance: https://github.com/kubernetes/kubernetes/pull/84740

mortent

comment created time in 16 minutes

pull request commentkubernetes/kubernetes

Migrating pkg/kubelet/winstats to structured logging

@hexxdump: The following tests failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
pull-kubernetes-unit 2424c7388dc0a3e960bbf0ae5fabd31e9e262829 link /test pull-kubernetes-unit
pull-kubernetes-e2e-kind 2424c7388dc0a3e960bbf0ae5fabd31e9e262829 link /test pull-kubernetes-e2e-kind

Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR.

<details>

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. </details> <!-- test report -->

hexxdump

comment created time in 21 minutes

issue commentkubernetes-sigs/cri-tools

FATA[0000] getting the runtime version: rpc error: code = Internal desc = stream terminated by RST_STREAM with error code: PROTOCOL_ERROR

Hey @Ruilkyu, may I ask you to share more information about this issue? To which server (container runtime) are you trying to connect? Which crictl version are you using?

Ruilkyu

comment created time in 24 minutes

push eventkubernetes/kubernetes

Jordan Liggitt

commit sha ba4aade23a76969d7598fb7ac170891c183e7327

Deflake TestPrepareRun

view details

Kubernetes Prow Robot

commit sha 348fc2e41463a1e8087b054315befe57b7b5146c

Merge pull request #99908 from liggitt/prepare-run-flake Deflake TestPrepareRun

view details

push time in 24 minutes

PR merged kubernetes/kubernetes

Reviewers
Deflake TestPrepareRun approved area/apiserver area/deflake cncf-cla: yes kind/flake lgtm needs-priority needs-triage release-note-none sig/api-machinery size/S

What type of PR is this?

the max-in-flight filter takes a few milliseconds to initialize and raced the /healthz get, producing this flake:

=== RUN   TestPrepareRun
W0306 19:41:03.993313   72776 authorization.go:47] Authorization is disabled
W0306 19:41:03.994050   72776 authentication.go:40] Authentication is disabled
    genericapiserver_test.go:336: [+]ping ok
        [+]log ok
        [-]poststarthook/max-in-flight-filter failed: reason withheld
        healthz check failed
        
    genericapiserver_test.go:337: 
        	Error Trace:	genericapiserver_test.go:337
        	Error:      	Not equal: 
        	            	expected: 200
        	            	actual  : 500
        	Test:       	TestPrepareRun

https://storage.googleapis.com/k8s-gubernator/triage/index.html?pr=1&job=pull-kubernetes-unit&test=PrepareRun

/kind flake

NONE
+15 -4

3 comments

1 changed file

liggitt

pr closed time in 24 minutes

pull request commentkubernetes/kubernetes

Support specifying more than 5 ports in L4 ILB service

[APPROVALNOTIFIER] This PR is APPROVED

This pull-request has been approved by: <a href="https://github.com/kubernetes/kubernetes/pull/99595#issuecomment-792568581" title="Approved">cheftako</a>, <a href="https://github.com/kubernetes/kubernetes/pull/99595#pullrequestreview-601008351" title="Approved">MrHohn</a>, <a href="https://github.com/kubernetes/kubernetes/pull/99595#" title="Author self-approved">prameshj</a>

The full list of commands accepted by this bot can be found here.

The pull request process is described here

<details > Needs approval from an approver in each of these files:

Approvers can indicate their approval by writing /approve in a comment Approvers can cancel approval by writing /approve cancel in a comment </details> <!-- META={"approvers":[]} -->

prameshj

comment created time in 24 minutes

issue commentscikit-learn/scikit-learn

OOB calculation in Random Forests seems to use incorrect indices

Yup, I think you might be right, thanks for the report.

Ideally we would generate the oob samples at the same place where we're generating the bagging samples (i.e. in _parallel_build_trees) to avoid reproducing the bug in the future. But that might take up a lot of memory... I guess we'll have to sample n_estimators int seed for that.

Would you like to try and provide a fix @simontindemans ?

simontindemans

comment created time in 25 minutes

pull request commentkubernetes/kubernetes

Support specifying more than 5 ports in L4 ILB service

/lgtm /approve

prameshj

comment created time in 25 minutes

pull request commentkubernetes/kubernetes

podresources APIs: concrete resources apis: implement GetAllocatableResources

@fromanirh: The following tests failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
pull-kubernetes-e2e-gce-device-plugin-gpu 701b859a4002547c0f2cc566b403013b528c76d6 link /test pull-kubernetes-e2e-gce-device-plugin-gpu
pull-kubernetes-e2e-azure-file-windows-containerd 6239c1acb8c9b45bbd9404018d8203a8ec244e4f link /test pull-kubernetes-e2e-azure-file-windows-containerd

Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR.

<details>

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. </details> <!-- test report -->

fromanirh

comment created time in 26 minutes

pull request commentkubernetes/kubernetes

TestValidateNetworkPolicy refactor

/retest

danielamlins

comment created time in 30 minutes

pull request commentkubernetes/kubernetes

Migrating pkg/kubelet/winstats to structured logging

@hexxdump: The following tests failed, say /retest to rerun all failed tests:

Test name Commit Details Rerun command
pull-kubernetes-e2e-kind 68edcb2dca8507ff1d146e56c0c75fe948c4c05d link /test pull-kubernetes-e2e-kind
pull-kubernetes-unit 2424c7388dc0a3e960bbf0ae5fabd31e9e262829 link /test pull-kubernetes-unit

Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR.

<details>

Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. </details> <!-- test report -->

hexxdump

comment created time in 34 minutes

issue commentkubernetes/kubernetes

[Flaky Test][sig-network] Loadbalancing: L7 GCE [Slow] [Feature:Ingress] should conform to Ingress spec

I think that everything should be accessible - @freehan @bowei to confirm.

is the test testing kubernetes network or gce cloud-provider, gce ingress, ....?

Technically, many networking (and other) tests are relying on cloud-provider (e.g. all tests that exercise LoadBalanced service). So that isn't unique (do we have complaints about these tests too)? So what make ingress tests different is a separate out-of-tree controller, I guess.

[To be clear - I'm not trying to push sig-network to own that, i'm just trying to understand the arguments better.]

sayanchowdhury

comment created time in 36 minutes

issue commentkubernetes/kubernetes

Improve concurrency for hack/verify-structured-logging.sh

/assign

adisky

comment created time in 37 minutes