profile
viewpoint

DP19/aws-node-termination-handler 0

A Kubernetes Daemonset to gracefully handle EC2 instance shutdown

DP19/dashboard 0

General-purpose web UI for Kubernetes clusters

DP19/helm 0

The Kubernetes Package Manager

DP19/homebrew-php 0

💀 Homebrew/php (deprecated)

DP19/munki-enroll 0

Scripts to automatically enroll clients in Munki, allowing for very flexible manifest structures.

DP19/simple-node-js-react-npm-app 0

For an introductory tutorial on how to use Jenkins to build a simple Node.js and React application with npm.

DP19/teslamate 0

A self-hosted data logger for your Tesla 🚘

DP19/webapp 0

Sample Web App with Maven for Jenkins Demo

DP19/zabbix-cachet 0

Python script which sync Zabbix IT Services with Cachet

push eventDP19/helm

Simon Alling

commit sha 36c9c0e5202789f0f2c87fabf816900259de0fb8

Simplify chart installable check The error conveys at least as much information as the boolean. Signed-off-by: Simon Alling <alling.simon@gmail.com>

view details

Nguyen Hai Truong

commit sha 451e2158ad24594d8c67e2e4a99aefc5ffa84a9f

Remove duplicated words (#7336) Although it is spelling mistakes, it might make an affects while reading. Signed-off-by: Nguyen Hai Truong <truongnh@fujitsu.com> Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

Jorge I. Gasca

commit sha b7ff1e29327d33b6fe9b44b11d84c182d1adea45

fix(cmd): Fixes logging on action conf init error (#6909) * fix(cmd): Fixes logging on action conf init error Errors relating to initializing the action config cause helm to exit silently, except when in debug mode. This now emits the useful error. Closes #6863 Signed-off-by: Jorge Gasca <jorge.ignacio.gasca@gmail.com> * Remove unnecessary formatting of err struct Signed-off-by: Jorge Gasca <jorge.ignacio.gasca@gmail.com> Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

Hu Shuai

commit sha 00769c4512c9c145eafb84f60e8a473de83b0209

Fix a typo "update" -> "updates" (#7346) Signed-off-by: Hu Shuai <hus.fnst@cn.fujitsu.com> Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

EItanya

commit sha 17dc43f0547d4fd7205de07428b97fbc4a5276fd

added config file string Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

EItanya

commit sha 2f534f97424ce4eaa3356e199e453e94bd65b4c6

fix test Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

EItanya

commit sha 640d527190376a8b4945575f767065ebd47a83cb

removed panic, and replaced with error Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

EItanya

commit sha f604105547dc260f4388792de44a6eb9d2a1353f

fixed to mirror master Signed-off-by: EItanya <eitan.yarmush@solo.io>

view details

Andreas Lindhé

commit sha de1996e500c75faccbb8aadc1d2e8b788f83ac3b

Add comments about release Version variable I was looking into the `get` command, and got tripped up by the `Version` variable. It was unclear to me what Version represents, since it's called REVISION when doing e.g., `helm list`. But even after knowing this, it was not very clear to me why we (implicitly) set the Version variable to 0 but never seem to use it. `mhickey` explained to me on Slack that this gets the latest revision of the release. Makes sense, but I added a comment about that too, to clarify. Signed-off-by: Andreas Lindhé <andreas@lindhe.io>

view details

EItanya

commit sha 872d9bcb0e0058a261fc96ec168f0b34b2809fa6

Merge branch 'master' of github.com:helm/helm into helm-7351

view details

liuming216448

commit sha cca68288063d235a4c32ef1ba822dd487317c8dc

fix: allow to rollback to previous version even if no deployed releases(#6978) Signed-off-by: liuming <hit_oak_tree@126.com>

view details

liuming

commit sha e7adc4b5d73ec76a84a5e8ae258823b3d2990a74

Merge remote-tracking branch 'helm/master'

view details

liuming

commit sha 299ccd9e8878e6194d9d2a19affd4fbe0ee65f38

Merge remote-tracking branch 'helm/master'

view details

Liu Ming

commit sha fe308142956d1641d614a7414f5d9b6717f2a1e7

Merge remote-tracking branch 'helm/master'

view details

Liu Ming

commit sha 98ec2760c800ddcd75d27c75318cd979fb1c6a39

Merge remote-tracking branch 'helm/master'

view details

Liu Ming

commit sha bd08203a0fc742d3abebcc333f17968239be5cbb

Merge remote-tracking branch 'helm/master'

view details

Adam Reese

commit sha bdf6f48704ed9e09d7fa636f025a3e2d344d42d4

fix(pkg/kube): continue deleting objects when one fails * Continue deleting objects when one fails to minimize the risk of an upgrade ending in an unrecoverable state * Exclude failed deleted object from the returned result set Signed-off-by: Adam Reese <adam@reese.io>

view details

Liu Ming

commit sha 376a5ca55074ffa339faa40cdecca246404056ea

Merge remote-tracking branch 'helm/master'

view details

Liu Ming

commit sha a1685b737dd6846bdc35c281bf841b9cc43cb104

Merge remote-tracking branch 'helm/master' Signed-off-by: Liu Ming <hit_oak_tree@126.com>

view details

Liu Ming

commit sha b99d493a9e1bce54d4ff61d90d67a9221906bbfc

Merge remote-tracking branch 'helm/master'

view details

push time in a month

Pull request review commentaws/aws-node-termination-handler

Add retries when reponse from IMDSv2 retruns a 401

 func main() { 	for _, fn := range monitoringFns { 		go func(monitor monitor.Monitor) { 			log.Log().Msgf("Started monitoring for %s events", monitor.Kind())+			var previousErr error+			var duplicateErrCount int 			for range time.Tick(time.Second * 2) { 				err := monitor.Monitor() 				if err != nil { 					log.Log().Msgf("There was a problem monitoring for %s events: %v", monitor.Kind(), err) 					metrics.ErrorEventsInc(monitor.Kind())+					if err == previousErr {+						duplicateErrCount+++					} else {+						duplicateErrCount = 0+						previousErr = err+					}+					if duplicateErrCount >= duplicateErrThreshold {+						log.Log().Msg("Stopping NITH - Duplicate Error Threshold hit.")

Updated

DP19

comment created time in a month

PullRequestReviewEvent

push eventDP19/aws-node-termination-handler

David Pait

commit sha f459e124aedef90fae17d5b0c69f655b00b11238

fix report card

view details

push time in a month

push eventDP19/aws-node-termination-handler

David Pait

commit sha 99f8ee43e32da574879892b7f551f03cbce6b55e

fix spelling of NTH

view details

push time in a month

push eventDP19/aws-node-termination-handler

David Pait

commit sha 7399d7951b265351566638b83cb3e508786fb3fb

add missing equals to duplicateErrorThershold check in main function

view details

push time in a month

pull request commentaws/aws-node-termination-handler

Add retries when reponse from IMDSv2 retruns a 401

@bwagner5 - thanks for the feedback! I've move this logic to the Request function and add two new local vars to the monitor loop to track the previous error and if you get the same error 3 times in a row it should panic. This way it should cover more than just the 401 error but any error that's duplicated

DP19

comment created time in a month

push eventDP19/aws-node-termination-handler

David Pait

commit sha e48b20326dd41a164eec2a107b279c7670afb75a

Revert "Add retries when reponse from IMDSv2 retruns a 401" This reverts commit c1e34774098384db6b33c0774ec7dcf7a5c9bce7.

view details

David Pait

commit sha b0869444b859dbc984541fa9712a11912518a198

move IMDSv2 401 retries to Request function. move Panic to main function. add tests for 401 retries.

view details

push time in a month

issue commentaws/aws-node-termination-handler

Unable to parse metadata response: Metadata request received http status code: 401

@bwagner5 - Just opened a PR to put some retries in for getting info for both scheduled maintenance events and spot instance events. Could use some guidance on how best to test this!

benzch

comment created time in a month

PR opened aws/aws-node-termination-handler

Add retries when reponse from IMDSv2 retruns a 401

Issue #, if available: #229 Description of changes: Added a for loop for getting scheduled maintenance events and spot instance events. Set the retry limit to 1.

Not sure of the best way to add this to the test suite. Also curious if throwing a panic to sovle the second item in the issue is what's best to do here or not. By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

+45 -28

0 comment

1 changed file

pr created time in a month

push eventDP19/aws-node-termination-handler

David Pait

commit sha c1e34774098384db6b33c0774ec7dcf7a5c9bce7

Add retries when reponse from IMDSv2 retruns a 401

view details

push time in a month

fork DP19/aws-node-termination-handler

A Kubernetes Daemonset to gracefully handle EC2 instance shutdown

https://aws.amazon.com/ec2

fork in a month

push eventDP19/aws-node-termination-handler

David Pait

commit sha 8557cf7ec0311d1f975e4355295ebe7a0773aae8

Update ec2metadata package to include AvailabilityZone info (#210)

view details

Bryan™

commit sha 1aca361687645c7d4b31f4a1e8b952d1f09541ab

Integrate AEMM with NTH integration tests (#209) * use aemm in e2e tests * renaming ec2-metadata-test-proxy to webhook-test-proxy * reorder pod installation in e2e and renamings * intregrate aemm with eks cluster test * addressing shellcheck, fmt issues * update AEMM version and address pr comments

view details

crhuber

commit sha 28f6f52be979f0673fcee28784e0464a7658724e

Add PodMonitor for Prometheus Metrics (#206) * Add PodMonitor * remove quotes on ports * add readme

view details

Bryan™

commit sha 6f9d2b056566b4b906026367e74f0fdb17923c45

bump chart.yaml version (#211)

view details

Jerad C

commit sha d0610b4ced66c0fe431a03d2e51a9d77f924841c

Fix log output on e2e tests when a failure occurs (#163) (#212) * fix NTH log output on failure in local suite * add NTH log output on failure to EKS suite * remove trailing whitespace

view details

Brandon Wagner

commit sha af1722728d918e25100a9d3786afd121f461cf27

cleanup webhook secret after each test run (#213)

view details

Brandon Wagner

commit sha 8bf43a9fd02c5531db405fbaef62eb9678460881

webhook-http-proxy-test: print squid logs on failure and improve access log grep (#214)

view details

Brandon Wagner

commit sha 4a740f7b75aa50a429d12fe092f501c59238b3ef

squid fixes for webhook-http-proxy-test (#216) * squid fixes for webhook-http-proxy-test * moved squid pod lookup to a function

view details

Bryan™

commit sha b52d4675a716bb4150033ad3e8f0488def39a825

AEMM & Windows integrations (#215) * update aemm in e2e tests to support targetNodeOs; bumping aemm version * updating scripts+travis to support Windows deployments * bump AEMM version * update Windows version * fix push-docker-images logic to populate manifest correctly * addressing pr comments

view details

Bryan™

commit sha e68df3460c92a4db705c37319108989585b07e62

add readme for e2e tests (#218)

view details

Brandon Wagner

commit sha bad975d59e330b8af966b087f23cbf8c5692ecbb

only deploy to worker node on e2e tests (#217) * only deploy to worker node on e2e tests * fix webhook tests without using host networking * fix webhook port consistency * print gomod download logs

view details

Brandon Wagner

commit sha 91b481abceb22816019b3953ec970ae16f31877a

fix webhook test stability (#219)

view details

Jerad C

commit sha a91b0e788596ceb43b8a1893573ae9c899250fdb

add missing helm chart properties (#220) (#222)

view details

Thomas O'Neill

commit sha 95fdc025df8f92b725f988869854917d961cc14b

Fix Helm chart comment on enableSpotInterruptionDraining default behavior (#221) * default enableSpotInterruptionDraining to true * enableSpotInterruptionDraining default to "true" * Update README for helm chart, and roll back value change Co-authored-by: Thomas O'Neill <toneill@new-innov.com>

view details

Prathibha Datta Kumar

commit sha 1943f8bce4f9a1ef73a52c8973712aac1e1c3462

fix missing new line for updateStrategy (#227)

view details

Prathibha Datta Kumar

commit sha 292e4af3844abae869fb95bcd8f37255dfbf5286

updating default updateStrategy and bumping version for helm chart release (#228)

view details

Bryan™

commit sha ca45b8f8274c25be8c5180e51efdea4b52e7e847

fix manifest updating in push-docker-images (#231)

view details

Brandon Wagner

commit sha 3e355f41607f887a3ebc7e39d38137cfb193c6eb

Add Amazon EC2 Spot Instances Integration Roadmap to readme (#232)

view details

Brandon Wagner

commit sha e48c1f8b965f7e608688fb59a555f87203bd1326

upgrade to go 1.15 (#233) * upgrade to go 1.15 * remove duplicate linker flags * remove -s linker flag from windows

view details

Paulo Martins

commit sha c685bbaf7a614305bed5e33f4aed867e15a041b3

fix identation on PodMonitor (#235)

view details

push time in a month

PR merged DP19/aws-node-termination-handler

update

Issue #, if available:

Description of changes:

By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

+1897 -1017

0 comment

75 changed files

DP19

pr closed time in a month

PR opened DP19/aws-node-termination-handler

update

Issue #, if available:

Description of changes:

By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

+1897 -1017

0 comment

75 changed files

pr created time in a month

issue closedkubernetes/dashboard

Failed Job has green checkmark

Environment

<!-- Describe your setup. Versions of Node.js, Go etc. are needed only from developers. -->

Installation method: Helm
Kubernetes version: v1.16.8-eks-fd1ea7
Dashboard version: v2.0.3
Operating system: Amazon Linux
Steps to reproduce
  1. Launch Job that is designed to fail by having a pod terminate with error
Observed result

The Job in the UI will show a green checkmark but the pod will show it's terminated with Error and kubectl will report the job as failed

Expected result

In v1 it showed the job with a red exclamation point indicating the job as failed, I would expect v2 to do the same

closed time in 2 months

DP19

fork DP19/dashboard

General-purpose web UI for Kubernetes clusters

fork in 2 months

issue openedkubernetes/dashboard

Failed Job has green checkmark

Environment

<!-- Describe your setup. Versions of Node.js, Go etc. are needed only from developers. -->

Installation method: Helm
Kubernetes version: v1.16.8-eks-fd1ea7
Dashboard version: v2.0.3
Operating system: Amazon Linux
Steps to reproduce
  1. Launch Job that is designed to fail by having a pod terminate with error
Observed result

The Job in the UI will show a green checkmark but the pod will show it's terminated with Error and kubectl will report the job as failed

Expected result

In v1 it showed the job with a red exclamation point indicating the job as failed, I would expect v2 to do the same

created time in 2 months

startedauth0/java-jwt

started time in 2 months

startedjwtk/jjwt

started time in 2 months

startedtelepresenceio/telepresence

started time in 2 months

issue commentaws/aws-node-termination-handler

Provide handling for ASG 'rebalance' events

Another option is to enabled AZRebalance in the Suspend Processes section of the ASG. This will prevent the ASG from automatically trying to balance the nodes across AZ's. While this doesn't really solve the issue of the termination handling, it will stop it from happening in the first place

spolspol

comment created time in 2 months

pull request commentaws/aws-node-termination-handler

Update ec2metadata package to include AvailabilityZone info

@bwagner5 - Thanks! I did update the e2e tests to include this new field and saw that it ended up being an empty string

2020/07/21 14:48:07 Startup Metadata Retrieved: {InstanceID:i-1234567890abcdef0 InstanceType:m4.large PublicHostname:ec2-12-34-56-89.compute-1.amazonaws.com PublicIP:12.34.56.89 LocalHostname:ip-87-65-43-21.ec2.internal LocalIP:87.65.43.21 AvailabilityZone:{}}

2020/07/21 14:48:21 Got interruption event from channel {InstanceID:i-1234567890abcdef0 InstanceType:m4.large PublicHostname:ec2-12-34-56-89.compute-1.amazonaws.com PublicIP:12.34.56.89 LocalHostname:ip-87-65-43-21.ec2.internal LocalIP:87.65.43.21 AvailabilityZone:{}} {EventID:spot-itn-b8eb1e0bd5080acb927a7f124aa20792d925f8c67e1bc57a084086ea2887bb40 Kind:SPOT_ITN Description:Spot ITN received. Instance will be interrupted at 2020-07-21T14:50:05Z

Is this info just not included in the testing environment? I was able to curl this endpoint on my eks nodes and it did return the correct info, just want to double check this was a testing env issue and not something else!

DP19

comment created time in 2 months

issue commentaws/aws-node-termination-handler

Improve InterruptionEvent with more details

@bwagner5 Thanks! PR #210 adds this info

mmiranda

comment created time in 2 months

PR opened aws/aws-node-termination-handler

Update ec2metadata package to include AvailabilityZone info

Issue #, if available: 183

Description of changes:

Added AvailabilityZone field to NodeMetadata struct to be used in the webTemplate. Added tests and updated docs per this change.

By submitting this pull request, I confirm that my contribution is made under the terms of the Apache 2.0 license.

+15 -9

0 comment

4 changed files

pr created time in 2 months

fork DP19/aws-node-termination-handler

A Kubernetes Daemonset to gracefully handle EC2 instance shutdown

https://aws.amazon.com/ec2

fork in 2 months

issue commentaws/aws-node-termination-handler

Improve InterruptionEvent with more details

I'd like to take stab at this!

mmiranda

comment created time in 2 months

issue commentdocker/for-mac

[Kubernetes failed to start] cannot list Stacks from compose.docker.com/v1beta1 api server: timed out waiting for the condition

I saw this issue as well, however kubernetes was reachable. Looking into it a little more i saw some pods in the docker namespace were pending and couldn't be put on nodes due to some taints that they had. Being its a single node cluster I took a look and saw that there was a taint for "Disk pressure" meaning the disk image for docker for desktop was almost full. I increased the size of the disk and once it was applied and restarted it came up without issue.

Could be the same issue you're seeing

mogi86

comment created time in 2 months

more