-
-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug- Veth is not releasing when the MicroVM restart #22
Comments
results=
results=
|
Thanks for the bug report, I'll take a look when I have available time to work on this. |
Bug is due to the following issue that has been solved in firecracker-go-sdk upstream firecracker-microvm/firecracker-go-sdk#428. Unfortunately, this plugin is not up-to-date with the firecraker-go-sdk upstream. |
@Joffref I tried cherry-picking that commit on top of the latest pre-1.0-compatible firecracker-go-sdk and basing the firecracker-task-driver off of it, but I'm still getting the same issue. Is there anything else I need to do to get this integrated with the task driver? My forked firecracker-go-sdk and firecracker-task-driver are available at https://github.com/radixbio/firecracker-go-sdk and https://github.com/radixbio/firecracker-task-driver respectively for context. Thanks for linking the issue!! |
@hitendra-galaxyweblinks were you using the cni-plugins from my fork? if not that the cause of the segfault. Unfortunately by mistake a erased my fork with the changes need a month ago, but luckily seems the current sdk fixed the issue. I'll try to try this during the week should be and easy fix. |
I am working on a project where I am deploying the micro VM using nomad, the driver is working fine but there is an issue, when the VM is restarting or when we are updating the job with new rootfs, the VM is failing to start. When I dug more, I found the driver is unable to assign IP to the new VM as the IP range is exhausted. When I troubleshoot more I found that there are so many Firecracker VM is created, with each restart it provision more and more VMs uncontrollably and exhaust the whole IP range. Kindly refer to the screenshot to support my case. I am still trying to figure out this behavior of the driver. Technically it should update the rootfs and restart the VM with new rootfs and assign the same IP or new one but why it is creating the VM in the background? I would really appreciate the help here.
The text was updated successfully, but these errors were encountered: