HN Theater @HNTheaterMonth

The best talks and videos of Hacker News.

Hacker News Comments on
BAFUG June 2015 - Building a Distributed Hypervisor by Michael Berman (TidalScale)

iXsystems, Inc. · Youtube · 110 HN points · 1 HN comments
HN Theater has aggregated all Hacker News stories and comments that mention iXsystems, Inc.'s video "BAFUG June 2015 - Building a Distributed Hypervisor by Michael Berman (TidalScale)".
Youtube Summary
Bay Area FreeBSD User Group, June 2015 - Building a distributed hypervisor by Michael Berman (TidalScale) at Citrix Startup Accelerator
HN Theater Rankings

Hacker News Stories and Comments

All the comments and stories posted to Hacker News that reference this video.
Mar 07, 2016 · 108 points, 3 comments · submitted by xj9
mtanski
This is interesting but short on details. Is it essentially creating a virtual many core / NUMA machine? If that's so I wonder what the overhead for such things as emulating x86 cache coherency.
tg2
Furthermore, fundamental things such as the fact CPU interconnects are in the 160+Gbps range compared to (in their example) 10gbps network prevents this from scaling as a single system would. Even in a single dual socket system it is not uncommon to pin (via numactl) processes to a single CPU in order to get full performance out of applications and prevent saturating this interconnect.

Accessing memory from a remote core over the network is seriously handicapped in this respect, and purpose built cluster systems such as infiniband rdma with bandwidth in the 100gbps realm still have issues with this latency. A single stick of previous-gen ddr3-1600 can exceed that bandwidth, with 15x faster access time.

To give an example, a very low latency (non Ethernet) network is around 1300μs, local dram (from the same socket) is 60ns. L3 cache on the same socket is 15ns.

You can be smart about pinning memory allocation to the local CPU core requesting it, and minimizing thread migration to another host, but there is no magic bullet to getting past these theoretical limits. Accessing packets arriving on a network card in host A from host B's core would halve the network bandwidth unless there is a dedicated network for the clustering.

That being said, virtualization isn't perfect either and the overhead can be substantial as we've seen when comparing to containers on bare metal.

I'd really like to take this for a test drive and benchmark it with some R jobs.

Even if it's slower, the value of not having to use cluster-aware toolkits is valuable to many, not to mention simplicity of operation.

I wish they would release an open source version before getting sucked up as a stop-loss by Intel.

mtanski
I imagine that a untuned guest operating scheduler would wreck havoc in this scenario.
This video from the Bay Area FreeBSD User Group, June 2015, does a decent job of explaining how TidalScale implements their hypervisor, spoiler, it's based off FreeBSD and Bhyve.

https://www.youtube.com/watch?v=f-ug6B6ycng

Jul 29, 2015 · 2 points, 0 comments · submitted by tachion
HN Theater is an independent project and is not operated by Y Combinator or any of the video hosting platforms linked to on this site.
~ yaj@
;laksdfhjdhksalkfj more things
yahnd.com ~ Privacy Policy ~
Lorem ipsum dolor sit amet, consectetur adipisicing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation ullamco laboris nisi ut aliquip ex ea commodo consequat. Duis aute irure dolor in reprehenderit in voluptate velit esse cillum dolore eu fugiat nulla pariatur. Excepteur sint occaecat cupidatat non proident, sunt in culpa qui officia deserunt mollit anim id est laborum.