It’s part of Cisco’s push to “Merakify” its business.
Starting up a virtual machine (VM) is not an easy task, there are a lot of things going on hidden from the plain sight and studying it alone is a challenging task. The goal of this post is to simplify the process of learning how the oVirt hypervisor works. The concept of the oVirt is illustrated in the process of starting up the VM, which covers everything from top to the bottom.
Disclaimer: I am an engineer working close to the host part of the oVirt, therefore my knowledge of the engine is limited.

Let me start with explaining the main parts oVirt hypervisor architecture. It will help a lot in understanding the overall process of the VM startup flow. Following is the simplified architecture, where I omit auxiliary components and support scripts. This will allow me to focus on the core concept without distractions.
The architecture comprises of three main components: 1) web UI and engine, 2) VDSM, 3) guest agent.
The web UI is where the user makes the first contact with the oVirt hypervisor. The web UI is the main command center of the engine and allows a user to control all aspects and states of Continue reading
Like it or not, operations teams will need to learn how to code.
![]() |
| CCNA Interview Questions |
One of my readers was wondering about the stability and scalability of large layer-2 domains implemented with VXLAN. He wrote:
If common BUM traffic (e.g. ARP) is being handled/localized by the network (e.g. NSX or ACI), and if we are managing what traffic hosts can send with micro-segmentation style filtering blocking broadcast/multicast, are large layer-2 domains still a recipe for disaster?
There are three major (fundamental) problems with large L2 domains:
Read more ...![]() |
| Fig 1.1- Cisco Router as Terminal Server |
The number of exposed and vulnerable devices online has remained largely unchanged since researchers began exploring SMI in 2010.
docker run -p 6343:6343/udp -p 8008:8008 sflow/flow-trendThe simplest way to run the software is using the docker. Configure network devices to send standard sFlow telemetry to Flow Trend. Access the web user interface on port 8008.
This is a liveblog of the HashiConf 2017 session titled “Cloud Native Infrastructure.” The speaker is Kris Nova, a Senior Developer Advocate at Microsoft. Kris, along with Justin Garrison, authored the O’Reilly Cloud Native Infrastructure book (more information here). As one of the last sessions (if not the last session) I’ll be able to attend, I’m looking forward to this session.
Kris is a self-confessed Linux lover, loves writing in Golang, is a Kubernetes maintainer, and works on Azure at Microsoft.
So, what is “cloud-native infrastructure”? To answer that, Nova first tries to answer “what is a cloud?” Nova breezes by that definition without going into any real detail (or any real definition), and proceeds to talk about what infrastructure is. Again, Nova breezes by that without providing any real definition or depth, and proceeds to ask “Why is infrastructure better in the cloud?” According to Nova, infrastructure is better in the cloud because management can be as simple as an HTTP request. The next few slides in Nova’s presentation compare the “traditional” ways of managing infrastructure (provisioning switches, patching cables, troubleshooting problems) are now, when infrastructure is in the cloud, as simple as a series Continue reading
HashiConf 2017 is a wrap for me, and as I’m sitting here at the airport lounge in Austin I’d thought I’d post links back to the liveblogs I published as well as a few thoughts on the conference overall.
First, here are links to the liveblogs published during the event:
I think it was a pretty good event. The venue (JW Marriott in Austin) seemed roughly appropriate for the number of attendees (around 800, I believe), although some additional seating during meal times would have been a good idea. The conference Wi-Fi was mostly OK, though it had its moments.
The quality of sessions varied; some sessions were very good; others, not so much (unfortunately). It would have been good to see a clearer breakdown of the sessions according to area/theme. They had 3 content tracks, but it wasn’t really clear to me if the tracks had any central theme. I, personally, bounced around all three tracks.
I did like the inclusion of high-top tables at the Continue reading
As an economic powerhouse and with a rising military and political presence around the world, you would expect, given the inherent political nature of supercomputing, that China would have multiple and massive supercomputing centers as well as a desire to spread its risk and demonstrate its technical breadth by investing in many different kinds of capability class supercomputers.
And this is precisely what China is doing, including creating its own offload accelerator, based on digital signal processors. This Matrix2000 DSP accelerator, which was unveiled at the ISC16 supercomputing event last year and which is being created by the National University …
China Arms Upgraded Tianhe-2A Hybrid Supercomputer was written by Timothy Prickett Morgan at The Next Platform.