|By Don MacVittie||
|December 12, 2012 10:00 AM EST||
All the goodness FPGAs bring hardware in general, and ADO hardware in particular.
In two previous installments, I talked at a high level about the uses of FPGAs, risk mitigation, and the potential benefits. Today I’d like to delve into the benefits that the industry in general, and F5 in particular, gain from using FPGAs, and why it matters to IT. If you’re a regular reader, you know that I try not to be a chorus line for F5 solutions, but don’t shy away from talking about them when it fits the topic. That will continue with this post. While I will use F5 for the specifics, the benefits can be generalized to the bulk of the industry.
Used to be, way back in the day, everyone walked everywhere. That worked for a long period of world history. The horse was adopted for longer trips, and it about doubled travel speed, but still, the bulk of the world populace walked nearly all of the time. Then along came cars, and they enabled a whole lot of things. One of the great benefits that the automobile introduced was the ability to be more agile. By utilizing the machinery, you could move from one town to another relatively quickly. You could even work in a town 30 miles – a days’ walk for a physically fit person – from your home. At this point in human – or at least first world – history, walking is a mode of transportation that is rarely used for important events. There are some cities so tightly packed that walking makes sense, but for most of us, we take a car the vast majority of the time. When speed is not of the essence – say when you take a walk with a loved one – the car is left behind, but for day-to-day transport, the car is the go-to tool.
There is a corollary to this phenomenon in the Application Delivery world. While in some scenarios, a software ADC will do the trick, there are benefits to hardware that mean if you have it, you’ll use the hardware much more frequently. This is true of far more than ADCs, but bear with me, I do work for an ADC vendor . There are some things that can just be done more efficiently in hardware, and some things that are best left (normally due to complexity) to software. In the case of FPGAs, low-level operations that do a lot of repetitive actions are relatively easily implemented – even to the point of FPGA and/or programming tools for FPGAs coming with certain pre-built layouts at this point. As such, certain network processing that is latency-sensitive and can be done with little high-level logic are well suited to FPGA processing. When a packet can be processed in X micro-seconds in FPGA, or in X^3 milliseconds by the time it passes through the hardware, DMA transfer, firmware/network stack, and finally lands in software that can manipulate it, definitely go with the FPGA option if possible.
And that’s where a lot of the benefits of FPGAs in the enterprise are being seen. Of course you don’t want to have your own FPGA shop and have to maintain your own installation program to reap the benefits. But vendors have sets of hardware that are largely the same and are produced en-masse. It makes sense that they would make use of FPGAs, and they do. Get that packet off the wire, and if it meets certain criteria, turn it around and get it back on the wire with minor modifications.
But that’s not all. While it was a great step to be able to utilize FPGAs in this manner and not have to pay the huge up-front fees of getting an ASIC designed and a run of them completed, the use of FPGAs didn’t stop there – indeed, it is still growing and changing. The big area that has really grown the usage of ever-larger FPGAs is in software assistance. Much like BIOS provides discrete functionality that software can call to achieve a result, FPGAs can define functions with register interface that are called directly from software – not as a solution, but as an incremental piece of the solution. This enables an increase in the utilization of FPGAs and if the functions are chosen carefully, an improvement in the overall performance of the system the FPGAs are there to support. It is, essentially, offloading workload from software. When that offload is of computationally intensive operations, the result can be a huge performance improvement. Where a software solution might have a function call, hardware can just do register writes and reads, leaving the system resources less taxed. Of course if the operation requires a lot of data storage memory, it still will, which is why I mentioned “computationally expensive”.
The key thing is to ask your vendor (assuming they use FPGAs) what they’re doing with them, and what benefit you see. It is a truth that the vast majority of vendors go to FPGAs for their own benefit, but that is not exclusive of making things better for customers. So ask them how you, as a customer, benefit.
And when you wonder why a VM can’t perform every bit as well as custom hardware, well the answer is at least partially above. The hardware functionality of custom devices must be implemented in software for a VM, and that software then runs on not one, but two operating systems, and eventually calls general purpose hardware. While VMs, like feet, are definitely good for some uses, when you need your app to be the fastest it can possibly be, hardware – specifically FPGA enhanced hardware – is the best answer, much as the car is the best answer for daily travel in most of the world. Each extra layer – generic hardware, the host operating system, the virtual network, and the guest operating system – adds cost to processing. The lack of an FPGA does too, because those low-level operations must be performed in software.
So know your needs, use the right tool for the job. I would not drive a car to my neighbors’ house – 200 feet away – nor would I walk from Green Bay to Cincinnati (just over 500 miles). Know what your needs are and your traffic is like, then ask about FPGA usage. And generalize this… To network switches, WAPs, you name it. You’re putting it into your network, so that IS your business.
Walking in Ust-Donetsk
And yeah, you’ll hear more on this topic before I wrap up the Bare Metal Blog series, but for now, keep doing what you do so well, and I’ll be back with more on testing soon.
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor – all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
Dec. 8, 2016 06:15 AM EST Reads: 5,909
Get deep visibility into the performance of your databases and expert advice for performance optimization and tuning. You can't get application performance without database performance. Give everyone on the team a comprehensive view of how every aspect of the system affects performance across SQL database operations, host server and OS, virtualization resources and storage I/O. Quickly find bottlenecks and troubleshoot complex problems.
Dec. 8, 2016 05:00 AM EST Reads: 2,159
As we enter the final week before the 19th International Cloud Expo | @ThingsExpo in Santa Clara, CA, it's time for me to reflect on six big topics that will be important during the show. Hybrid Cloud: This general-purpose term seems to provide a comfort zone for many enterprise IT managers. It sounds reassuring to be able to work with one of the major public-cloud providers like AWS or Microsoft Azure while still maintaining an on-site presence.
Dec. 8, 2016 04:30 AM EST Reads: 2,913
Between 2005 and 2020, data volumes will grow by a factor of 300 – enough data to stack CDs from the earth to the moon 162 times. This has come to be known as the ‘big data’ phenomenon. Unfortunately, traditional approaches to handling, storing and analyzing data aren’t adequate at this scale: they’re too costly, slow and physically cumbersome to keep up. Fortunately, in response a new breed of technology has emerged that is cheaper, faster and more scalable. Yet, in meeting these new needs they...
Dec. 8, 2016 01:30 AM EST Reads: 1,925
I’m a huge fan of open source DevOps tools. I’m also a huge fan of scaling open source tools for the enterprise. But having talked with my fair share of companies over the years, one important thing I’ve learned is that you can’t scale your release process using open source tools alone. They simply require too much scripting and maintenance when used that way. Scripting may be fine for smaller organizations, but it’s not ok in an enterprise environment that includes many independent teams and to...
Dec. 8, 2016 12:30 AM EST Reads: 701
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, discussed how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team at D...
Dec. 8, 2016 12:30 AM EST Reads: 1,136
In IT, we sometimes coin terms for things before we know exactly what they are and how they’ll be used. The resulting terms may capture a common set of aspirations and goals – as “cloud” did broadly for on-demand, self-service, and flexible computing. But such a term can also lump together diverse and even competing practices, technologies, and priorities to the point where important distinctions are glossed over and lost.
Dec. 7, 2016 08:45 PM EST Reads: 1,640
SYS-CON Events announced today that Dataloop.IO, an innovator in cloud IT-monitoring whose products help organizations save time and money, has been named “Bronze Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Dataloop.IO is an emerging software company on the cutting edge of major IT-infrastructure trends including cloud computing and microservices. The company, founded in the UK but now based in San Fran...
Dec. 7, 2016 07:00 PM EST Reads: 523
Monitoring of Docker environments is challenging. Why? Because each container typically runs a single process, has its own environment, utilizes virtual networks, or has various methods of managing storage. Traditional monitoring solutions take metrics from each server and applications they run. These servers and applications running on them are typically very static, with very long uptimes. Docker deployments are different: a set of containers may run many applications, all sharing the resource...
Dec. 7, 2016 06:30 PM EST Reads: 5,682
In his session at 19th Cloud Expo, Claude Remillard, Principal Program Manager in Developer Division at Microsoft, contrasted how his team used config as code and immutable patterns for continuous delivery of microservices and apps to the cloud. He showed how the immutable patterns helps developers do away with most of the complexity of config as code-enabling scenarios such as rollback, zero downtime upgrades with far greater simplicity. He also demoed building immutable pipelines in the cloud ...
Dec. 7, 2016 06:00 PM EST Reads: 1,878
Join Impiger for their featured webinar: ‘Cloud Computing: A Roadmap to Modern Software Delivery’ on November 10, 2016, at 12:00 pm CST. Very few companies have not experienced some impact to their IT delivery due to the evolution of cloud computing. This webinar is not about deciding whether you should entertain moving some or all of your IT to the cloud, but rather, a detailed look under the hood to help IT professionals understand how cloud adoption has evolved and what trends will impact th...
Dec. 7, 2016 06:00 PM EST Reads: 2,679
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
Dec. 7, 2016 05:00 PM EST Reads: 1,932
The 20th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held June 6-8, 2017, at the Javits Center in New York City, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Containers, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal ...
Dec. 7, 2016 03:45 PM EST Reads: 2,275
You have great SaaS business app ideas. You want to turn your idea quickly into a functional and engaging proof of concept. You need to be able to modify it to meet customers' needs, and you need to deliver a complete and secure SaaS application. How could you achieve all the above and yet avoid unforeseen IT requirements that add unnecessary cost and complexity? You also want your app to be responsive in any device at any time. In his session at 19th Cloud Expo, Mark Allen, General Manager of...
Dec. 7, 2016 03:30 PM EST Reads: 1,794
Financial Technology has become a topic of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 20th Cloud Expo at the Javits Center in New York, June 6-8, 2017, will find fresh new content in a new track called FinTech.
Dec. 7, 2016 02:15 PM EST Reads: 2,207
Information technology is an industry that has always experienced change, and the dramatic change sweeping across the industry today could not be truthfully described as the first time we've seen such widespread change impacting customer investments. However, the rate of the change, and the potential outcomes from today's digital transformation has the distinct potential to separate the industry into two camps: Organizations that see the change coming, embrace it, and successful leverage it; and...
Dec. 7, 2016 02:15 PM EST Reads: 3,403
"Dice has been around for the last 20 years. We have been helping tech professionals find new jobs and career opportunities," explained Manish Dixit, VP of Product and Engineering at Dice, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Dec. 7, 2016 01:45 PM EST Reads: 1,142
Rapid innovation, changing business landscapes, and new IT demands force businesses to make changes quickly. In the eyes of many, containers are at the brink of becoming a pervasive technology in enterprise IT to accelerate application delivery. In this presentation, attendees learned about the: The transformation of IT to a DevOps, microservices, and container-based architecture What are containers and how DevOps practices can operate in a container-based environment A demonstration of how ...
Dec. 7, 2016 12:45 PM EST Reads: 1,120
SYS-CON Events has announced today that Roger Strukhoff has been named conference chair of Cloud Expo and @ThingsExpo 2017 New York. The 20th Cloud Expo and 7th @ThingsExpo will take place on June 6-8, 2017, at the Javits Center in New York City, NY. "The Internet of Things brings trillions of dollars of opportunity to developers and enterprise IT, no matter how you measure it," stated Roger Strukhoff. "More importantly, it leverages the power of devices and the Internet to enable us all to im...
Dec. 7, 2016 12:30 PM EST Reads: 797
Cloud Expo, Inc. has announced today that Andi Mann returns to 'DevOps at Cloud Expo 2017' as Conference Chair The @DevOpsSummit at Cloud Expo will take place on June 6-8, 2017, at the Javits Center in New York City, NY. "DevOps is set to be one of the most profound disruptions to hit IT in decades," said Andi Mann. "It is a natural extension of cloud computing, and I have seen both firsthand and in independent research the fantastic results DevOps delivers. So I am excited to help the great t...
Dec. 7, 2016 12:00 PM EST Reads: 683