|By Joe Zeto||
|July 18, 2012 01:51 PM EDT||
Consumers are continuing to adopt multiple connected devices and video content is expected to reach more than 70 percent of global traffic. This growth and the increased reliance on wireless networks is putting stress on existing 802.11a/b/g/n networks. As a result of this high usage, users are likely to experience deteriorated performance, choppy videos and slower load times. At a time when IT managers report that network users are now averaging more than one Wi-Fi connected device per person, solutions to handle the rapid growth of devices are at a premium.
The next generation of the 802.11 standard, or IEEE 802.11ac, promises to finally break the wireless Ethernet gigabit barrier. This technology will deliver higher bandwidth while retaining better quality of experience (QoE) for end users, and is expected to be adopted rapidly into all markets: residential, enterprise, carrier and large venue.
Some of the first applications for 802.11ac's faster speeds will better residential video streaming, data syncing between mobile devices, and data backup. Streaming digital media between devices faster and simultaneously connecting more wireless devices will be some of the starting benefits for consumers and enterprises. In terms of service providers, they will be able to deploy the new technology to offload traffic from congested 3G and 4G-LTE cellular networks, and in dense operator hotspots 802.11ac will supply better performance to more users.
To date, all 802.11 revisions have focused on increasing transport speeds, which lead to higher traffic delivery rates and ultimately to faster response times as experienced by the end user. The introduction of 802.11n brought advances of MIMO (multiple-in, multiple-out) to deliver traffic over multiple spatial streams, and packet aggregation. MIMO delivered marked improvements in physical transport rates, enabling more bits per second to be transmitted than ever before over Wi-Fi. Packet aggregation delivered equally impressive improvements in transport experience, allowing devices to send more data once they had gained access to the wireless media. The new 802.11ac protocol is continuing down this path by preserving aggregation techniques, advancing the physical transport rates yet again, and introducing the concept of parallel transport into Wi-Fi through a technique known as Multi-User MIMO (MU-MIMO), where multiple client devices are receiving packets concurrently.
This is the first time Wi-Fi history that directed traffic can be delivered to multiple client devices at the same time. This ability has significant impact on delivery of content to any location with multiple users, especially where content is revenue-generating or critical.
Achieving Increased Gigabit+ Performance with 802.11ac
In order to reach the best performance, 802.11ac uses a variety of advancements and addresses the need for performance improvement through three primary initiatives:
- Increasing Raw Bandwidth allows for the higher speeds associated with 802.11ac. It makes use of a higher rate encoding scheme known as 256-QAM, which transmits 33 percent more data than the 64-QAM used in the 802.11n standard. Signal-to-noise ratios that worked for 802.11n are no longer sufficient because the difference in detectable signal level is now significantly smaller.
- Multi-user Support makes 802.11ac a real information superhighway, unlike its predecessors that only allowed one device to transmit at a time. MU-MIMO allows an access point to transmit data to multiple client devices on the same channel at the same time. It works by directing some of the spatial streams to one client and other spatial streams to a second client. MU-MIMO is critical to performance improvements in environments with high client counts.
- Individual Client Channel Optimization is also a major performance booster. The concept behind channel optimization is transmit beamforming (TxBF). The reflections and attenuations, common during the transmission of 802.11 signals, have a significant performance impact on overall network performance. With TxBF, the access point communicates with the client devices to determine the types of impairment that are present in the environment. Then the access point "precodes" the transmitted frame with the inverse of the impairment such that when the next frame is transmitted and transformed by the medium, it is received as a clean frame by the client. Since no two clients are in the same location, TxBF needs to be applied on a client-by-client basis and constantly updated to reflect the changing environment.
Overcoming Technical Challenges
One of the biggest frustration for developers and users of 802.11 is that it needs to work with previous versions. It can also be extremely difficult to identify the root cause of development problems. For example, when an application performs poorly, it is often hard to determine if it is due to an environmental, client, or network issue. The various devices in an 802.11 network are highly correlated so an issue in one area quickly ripples through to many other areas. Developers have lacked an effective means to assess the total picture from the RF to the application layer.
IEEE 802.11ac makes this problem significantly more challenging. In addition to being deployed into an existing environment with ten years' worth of previous releases, 802.11ac makes use of advanced technologies that are substantially more complex and demanding than previous versions. This latest generation of 802.11 requires a rethinking of how the technology is developed and tested to include a much more holistic view through the product development life cycle.
Traditionally, the RF section is verified using one set of equipment, and then the upper layer functions are tested using a second set of tools. The overall technical complexity and the introduction of new technologies such as TxBF demand coordination and control between the different layers of the protocol stack. Without this coordination, it would be difficult to utilize these functions and to quickly pinpoint performance issues.
802.11ac brings the promise of moving Wi-Fi into the limelight as a trusted and capable communication protocol, and will require equipment and rigor to match. The new generation of testing should be able to decode every frame in real-time and determine each frame's RF characteristics, as well as their frame-level performance, and generate every frame without limitation in real-time to adequately test receiver performance. Previous approaches use a digitized data record approach for both generation and analysis, creating or capturing what are known as I/Q files, and equipment typically adapted from the general-purpose RF domain. This result in equipment being capable of a single spatial stream, and able to generate or capture a small fraction of the frames required to perform testing. To meet the need, the approach needs to be able to generate and analyze all frames in real-time to the limit of the specification, tightly integrate RF and MAC functionality in 802.11ac, and include integral, real-time channel emulation to address TxBF performance.
Increasing Performance for All Markets
Gigabit+ performance for residential, enterprise, carrier and large venue markets is possible with the 802.11ac standard. But to realize the performance and density promise, chip and hardware developers must navigate some significant technical challenges, as detailed in this article. They must ensure graceful migrations from existing deployed solutions by providing backward compatibility and delivering high performance RF transmission and receive performance with a wide variety of signals. They must maintain high performance to multiple clients under the channel conditions that will exist in real deployments, while at the same time provide the high reliability and feature robustness to enable enterprise and carrier grade 802.11 adoption. Ultimately, the developers need to ensure that the key application traffic - most notably video - can be delivered with quality.
JFrog has announced a powerful technology for managing software packages from development into production. JFrog Artifactory 4 represents disruptive innovation in its groundbreaking ability to help development and DevOps teams deliver increasingly complex solutions on ever-shorter deadlines across multiple platforms JFrog Artifactory 4 establishes a new category – the Universal Artifact Repository – that reflects JFrog's unique commitment to enable faster software releases through the first pla...
Oct. 9, 2015 12:30 AM EDT Reads: 627
Overgrown applications have given way to modular applications, driven by the need to break larger problems into smaller problems. Similarly large monolithic development processes have been forced to be broken into smaller agile development cycles. Looking at trends in software development, microservices architectures meet the same demands. Additional benefits of microservices architectures are compartmentalization and a limited impact of service failure versus a complete software malfunction....
Oct. 9, 2015 12:00 AM EDT Reads: 184
IT data is typically silo'd by the various tools in place. Unifying all the log, metric and event data in one analytics platform stops finger pointing and provides the end-to-end correlation. Logs, metrics and custom event data can be joined to tell the holistic story of your software and operations. For example, users can correlate code deploys to system performance to application error codes.
Oct. 9, 2015 12:00 AM EDT Reads: 211
Somebody call the buzzword police: we have a serious case of microservices-washing in progress. The term “microservices-washing” is derived from “whitewashing,” meaning to hide some inconvenient truth with bluster and nonsense. We saw plenty of cloudwashing a few years ago, as vendors and enterprises alike pretended what they were doing was cloud, even though it wasn’t. Today, the hype around microservices has led to the same kind of obfuscation, as vendors and enterprise technologists alike ar...
Oct. 9, 2015 12:00 AM EDT Reads: 446
With containerization using Docker, the orchestration of containers using Kubernetes, the self-service model for provisioning your projects and applications and the workflows we built in OpenShift is the best in class Platform as a Service that enables introducing DevOps into your organization with ease. In his session at DevOps Summit, Veer Muchandi, PaaS evangelist with RedHat, will provide a deep dive overview of OpenShift v3 and demonstrate how it helps with DevOps.
Oct. 8, 2015 11:30 PM EDT Reads: 649
The web app is agile. The REST API is agile. The testing and planning are agile. But alas, data infrastructures certainly are not. Once an application matures, changing the shape or indexing scheme of data often forces at best a top down planning exercise and at worst includes schema changes that force downtime. The time has come for a new approach that fundamentally advances the agility of distributed data infrastructures. Come learn about a new solution to the problems faced by software organ...
Oct. 8, 2015 11:00 PM EDT Reads: 860
Between the compelling mockups and specs produced by analysts, and resulting applications built by developers, there exists a gulf where projects fail, costs spiral, and applications disappoint. Methodologies like Agile attempt to address this with intensified communication, with partial success but many limitations. In his session at DevOps Summit, Charles Kendrick, CTO and Chief Architect at Isomorphic Software, will present a revolutionary model enabled by new technologies. Learn how busine...
Oct. 8, 2015 10:45 PM EDT Reads: 275
SYS-CON Events announced today that Dyn, the worldwide leader in Internet Performance, will exhibit at SYS-CON's 17th International Cloud Expo®, which will take place on November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Dyn is a cloud-based Internet Performance company. Dyn helps companies monitor, control, and optimize online infrastructure for an exceptional end-user experience. Through a world-class network and unrivaled, objective intelligence into Internet condit...
Oct. 8, 2015 10:00 PM EDT Reads: 588
The last decade was about virtual machines, but the next one is about containers. Containers enable a service to run on any host at any time. Traditional tools are starting to show cracks because they were not designed for this level of application portability. Now is the time to look at new ways to deploy and manage applications at scale. In his session at @DevOpsSummit, Brian “Redbeard” Harrington, a principal architect at CoreOS, will examine how CoreOS helps teams run in production. Attende...
Oct. 8, 2015 09:45 PM EDT Reads: 1,249
The APN DevOps Competency highlights APN Partners who demonstrate deep capabilities delivering continuous integration, continuous delivery, and configuration management. They help customers transform their business to be more efficient and agile by leveraging the AWS platform and DevOps principles.
Oct. 8, 2015 09:30 PM EDT Reads: 231
Containers are revolutionizing the way we deploy and maintain our infrastructures, but monitoring and troubleshooting in a containerized environment can still be painful and impractical. Understanding even basic resource usage is difficult - let alone tracking network connections or malicious activity. In his session at DevOps Summit, Gianluca Borello, Sr. Software Engineer at Sysdig, will cover the current state of the art for container monitoring and visibility, including pros / cons and li...
Oct. 8, 2015 09:30 PM EDT Reads: 206
Containers are changing the security landscape for software development and deployment. As with any security solutions, security approaches that work for developers, operations personnel and security professionals is a requirement. In his session at @DevOpsSummit, Kevin Gilpin, CTO and Co-Founder of Conjur, will discuss various security considerations for container-based infrastructure and related DevOps workflows.
Oct. 8, 2015 09:15 PM EDT Reads: 220
Manufacturing has widely adopted standardized and automated processes to create designs, build them, and maintain them through their life cycle. However, many modern manufacturing systems go beyond mechanized workflows to introduce empowered workers, flexible collaboration, and rapid iteration. Such behaviors also characterize open source software development and are at the heart of DevOps culture, processes, and tooling.
Oct. 8, 2015 09:00 PM EDT Reads: 1,091
Any Ops team trying to support a company in today’s cloud-connected world knows that a new way of thinking is required – one just as dramatic than the shift from Ops to DevOps. The diversity of modern operations requires teams to focus their impact on breadth vs. depth. In his session at DevOps Summit, Adam Serediuk, Director of Operations at xMatters, Inc., will discuss the strategic requirements of evolving from Ops to DevOps, and why modern Operations has begun leveraging the “NoOps” approa...
Oct. 8, 2015 06:00 PM EDT Reads: 138
Internet of Things (IoT) will be a hybrid ecosystem of diverse devices and sensors collaborating with operational and enterprise systems to create the next big application. In their session at @ThingsExpo, Bramh Gupta, founder and CEO of robomq.io, and Fred Yatzeck, principal architect leading product development at robomq.io, discussed how choosing the right middleware and integration strategy from the get-go will enable IoT solution developers to adapt and grow with the industry, while at th...
Oct. 8, 2015 06:00 PM EDT Reads: 2,163
In their session at DevOps Summit, Asaf Yigal, co-founder and the VP of Product at Logz.io, and Tomer Levy, co-founder and CEO of Logz.io, will explore the entire process that they have undergone – through research, benchmarking, implementation, optimization, and customer success – in developing a processing engine that can handle petabytes of data. They will also discuss the requirements of such an engine in terms of scalability, resilience, security, and availability along with how the archi...
Oct. 8, 2015 05:00 PM EDT Reads: 399
“All our customers are looking at the cloud ecosystem as an important part of their overall product strategy. Some see it evolve as a multi-cloud / hybrid cloud strategy, while others are embracing all forms of cloud offerings like PaaS, IaaS and SaaS in their solutions,” noted Suhas Joshi, Vice President – Technology, at Harbinger Group, in this exclusive Q&A with Cloud Expo Conference Chair Roger Strukhoff.
Oct. 8, 2015 05:00 PM EDT Reads: 438
DevOps has often been described in terms of CAMS: Culture, Automation, Measuring, Sharing. While we’ve seen a lot of focus on the “A” and even on the “M”, there are very few examples of why the “C" is equally important in the DevOps equation. In her session at @DevOps Summit, Lori MacVittie, of F5 Networks, will explore HTTP/1 and HTTP/2 along with Microservices to illustrate why a collaborative culture between Dev, Ops, and the Network is critical to ensuring success.
Oct. 8, 2015 04:45 PM EDT Reads: 129
Application availability is not just the measure of “being up”. Many apps can claim that status. Technically they are running and responding to requests, but at a rate which users would certainly interpret as being down. That’s because excessive load times can (and will be) interpreted as “not available.” That’s why it’s important to view ensuring application availability as requiring attention to all its composite parts: scalability, performance, and security.
Oct. 8, 2015 03:00 PM EDT Reads: 393
Saviynt Inc. has announced the availability of the next release of Saviynt for AWS. The comprehensive security and compliance solution provides a Command-and-Control center to gain visibility into risks in AWS, enforce real-time protection of critical workloads as well as data and automate access life-cycle governance. The solution enables AWS customers to meet their compliance mandates such as ITAR, SOX, PCI, etc. by including an extensive risk and controls library to detect known threats and b...
Oct. 8, 2015 03:00 PM EDT Reads: 209