Click here to close now.




















Welcome!

Microservices Expo Authors: Carmen Gonzalez, Yeshim Deniz, Liz McMillan, Pat Romanski, Samuel Scott

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog, API Journal, IoT User Interface, Agile Computing, Cloud Security

@CloudExpo: Article

Beyond Intrusion Detection: Eight Best Practices for Cloud SIEM Deployment

Improving visibility across the enterprise

For all the right reasons, your company has been thinking about deploying SIEM…to create an alert system when those with less than good intentions come knocking; to remediate potential network threats; to comply with federal, state or industry regulations; and identify the risks and vulnerabilities throughout the enterprise IT infrastructure and architecture. If you maintain even a modest (SMB -> Fortune 1000) organization that has any online identity, SIEM should be the cornerstone of your asset protection strategy.

First and foremost, SIEM (and to a certain extent log management) is about visibility. Who is doing what and when on your network. It is as much about understanding the holistic landscape of your infrastructure as it is protecting proprietary assets. Without it, it’s akin to coaching the Big Game without any idea who is the opponent; or for that matter if you even have a starting left guard.

But fun metaphors aside, SIEM is a critical enterprise tool. And just like any enterprise solution, it requires forethought, vigilance and most importantly, a good game plan. And when deployed properly it can change your IT department from infrastructure-based, to information-centric. And as such you get to make better decisions, faster.

And with every technology there are best practices and pitfalls. In past articles I have spoken at length regarding the advantages of deploying and managing SIEM in the cloud. Many of these surround the affordability, manageability, control and capability of the solution. For many, security from the cloud is still an emerging concept. But for those who’ve already made the leap, they are reaping the significant benefits. But I want to move beyond the arguments of “going cloud” when deciding on security solutions. Today I want to focus on what happens next. How do you start collecting that ROI once a cloud-based security-as-a-service has been chosen?

The reason most enterprise deployments fail (on premise or cloud) can be typically traced to two causes: (1.) Lack of buy-in from the executive level or employee resistance to change, but more often the culprit is (2.) lack of vision or process. Too many companies jump in and apply a solution because they heard it was important or were sold a Porsche when all they needed was a family SUV. Of course one of the benefits of cloud-based security is the ability to "buy" the SUV and instantly scale up to that Porsche, if and when, the business need requires it (without touching CapEx budgets!)! But with that here are 8 best practices you should implement when moving forward with your cloud-based security initiative:

Best Practice #1: Identify your goals and match your scope to them. There are five questions you need to ask before moving forward with any deployment. 1. WHY do you need SIEM (compliance? user and/or partner expansion? BYOD? Breach detection?) HOW will SIEM be deployed to properly address these issues (what processes, functionality and capabilities are needed; which needs to be outsourced/replaced/improved) WHAT needs to be collected, analyzed and reported? HOW BIG does the deployment need to scale to accurately and cost effectively meet your specific business need? And WHERE is the information situated that should/must be monitored?

Best practice #2: Incremental usage. The quickest route to success is taking baby steps. The idea is to prove the concept and then expand the scope. To some this might be to start with log management and add  SIEM once you understand the requirements, commitment and volume. Now because security-as-a-service is so flexible and can ramp up or down instantly, an easy entry point might be to start with only those elements that fulfill compliance. The project might be overwhelming, but if you take it in bite-sized phases, you will find the victories come easier and the ROI is justified. When dealing with a cloud security deployment, it is easy to turn on the fire hose when only a garden hose is needed. But the beauty of a cloud deployment is the ease and flexibility of scaling. Again, another example of incremental usage would be either to apply SIEM against specific use case scenarios or possibly just migrate a division or a department or a function (as opposed to the entire enterprise).

Best Practice #3: Determine what IS and ISN’T a threat to your network. Returning to the fire hose metaphor, when deploying a SIEM initiative, it is very easy to get lost in a sea of data. It can be like trying to drink from that proverbial fire hose. The trick is to recognize what constitutes a true risk and eliminate false positives. And this requires some internal analysis to create a series of rules that sift out the white noise and differentiate “normal” traffic from suspicious activity. For instance, if there is an attempted access to your partner portal from Russia—is that normal? Do you even have a partner in Minsk? But even a simple filter isn’t quite enough. Risk is three dimensional and it can hide in plain sight. That’s why you continue to filter based on time of day, IP address, server, attempts, network availability and a myriad of other forensic qualifiers before the alert is grave enough to require immediate attention.

Best practice #4: Map response plans. Now that an incident gets your attention, what do you do? Do you launch an account investigation, suspend the user, deactivate a password, apply a  denial-of-service against the IP or a number of remediations based on the severity, vulnerability and identity of the transgressor. This goes back to workflow and process. Who is going to what to whom and how? SIEM is a process-reliant technology. You simply can’t flip a switch and say you’ve put up a magic forcefield around your network. Your response plan is your blueprint to closing the vulnerability gaps and ensuring compliance.

Best practice #5 Correlate data from multiple sources. The practice of situational awareness is what adds the muscle into a SIEM initiative. Like #4, it isn’t enough to plug in a solution and press “go.” Situational awareness takes into account a multitude of different endpoints, servers, data streams, assets and inventories, events and flows, from across the enterprise and puts information into context.  Context is the most important portion of risk assessment.  For example, a shark is a threat. However if that shark is 10 miles away, it is not a direct or immediate threat. Doesn't mean you're not vulnerable if that shark gets hungry. Having an engine that not only creates accurate perspective, but analyzes, understands and acts upon behaviors is key. And to do that a centralized SIEM engine needs the data from more than just a single source or single server.

Best Practice #6: Requires Real time monitoring 7/24/365. For many companies this is a challenge, but hackers don’t sleep. And although a great deal of SIEM and Log Management is automated, it still requires the vigilance of 24 hour monitoring. Trees might be falling in the forest, but if there is no one to see them, breaches occur, networks are compromised. I’ve witnessed plenty of IT departments that don’t have the resources. Again, this is a considerable advantage that security-as-s-service provides and allows you to sleep just a little better at night. Knowing that this one crucial element of your security is professionally addressed without additional staff or budget makes the cloud that much more valuable.

Best Practice #7 Remain calm! One thing we’ve noticed is that soon after the deployment of a SIEM/Log Management it seems there are alerts and issues you never dreamed about. Things are bound to look worse before they get better and it can seem overwhelming; kind of opening a Pandora’s Box of malware and botnets.  For the most part it is because you now know what you didn’t know before.  In some respect it is like looking at your hotel room comforter under black light and a microscope. But once you realize what you’re looking at and that much or the remediation can be automated, soon, (with a bit of fine tuning and normalizing correlation feeds) you will be measure that the anomalous events lessen and the alert prioritizations allow you to make timely and intelligent decisions.

Best practice #8: Evolution. Security is a moving target. You need to revisit you processes and workflows every few months to make sure you are up to date with compliance requirements, new users/access points and expanded or redefined workflows. This is more than recognizing the latest virus threats. New users access your network with regularity. New layers of regulations are added. There are new applications requiring monitoring. All in all, by giving your cloud-based SIEM and log management solutions the new and necessary data, your enterprise will be more secure than it was yesterday.

More Stories By Kevin Nikkhoo

With more than 32 years of experience in information technology, and an extensive and successful entrepreneurial background, Kevin Nikkhoo is the CEO of the dynamic security-as-a-service startup Cloud Access. CloudAccess is at the forefront of the latest evolution of IT asset protection--the cloud.

Kevin holds a Bachelor of Science in Computer Engineering from McGill University, Master of Computer Engineering at California State University, Los Angeles, and an MBA from the University of Southern California with emphasis in entrepreneurial studies.

@MicroservicesExpo Stories
The 5th International DevOps Summit, co-located with 17th International Cloud Expo – being held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA – announces that its Call for Papers is open. Born out of proven success in agile development, cloud computing, and process automation, DevOps is a macro trend you cannot afford to miss. From showcase success stories from early adopters and web-scale businesses, DevOps is expanding to organizations of all sizes, including the ...
The 17th International Cloud Expo has announced that its Call for Papers is open. 17th International Cloud Expo, to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, APM, APIs, Microservices, Security, Big Data, Internet of Things, DevOps and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding bu...
ElasticBox, the agile application delivery manager, announced freely available public boxes for the DevOps community. ElasticBox works with enterprises to help them deploy any application to any cloud. Public boxes are curated reference boxes that represent some of the most popular applications and tools for orchestrating deployments at scale. Boxes are an adaptive way to represent reusable infrastructure as components of code. Boxes contain scripts, variables, and metadata to automate proces...
To support developers and operations professionals in their push to implement DevOps principles for their infrastructure environments, ProfitBricks, a provider of cloud infrastructure, is adding support for DevOps tools Ansible and Chef. Ansible is a platform for configuring and managing data center infrastructure that combines multi-node software deployment, ad hoc task execution, and configuration management, and is used by DevOps professionals as they use its playbooks functionality to autom...
It’s been proven time and time again that in tech, diversity drives greater innovation, better team productivity and greater profits and market share. So what can we do in our DevOps teams to embrace diversity and help transform the culture of development and operations into a true “DevOps” team? In her session at DevOps Summit, Stefana Muller, Director, Product Management – Continuous Delivery at CA Technologies, answered that question citing examples, showing how to create opportunities for ...
Puppet Labs has announced the next major update to its flagship product: Puppet Enterprise 2015.2. This release includes new features providing DevOps teams with clarity, simplicity and additional management capabilities, including an all-new user interface, an interactive graph for visualizing infrastructure code, a new unified agent and broader infrastructure support.
Skeuomorphism usually means retaining existing design cues in something new that doesn’t actually need them. However, the concept of skeuomorphism can be thought of as relating more broadly to applying existing patterns to new technologies that, in fact, cry out for new approaches. In his session at DevOps Summit, Gordon Haff, Senior Cloud Strategy Marketing and Evangelism Manager at Red Hat, discussed why containers should be paired with new architectural practices such as microservices rathe...
SYS-CON Events announced today that Pythian, a global IT services company specializing in helping companies leverage disruptive technologies to optimize revenue-generating systems, has been named “Bronze Sponsor” of SYS-CON's 17th Cloud Expo, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Founded in 1997, Pythian is a global IT services company that helps companies compete by adopting disruptive technologies such as cloud, Big Data, advance...
Puppet Labs is pleased to share the findings from our 2015 State of DevOps Survey. We have deepened our understanding of how DevOps enables IT performance and organizational performance, based on responses from more than 20,000 technical professionals we’ve surveyed over the past four years. The 2015 State of DevOps Report reveals high-performing IT organizations deploy 30x more frequently with 200x shorter lead times. They have 60x fewer failures and recover 168x faster
Containers are not new, but renewed commitments to performance, flexibility, and agility have propelled them to the top of the agenda today. By working without the need for virtualization and its overhead, containers are seen as the perfect way to deploy apps and services across multiple clouds. Containers can handle anything from file types to operating systems and services, including microservices. What are microservices? Unlike what the name implies, microservices are not necessarily small,...
DevOps has traditionally played important roles in development and IT operations, but the practice is quickly becoming core to other business functions such as customer success, business intelligence, and marketing analytics. Modern marketers today are driven by data and rely on many different analytics tools. They need DevOps engineers in general and server log data specifically to do their jobs well. Here’s why: Server log files contain the only data that is completely full and accurate in th...
All major researchers estimate there will be tens of billions devices - computers, smartphones, tablets, and sensors - connected to the Internet by 2020. This number will continue to grow at a rapid pace for the next several decades. With major technology companies and startups seriously embracing IoT strategies, now is the perfect time to attend @ThingsExpo, November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Learn what is going on, contribute to the discussions, and e...
SYS-CON Events announced today that G2G3 will exhibit at SYS-CON's @DevOpsSummit Silicon Valley, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. Based on a collective appreciation for user experience, design, and technology, G2G3 is uniquely qualified and motivated to redefine how organizations and people engage in an increasingly digital world.
Whether you like it or not, DevOps is on track for a remarkable alliance with security. The SEC didn’t approve the merger. And your boss hasn’t heard anything about it. Yet, this unruly triumvirate will soon dominate and deliver DevSecOps faster, cheaper, better, and on an unprecedented scale. In his session at DevOps Summit, Frank Bunger, VP of Customer Success at ScriptRock, will discuss how this cathartic moment will propel the DevOps movement from such stuff as dreams are made on to a prac...
Docker containerization is increasingly being used in production environments. How can these environments best be monitored? Monitoring Docker containers as if they are lightweight virtual machines (i.e., monitoring the host from within the container), with all the common metrics that can be captured from an operating system, is an insufficient approach. Docker containers can’t be treated as lightweight virtual machines; they must be treated as what they are: isolated processes running on hosts....
SYS-CON Events announced today that DataClear Inc. will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. The DataClear ‘BlackBox’ is the only solution that moves your PC, browsing and data out of the United States and away from prying (and spying) eyes. Its solution automatically builds you a clean, on-demand, virus free, new virtual cloud based PC outside of the United States, and wipes it clean...
SYS-CON Events announced today that HPM Networks will exhibit at the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. For 20 years, HPM Networks has been integrating technology solutions that solve complex business challenges. HPM Networks has designed solutions for both SMB and enterprise customers throughout the San Francisco Bay Area.
SYS-CON Events announced today the Containers & Microservices Bootcamp, being held November 3-4, 2015, in conjunction with 17th Cloud Expo, @ThingsExpo, and @DevOpsSummit at the Santa Clara Convention Center in Santa Clara, CA. This is your chance to get started with the latest technology in the industry. Combined with real-world scenarios and use cases, the Containers and Microservices Bootcamp, led by Janakiram MSV, a Microsoft Regional Director, will include presentations as well as hands-on...
Microservice architecture is fast becoming a go-to solution for enterprise applications, but it's not always easy to make the transition from an established, monolithic infrastructure. Lightweight and loosely coupled, building a set of microservices is arguably more difficult than building a monolithic application. However, once established, microservices offer a series of advantages over traditional architectures as deployment times become shorter and iterating becomes easier.
The pricing of tools or licenses for log aggregation can have a significant effect on organizational culture and the collaboration between Dev and Ops teams. Modern tools for log aggregation (of which Logentries is one example) can be hugely enabling for DevOps approaches to building and operating business-critical software systems. However, the pricing of an aggregated logging solution can affect the adoption of modern logging techniques, as well as organizational capabilities and cross-team ...