Big Switch is an software-defined networking (SDN) startup with plenty of credentials. The company was co-founded in 2010 by Cisco veteran Kyle Forster and Guido Appenzeller, the Stanford researcher who led the development of the OpenFlow standard. Big Switch raised nearly $40 million in venture capital to date, from investors including Goldman Sachs, Redpoint Ventures, Index Ventures, and Khosla Ventures, and it finally made its debut today.
Big Switch exited stealth today with the official launch of their Big Switch Product Suite whose goal is to virtualize the network with an Open SDN with the slogan apps, apps, and more apps.
Today BigSwitch announced it’s first product line that is generally available. For those unaware, BigSwitch Networks is a VC whose heritage was deeply seeded by researchers and PHD’s from Stanford, UC-Berkeley along with, product managers and executives from traditional networking companies.
BigSwitch’s largest contribution to the community and that is the open source OpenFlow controller FloodLight. Just as Nicira who was recently acquired by VMWare open sourced their vSwitch Open vSwitch, BigSwitch did the same with FloodLight. In the spirit of enabling horizontalization these projects allow researchers, consumers and even competitors to leverage R&D from these startups. We have enough concepts to keep us going for the foreseeable future, it is time to start executing them as more products continue to emerge.
Big Switch Product Suite
The Big Network Controller serves as the network application platform powering two other products Big Tap and Big Virtual Switch. These three elements account for the startup’s new Big Switch Product Suite Big Tap is a monitoring tool while the Big Virtual Switch is being pegged as a virtualization layer that does for the networking space what VMware did for servers. Appenzeller, who serves as his company’s CEO, elaborated in a statement:
“Big Switch Networks ecosystem ensures customers the broadest range of choice in physical and virtual infrastructure and applications, coupled with the industry’s only open network application platform,” Appenzeller said. “Our Big Virtual Switch network virtualization application enables up to 50% more VMs per rack, resulting in as much as $500K/rack in CAPEX savings and $30K/rack/year OPEX savings, based on a 40 server rack.”
The entire Big Switch line-up is based on open networks, but the amount of support the startup managed to garner is even more noteworthy. The company’s ecosystem consists of over two dozen big name vendors including Arista Networks, Broadcom, Brocade, Juniper and Palo Alto Networks as well as Microsoft and Dell. Surprisingly there is no mention of Cisco, which has been slow to adopt SDN in spite of an apparent effort to jump on the bandwagon.
My ANGLE – A Networking Renaissance
We are seeing the beginning of a networking renaissance or a innovation boom in the networking space. This launch represents Big Switch’s “big splash” in the market after Nicira exited the market at $1b to VMware. Big Switch is setup to be “gold plated” venture and this is the first step for them to get a market position. Big Switch hopes to lead the emerging and lucrative software defined networking (SDN) category. Big Switch has sorted out their financing partners and are fully funded with a great management team. Simply put: they are “geared up” and ready to go full tilt.
Big Switch doesn’t have many customers yet, but based upon their recent traction with some big not yet announced enterprise customers they should do great. I’m expecting them to move fast and buildup their product and solutions and get them into the market as fast as possible. In the industry they call this a “greenfield” opportunity. We will be tracking them closely.
The goal of Big Switch is to be open on the platform and differentiate on their unique product suite. The question is “is the networking renaissance” real? Either way the overall the industry valuations are rising as a result.
Here is the CrowdANGLE on Twitter from some of our fav tweets.
John Furrier is founder, co-CEO, and Editor-in-Chief of SiliconANGLE, a new media company covering the intersection of computer science and social science. Furrier is also the co-founder and CEO of CrowdChat a social media platform for large-scale group conversations over hashtags. In addition to SiliconANGLE John runs Broadband Developments a private incubator and investment firm for creating new startups. Furrier lives in Palo Alto, California with his wife and four children.
A companion piece to Wikibon's Public Cloud Market Forecast 2015-2026, this research examines the revenue from SaaS, IaaS and PaaS vendors. The competitive environment surrounding the Public Cloud is in flux. SaaS remains turbulent with new entrants successfully gaining share and incumbent licensed software providers trying to develop SaaS offerings and reclaim leadership positions they have maintained for a decade or more. The IaaS segment leadership is beginning to crystalize as a function of scale, but PaaS is just formulating and finding its way. As such, enterprises need to be wary of which providers are winning and losing (and where), but more importantly what they themselves intend to accomplish with Public Cloud.
Cataloging and automated policy management are the key enablers of a virtual flash world, where storage snapshots are both King and Knave. Combining cataloging and automated policy management is the only solution to enabling storage copy reduction in harmony is risk management and compliance. This enables and justifies an all-flash data center, enables data to be available quicker to the business and other IT functions, and drives greater business and IT productivity and responsiveness. CIOs and senior management should create a small team of the best and brightest, create an optimized all-flash virtual environment with a programmatically integrated catalog in a subset of the datacenter, and demonstrate the practicality and benefits of this environment to the business and IT.
VMworld has grown to be one of the largest and most important technology industry events. Wikibon has attended this event for many years and will have its largest presence this year as part of a double-set of theCUBE. Coverage will examine the broad and diverse ecosystem including storage, cloud, networking and much more.
Wikibon believes latency storage vs. capacity storage is a key storage dimension, with different functional requirements and different cost profiles. Latency storage is found within the datacenter supporting more active applications, and in general has a high read bias. Latencies can vary from 1 millisecond down to a few microseconds; the lower (better) the latency the closer to the processor resources it is likely to be. It is also found for the metadata layer for capacity data. The boundary for latency storage will reduce down to 500 microseconds over the next three years. Capacity storage is found in archive, log, time-series databases for the Internet of things and many other similar applications. In general it is write-heavy. Latencies are generally above 1 millisecond, do not have to so close to the processor, and are suitable for remote private, public and hybrid cloud storage. Some parts of the capacity market place will have latencies as low as 500 microseconds over the next three years. Wikibon has added this dimension to the other storage dimensions projected, which include HDD vs. Flash, Hyperscale Server SAN vs. Enterprise Server SAN vs. Traditional SAN/NAS storage, Physical Capacity vs. Logical Capacity and SaaS Cloud vs. IaaS Cloud vs. PaaS Cloud. All these dimensions are projected for both Revenue and Terabytes. There is a strong correlation and interaction between the latency/capacity dimension and HDD/Flash dimension. Wikibon provides detailed breakdown of the storage projections for Premium clients.
With the EMC Federation in transition, VMware needs to take a greater leadership role in delivering cloud services. VMware's ability to lead their customers transitions to Hybrid Cloud requires them to deliver a more robust, more agile vCloud Air platform.
While the majority of Enterprise CIOs continue to have Hybrid Cloud near the top of their priorities, the solutions in the marketplace often fall short of expectations. Wikibon analysts look at the state of Hybrid Cloud from a C-Suite perspective.
Maturity by industry is a function of customer and prospect skills, competitive drivers within the industry, the nature of the data in question as well as the complexity of the analytics required, and most importantly the problem that the prospect or customer is trying to solve. Identifying where customers are in their journey and helping them to reach the next level is key to vendor success.
Both customers and vendors need to prioritize how they address adoption barriers. As with all emerging technologies, a full solution will often require extensive 3rd party participation, such as “data wrangling” and SQL data access tools. SQL data analysis offerings are maturing rapidly in the areas of application performance under greater user and data volume loads both from the Hadoop distribution vendors as well as 3rd parties. Other barriers, such as a skills gap across many roles, are more intractable. Smaller customers with fewer specialized practitioners in each role should include in their evaluations cloud-based solutions that are fully managed services.
Across all the primary roles involved in adopting Big Data applications, there are basic gaps in product maturity. However, IT leaders and practitioners should keep in mind that Big Data databases are part of a relatively immature ecosystems that requires advanced skills and integration technology in order to operate successfully. The ecosystem is evolving and maturing rapidly and there is a tremendous proliferation of technologies to augment early products.
Wikibon Research has forecasted Public Cloud revenues from 2015 through 2026; with the expectations that 33% of all IT spending ($500B) will move to Public Cloud services within the next decade. The impact this will have on IT vendors and IT strategy will be significant as more applications focus on mobile and real-time data analytics.