Click here to close now.

Welcome!

Oracle Authors: Carmen Gonzalez, Michael Jannery, Jnan Dash, Liz McMillan, Glenn Rossman

Related Topics: Web 2.0

Web 2.0: Blog Feed Post

One Line of Code That Changed the Web Forever

What impact will the emerging WebSocket standard have on the Web?

After presenting to a partner of Kaazing last week I got asked what impact the emerging WebSocket standard would have on the Web, assuming we continue down the path that has already been laid out.

The impact could be the same, or even more profound, as when we were first introduced to HTTP as a means to share static documents. The difference is that this time the targeted market is already defined – it is called the Web. I have over the past several years, half jokingly and half seriously, compared the current static Web with a push to talk radio (aka Walkie Talkie) and the new living Web with a cell phone. You can get by with the WT and solve most of your problems; after all it’s been around for a while and it works. If you want to communicate with a friend in “real-time” you can solve it by getting two WTs, one to talk and one to listen. With a new Web standard, WebSocket, entering the market, Web developers now have access to the equivalent of a cellphone – one channel for “talk” and “listen”.

What would you choose if both push-to-talk and cell phone were  available to you? What would developers choose if both technologies were readily available to them (e.g.: browser support)?

Well, if you are uncertain and feel like WebSocket is an unproven standard you might want to relate to this; remember the first time your friends started pushing you to buy a cellphone although you already had a stationary phone at home and one at the office that worked perfectly? Now several years later we all have at least one cell phone, each, and we can’t (at least I can’t) live without it.

This is exactly the same impact the following line of code will have on the Web in comparison with the current HTTP communication we are so used and accustomed too.

var mySocket = new WebSocket("ws://websocket.org/");

If you do understand the profound impact this one line of code will have you are in good shape and most likely are already using, extending, or pushing this new standard solution from W3C and IETF. If you are not, then let me take a short moment to explain why it is so important:

  1. HTTP was designed to deliver static documents, not to deliver transactional, dynamic, and real-time data updates.
  2. HTTP is by design stateless, so session state needs to be artificially maintained. Traditionally this is done by a legacy Web-tier solution such as an application server like Oracle WebLogic Server or IBM WebSpere.
  3. In every environment developers have access to a “socket” interface, which enables them to communicate using any format (read protocol) over a full-duplex connection. Not on the Web.
  4. Not having access to a standard, Web-friendly, socket API forces us to create transformation layers when sending data from a Web client, using HTTP, to a backend system relying on a different full-duplex TCP protocol e.g. XMPP, STOMP, AMQP.
  5. The above line of code opens the floodgates to use any TCP-based communication format, which in turn enables developers to freely innovate and create new types of Web applications that previously have not been feasible over the existing HTTP infrastructure.
  6. WebSocket offers a far better use of bandwidth by getting rid of unnecessary HTTP headers when information is shared. The improvement is at a ratio of up to 1000x.
  7. The latency to deliver data is greatly improved by eliminating the round trip of the HTTP request-response model, and by using the bandwidth more efficiently.

With the explosive growth of Web-enabled devices (yes, I’m thinking about the iPhone, iPad, Android, Galaxy, etc…) and the demand for more and live information, communication and distribution of data over the Web is growing exponentially. At this rate the growth of data distributed over the Web will out pace the performance principals of Moore’s Law, which we depend on to ensure that our hardware can keep up with our needs.

WebSocket traffic vs. HTTP traffic

For individuals this may not be too much of a concern, but for companies providing online services it will be, and already is, a huge and costly issue since it requires a tremendous amount of resources to deliver on the increasing demand for live information over the Web (read about Google’s move and Facebook’s move).

For example, when a user enters a single character ‘a’ in a search engine, a drop down list appears automatically showing possible search results starting with letter ‘a’. Behind the scenes an HTTP request has been issued asking the server for the information displayed in the drop down list. For every new character entered a new HTTP request is issued to the server to request for more information. The same HTTP characteristics you can find in collaborative online documents such as Google Docs, where each character entered generates a POST to ensure that users editing or looking at the same document can see each other changes in real-time.

Now, what was sent, what was received, and what was really needed?

There is a great article on websocket.org, called a “Quantum Leap in Scalability for the Web” that is outlining the difference between HTTP and WebSocket in terms of bandwidth utilization. In this article the sample application is a simple trading solution, but the math can be applied to any HTTP-based dynamic and transactional Web application.

In the article we have 0,665Gbps in header traffic to respond to 100,000 users per request.

What is the impact of using WebSocket technology? There are no sizable headers involved passing information between a client and a Websocket Gateway. Let’s apply the above math example to WebSocket technology as described by the article:

100,000 visitors receiving an update every second.  (WS wireframe = 2 byte) * 100,000 * 8 = 1,600,000 bps (0.001526Gbps).

Results from this easy math:

HTTP:// = 0,665Gbps versus WS:// = 0.001526Gbps.  In the above sample Websocket communication is 436 times more efficient. 436 times! We are talking about a gigantic leap of improvement, and that assuming that your cookies are not adding more data than this sample.

WebSocket is not a better Ajax!

Not only is the new standard improving bandwidth utilization it also gives us the ability to use any TCP-based high level communication format for our Web applications. This part of the HTML5 WebSocket standard has still yet to be fully appreciated. Right now most solutions and developers tinkering with the WebSocket APIs are looking at the new standard as merely a better replacement of XHR, or Ajax, when in fact it is a quantum leap forward in communicating over the Web that cannot be compared to XHR. With WebSocket we can now build client libraries in any Web technology supporting any TCP-based protocols. A simple example would be to extend the now widely used chat protocol XMPP to the Web (here is a demo site that lets you log in to Google Talk using XMPP over WebSocket) by providing a client-side implementation on top of Websocket APIs, or an advanced example would be to extend Java Message Service (JMS) over WebSocket such as the Kaazing WebSocket Gateway.

Scaling a WebSocket Solution

Web developers have been trying to work around the limitations of HTTP since the early days using techniques such as Comet, Reverse Ajax, or HTTP Streaming. With a move to persistent connections, or a stateful Web, server scalability of concurrent connections has been, and still is, a serious concern. Holding on to a thread on the server while the thread is not in use, combined with an Web-tier and infrastructure that was not designed for this, is not necessarily a scalable combination. Now, great strides have been made to ensure better scalability across technology stacks such as the use of NIO in Java.

At Kaazing we have always taken scalability and performance extremely serious and focused on making sure that our software is not in the way of scale or performance. As a matter of fact, we did a benchmark over the new year 2008/2009, to prove that scaling a WebSocket solution with persistent connections was not an issue, so we brought in a Java performance expert – Kirk Pepperdine – to help us and by the first weeks of Jan 09 we were running 1,000,000 concurrent connections on one single server. Now, is this practical? A more realistic scenario is running 1,000,000 users on a single rack or half a rack. This would enable us to have failover and high-availability, while still providing great performance and scale. So, last year we ran new tests together with DELL and Tibco to ensure not only great scale but also outstanding performance: DELL, Tibco, and Kaazing enable ‘The Fastest Million’ to revolutionize real-time data delivery over the Web.

In Conclusion

The simplest design ideas are often the innovations with the most impact. WebSocket as an idea and design is extremely “simple” and its impact on our industry will be profound. Of course, with simple ideas you also get the “doubters”. I remember one time when my co-founder John Fallows and I met with a renowned VC in Silicon Valley and he asked us:

“If this is such a great idea why has no one come up with this idea before?”

I guess you could ask humanity a similar question about why it took several thousands of years to invent the wheel – after all it’s so obvious and simple.

What is important to understand is that we now have at our disposal a very powerful tool that will enable us to communicate securely with anything over the Web, and that it is only our own imagination that will limit our ability to fully exploit the WebSocket standard to its full potential.

If you are having performance and scalability issues with your current Web solution then it is time to look at an enterprise WebSocket platform, such as the one Kaazing provides. To round off I’m just going to ask you one short question:

If you had a choice between building a Web application using HTTP and Websocket, and both were readily available to you, which one would you choose?

Read the original blog entry...

More Stories By Jonas Jacobi

Jonas has 21 years of experience leading the development of innovative technology products and services. Together with Kaazing’s Co-Founder & CTO John Fallows, he pioneered and championed the groundbreaking HTML5 WebSocket standard. Prior to co-founding Kaazing he served as VP of Product Management for Brane Corporation, a Silicon Valley startup dedicated to developing a market-leading enterprise platform for building model-driven apps. Before Brane, he spent 8+ years at Oracle where he served as a Java EE and open source Evangelist, and was Product Manager in the Oracle Application Server division for JavaServer Faces, Oracle ADF Faces, and Oracle ADF Faces Rich Client. He is a frequent speaker at international conferences on accelerating and scaling secure enterprise-grade WebComms (Web Communications).

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
The Workspace-as-a-Service (WaaS) market will grow to $6.4B by 2018. In his session at 16th Cloud Expo, Seth Bostock, CEO of IndependenceIT, will begin by walking the audience through the evolution of Workspace as-a-Service, where it is now vs. where it going. To look beyond the desktop we must understand exactly what WaaS is, who the users are, and where it is going in the future. IT departments, ISVs and service providers must look to workflow and automation capabilities to adapt to growing demand and the rapidly changing workspace model.
Cloud data governance was previously an avoided function when cloud deployments were relatively small. With the rapid adoption in public cloud – both rogue and sanctioned, it’s not uncommon to find regulated data dumped into public cloud and unprotected. This is why enterprises and cloud providers alike need to embrace a cloud data governance function and map policies, processes and technology controls accordingly. In her session at 15th Cloud Expo, Evelyn de Souza, Data Privacy and Compliance Strategy Leader at Cisco Systems, will focus on how to set up a cloud data governance program and s...
As organizations shift toward IT-as-a-service models, the need for managing and protecting data residing across physical, virtual, and now cloud environments grows with it. CommVault can ensure protection &E-Discovery of your data – whether in a private cloud, a Service Provider delivered public cloud, or a hybrid cloud environment – across the heterogeneous enterprise. In his session at 16th Cloud Expo, Randy De Meno, Chief Technologist - Windows Products and Microsoft Partnerships, will discuss how to cut costs, scale easily, and unleash insight with CommVault Simpana software, the only si...
Hadoop as a Service (as offered by handful of niche vendors now) is a cloud computing solution that makes medium and large-scale data processing accessible, easy, fast and inexpensive. In his session at Big Data Expo, Kumar Ramamurthy, Vice President and Chief Technologist, EIM & Big Data, at Virtusa, will discuss how this is achieved by eliminating the operational challenges of running Hadoop, so one can focus on business growth. The fragmented Hadoop distribution world and various PaaS solutions that provide a Hadoop flavor either make choices for customers very flexible in the name of opti...
Containers and microservices have become topics of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 16th Cloud Expo at the Javits Center in New York June 9-11 will find fresh new content in a new track called PaaS | Containers & Microservices Containers are not being considered for the first time by the cloud community, but a current era of re-consideration has pushed them to the top of the cloud agenda. With the launch of Docker's initial release in March of 2013, interest was revved up several notches. Then late last...
Roberto Medrano, Executive Vice President at SOA Software, had reached 30,000 page views on his home page - http://RobertoMedrano.SYS-CON.com/ - on the SYS-CON family of online magazines, which includes Cloud Computing Journal, Internet of Things Journal, Big Data Journal, and SOA World Magazine. He is a recognized executive in the information technology fields of SOA, internet security, governance, and compliance. He has extensive experience with both start-ups and large companies, having been involved at the beginning of four IT industries: EDA, Open Systems, Computer Security and now SOA.
HP and Aruba Networks on Monday announced a definitive agreement for HP to acquire Aruba, a provider of next-generation network access solutions for the mobile enterprise, for $24.67 per share in cash. The equity value of the transaction is approximately $3.0 billion, and net of cash and debt approximately $2.7 billion. Both companies' boards of directors have approved the deal. "Enterprises are facing a mobile-first world and are looking for solutions that help them transition legacy investments to the new style of IT," said Meg Whitman, Chairman, President and Chief Executive Officer of HP...
The industrial software market has treated data with the mentality of “collect everything now, worry about how to use it later.” We now find ourselves buried in data, with the pervasive connectivity of the (Industrial) Internet of Things only piling on more numbers. There’s too much data and not enough information. In his session at @ThingsExpo, Bob Gates, Global Marketing Director, GE’s Intelligent Platforms business, to discuss how realizing the power of IoT, software developers are now focused on understanding how industrial data can create intelligence for industrial operations. Imagine ...
Operational Hadoop and the Lambda Architecture for Streaming Data Apache Hadoop is emerging as a distributed platform for handling large and fast incoming streams of data. Predictive maintenance, supply chain optimization, and Internet-of-Things analysis are examples where Hadoop provides the scalable storage, processing, and analytics platform to gain meaningful insights from granular data that is typically only valuable from a large-scale, aggregate view. One architecture useful for capturing and analyzing streaming data is the Lambda Architecture, representing a model of how to analyze rea...
SYS-CON Events announced today that Vitria Technology, Inc. will exhibit at SYS-CON’s @ThingsExpo, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Vitria will showcase the company’s new IoT Analytics Platform through live demonstrations at booth #330. Vitria’s IoT Analytics Platform, fully integrated and powered by an operational intelligence engine, enables customers to rapidly build and operationalize advanced analytics to deliver timely business outcomes for use cases across the industrial, enterprise, and consumer segments.
SYS-CON Events announced today that Open Data Centers (ODC), a carrier-neutral colocation provider, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Open Data Centers is a carrier-neutral data center operator in New Jersey and New York City offering alternative connectivity options for carriers, service providers and enterprise customers.
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
The explosion of connected devices / sensors is creating an ever-expanding set of new and valuable data. In parallel the emerging capability of Big Data technologies to store, access, analyze, and react to this data is producing changes in business models under the umbrella of the Internet of Things (IoT). In particular within the Insurance industry, IoT appears positioned to enable deep changes by altering relationships between insurers, distributors, and the insured. In his session at @ThingsExpo, Michael Sick, a Senior Manager and Big Data Architect within Ernst and Young's Financial Servi...
PubNub on Monday has announced that it is partnering with IBM to bring its sophisticated real-time data streaming and messaging capabilities to Bluemix, IBM’s cloud development platform. “Today’s app and connected devices require an always-on connection, but building a secure, scalable solution from the ground up is time consuming, resource intensive, and error-prone,” said Todd Greene, CEO of PubNub. “PubNub enables web, mobile and IoT developers building apps on IBM Bluemix to quickly add scalable realtime functionality with minimal effort and cost.”
Sensor-enabled things are becoming more commonplace, precursors to a larger and more complex framework that most consider the ultimate promise of the IoT: things connecting, interacting, sharing, storing, and over time perhaps learning and predicting based on habits, behaviors, location, preferences, purchases and more. In his session at @ThingsExpo, Tom Wesselman, Director of Communications Ecosystem Architecture at Plantronics, will examine the still nascent IoT as it is coalescing, including what it is today, what it might ultimately be, the role of wearable tech, and technology gaps stil...
With several hundred implementations of IoT-enabled solutions in the past 12 months alone, this session will focus on experience over the art of the possible. Many can only imagine the most advanced telematics platform ever deployed, supporting millions of customers, producing tens of thousands events or GBs per trip, and hundreds of TBs per month. With the ability to support a billion sensor events per second, over 30PB of warm data for analytics, and hundreds of PBs for an data analytics archive, in his session at @ThingsExpo, Jim Kaskade, Vice President and General Manager, Big Data & Ana...
In the consumer IoT, everything is new, and the IT world of bits and bytes holds sway. But industrial and commercial realms encompass operational technology (OT) that has been around for 25 or 50 years. This grittier, pre-IP, more hands-on world has much to gain from Industrial IoT (IIoT) applications and principles. But adding sensors and wireless connectivity won’t work in environments that demand unwavering reliability and performance. In his session at @ThingsExpo, Ron Sege, CEO of Echelon, will discuss how as enterprise IT embraces other IoT-related technology trends, enterprises with i...
When it comes to the Internet of Things, hooking up will get you only so far. If you want customers to commit, you need to go beyond simply connecting products. You need to use the devices themselves to transform how you engage with every customer and how you manage the entire product lifecycle. In his session at @ThingsExpo, Sean Lorenz, Technical Product Manager for Xively at LogMeIn, will show how “product relationship management” can help you leverage your connected devices and the data they generate about customer usage and product performance to deliver extremely compelling and reliabl...
The Internet of Things (IoT) is causing data centers to become radically decentralized and atomized within a new paradigm known as “fog computing.” To support IoT applications, such as connected cars and smart grids, data centers' core functions will be decentralized out to the network's edges and endpoints (aka “fogs”). As this trend takes hold, Big Data analytics platforms will focus on high-volume log analysis (aka “logs”) and rely heavily on cognitive-computing algorithms (aka “cogs”) to make sense of it all.
One of the biggest impacts of the Internet of Things is and will continue to be on data; specifically data volume, management and usage. Companies are scrambling to adapt to this new and unpredictable data reality with legacy infrastructure that cannot handle the speed and volume of data. In his session at @ThingsExpo, Don DeLoach, CEO and president of Infobright, will discuss how companies need to rethink their data infrastructure to participate in the IoT, including: Data storage: Understanding the kinds of data: structured, unstructured, big/small? Analytics: What kinds and how responsiv...