Jaco Toledo, I have experience with HTML, CSS, PHP and MYSQL1.2k Views
Don’t think nginx will support quic anytime soon.If you’re interested in playing with QUIC today you’ll need to build the test QUIC server and client that are part of the Chromium project, get Google Chrome Canary and enable QUIC in chrome://flags.
Written 8 Feb
Zhifeng Mi1.1k Views
nginx is working on TCP level 7
and quic is based on UDP
so I don’t see any chance nginx will support quic.
And actually, you just need to develop a udp to tcp proxy to pass the data to nginx.
Now in Chrome if you enable quic, Chrome will start a proxy to send data to the
web server so if you don’t want to modify the whole architect in your server side.
develop a upd to tcp proxy to support quic will be enough.
Written 27 Dec 2015 · View Upvotes
One of the hottest topics for discussion today is the new HTTP/2 that is the most substantial protocol update since 1999, the year when the HTTP/1.1 version was released.
Among the key improvements brought by HTTP/2 are multiplexed streams, header compression, server push, and a binary protocol instead of textual one. These and other positive changes allowed to achieve good web pages loading results, including those having lots of additional files attached to them (e.g. styles, scripts, images, fonts, etc.). But what effect is this going to have upon REST APIs and their work?
The first thing to talk about is the format of data transmission. As it has been mentioned above, a binary format has substituted a textual one which is good news for servers but not for developers. The latter one is convenient for program debugging but the binary one is still a good innovation, as it is much easier for a computer to process such data. Consequently, the binary format will contribute to lightening some of the server processor load, which would be helpful for heavily loaded REST APIs.
The protocol is compatible with the HTTP/1.1 version. Status codes, methods and headers did not overcome any changes. The only thing that did change is the way of transmitting data via TCP connection.
HTTP/2 is much more effective when it comes to processing a large number of requests. One of the biggest HTTP/1.1 problems is that separate requests are expensive. The reason for this is that many TCP connections are created, and these are established due to the “3-way handshake” method. Then some time has to pass for “slow start” algorithm to make the connection effectively fast. HTTP/2 is devoid of such a drawback thanks to multiplexed streams. This allows to make several HTTP requests as if they were made concurrently. Therefore, all the styles, scripts and images will load faster and users will certainly notice the speed increase. Although in REST APIs there is no need to load a few files simultaneously, multiplexing will let to perform a number of concurrent server requests more effectively due to one TCP connection being used for all these requests. This will also lighten the server load, as there will be no necessity to allocate resources to maintain more than one TCP connection, which will save some memory though the work of processor will be a little bit more intensive.
One more improvement that came with HTTP/2 is header compression. API requests usually have much less headers than ordinary web pages. That is why the benefits it may bring are most likely to be minor.
“Server push” technology made it possible for server to initiate data transmission. Here is an example. A client requests a web page. The web server processes the request and, as a result, gets an idea of what resources the client needs additionally and thus can send them right after the response with the document. This way, the client does not need to send additional request to get the additional resources mentioned above. Unfortunately, it is difficult to find a RESTful API service where this technology can be truly useful.
Though many changes brought with HTTP/2 were created to speed web browsing up, some of themcan be used in the world of REST APIs as well. It is possible to achieve greater efficiency at a lower cost if the HTTP/2 strong points are taken into consideration when designing an application.
- 二进制的数据传输格式。对于负载较重的REST API服务比较有用。
- 多路复用流。对于一次性发送大量请求的Web加载比较有意义。对于REST API服务，依然可以带来一定的好处，降低服务器负载，节省内存。
- 首部压缩。对于REST API服务而言，通常首部的数据量都不是特别大，此项特性带来的好处并不大。
- 服务器推送对REST API服务的意义并不大。
Transitioning from SPDY to HTTP/2
Thursday, February 11, 2016
Last year we announced our intent to end support for the experimental protocol SPDY in favor of the standardized version, HTTP/2. HTTP/2 is the next-generation protocol for transferring information on the web, improving upon HTTP/1.1 with more features leading to better performance. Since then we’ve seen huge adoption of HTTP/2 from both web servers and browsers, with most now supporting HTTP/2. Over 25% of resources in Chrome are currently served over HTTP/2, compared to less than 5% over SPDY. Based on such strong adoption, starting on May 15th — the anniversary of theHTTP/2 RFC — Chrome will no longer support SPDY. Servers that do not support HTTP/2 by that time will serve Chrome requests over HTTP/1.1, providing the exact same features to users without the enhanced performance of HTTP/2.
At the same time, Chrome will stop supporting the TLS protocol extension NPN, which allows servers to negotiate SPDY and HTTP/2 connections with clients. NPN has been superseded by the TLS extension ALPN, published by the IETF in 2014. ALPN is already used 99% of the time to negotiate HTTP/2 with Chrome, and the remaining servers can gain ALPN support by upgrading their SSL library.
We are looking forward to HTTP/2 continuing to gain adoption, bringing us an even faster web.
Update: To better align with Chrome’s release cycle, SPDY and NPN support will be removed with the release of Chrome 51.
Posted by Bence Béky, Network Protocol Engineer and HTTP/2 Enthusiast