[Related tutorial recommendations: nodejs video tutorial]
RPC: Remote Procedure Call refers to remote procedure call, that is to say, there are two servers A and B. An application is deployed on server A and you want to call the application on server B. The functions/methods provided cannot be called directly because they are not in the same memory space. The semantics of the call and the data of the call need to be conveyed through the network.
Communication between servers
Same points
Differences
ajax
uses DNS
as the addressing service to obtain the IP address corresponding to the domain name, browse After getting the IP address, the server sends a request to obtain the data.
RPC
usually requests each other in the intranet, so it generally does not use DNS for addressing services. Because it is on the internal network, you can use a specified ID or a virtual VIP, such as v5:8001
, and then go to the addressing server to obtain the IP address corresponding to v5
.
ajax
Use the http protocol, which is a text protocol. When we interact with data, the file format is either html, Either it is a json object, which is in the form of key-value when using json.
RPC
adopts binary protocol. Using binary transmission, the packet it transmits looks like this [0001 0001 0111 0110 0010], which is all binary. Generally, those digits are used to represent a field. For example, the first 6 digits are a field, and so on.
This way there is no need for http to transmit the key in the json object, so the data volume is smaller.
Because the transmission is binary, it is more suitable for computers to understand, and the text protocol is more suitable for human understanding, so the computer takes much less time to interpret each field than the text protocol.
RPC uses binary to have smaller data volume and faster interpretation speed.
Simplex communication: Only the client can send messages to the server, or only the server can send messages to the server. Client sends messages
Half-duplex communication: Within a certain period of time, only the client can send messages to the server. After this period, the server can send messages to the client. . If time is divided into many time slices, within one time slice it is simplex communication
Full-duplex communication: the client and the server can communicate with each other
The main factors to consider when choosing one of these three communication methods are: implementation difficulty and cost. Full-duplex communication is more expensive than half-duplex communication. In some scenarios, half-duplex communication can still be considered.
ajax
is a half-duplex communication. http is a text protocol, but its bottom layer is a tcp protocol. The http text will undergo a conversion process from binary data flow to text at the tcp layer.
UnderstandingRPC
is just a deeper understanding of front-end technology.
buffer.from: from Some data creates a binary
const buffer1 = Buffer.from('geekbang') const buffer2 = Buffer.from([0, 1, 2, 3, 4]) <Buffer 67 65 65 6b 62 61 6e 67> <Buffer 00 01 02 03 04>
buffer.alloc: Create an empty binary
const buffer3 = Buffer.alloc(20) <Buffer 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00>
buffer.write(string, offset)
: Write a string buffer.writeInt8(value, offset)
: int8 represents 8-bit binary ( 8 bits represent an integer that can be represented by one byte), and offset is the number of bytes to be skipped before starting to write. buffer.writeInt16BE(value, offset)
: int16 (two bytes), which represents an integer that can be represented by 16 binary bits, that is, 32767. The program will report an error if it exceeds this number. const buffer = Buffer.from([1, 2, 3, 4]) // <Buffer 01 02 03 04> // 往第二个字节里面写入12 buffer.writeInt8(12, 1) // <Buffer 01 0c 03 04>
Big-endian BE and little-endian LE: The main reason is that the data arrangement of more than 2 bytes is different (writeInt8 has only one byte, so there is no big-endian and Little endian), in big endian, the low-order address is placed in the high-order bit, and in little endian, the low-order address is placed in the low-order bit. As follows:
const buffer = Buffer.from([1, 2, 3, 4]) buffer.writeInt16BE(512, 2) // <Buffer 01 02 02 00> buffer.writeInt16LE(512, 2) // <Buffer 01 02 00 02>
How does the binary transmitted by PC represent the fields? Now there is a binary package [00, 00, 00, 00, 00, 00, 00]. We assume that the first three bytes represent a field value, the next two bytes represent the value of a field, and the last two bytes also represent the value of a field. value. The writing method is as follows:
writeInt16BE(value, 0) writeInt16BE(value, 2) writeInt16BE(value, 4)
发现像这样写,不仅要知道写入的值,还要知道值的数据类型,这样就很麻烦。不如json格式那么方便。针对这种情况业界也有解决方案。npm有个库protocol-buffers
,把我们写的参数转化为buffer
:
// test.proto 定义的协议文件 message Column { required float num = 1; required string payload = 2; } // index.js const fs = require('fs') var protobuf = require('protocol-buffers') var messages = protobuf(fs.readFileSync('test.proto')) var buf = messages.Column.encode({ num: 42, payload: 'hello world' }) console.log(buf) // <Buffer 0d 00 00 28 42 12 0b 68 65 6c 6c 6f 20 77 6f 72 6c 64> var obj = messages.Column.decode(buf) console.log(obj) // { num: 42, payload: 'hello world' }
服务端代码:
const net = require('net') const LESSON_DATA = { 136797: '01 | 课程介绍', 136798: '02 | 内容综述', 136799: '03 | Node.js是什么?', 136800: '04 | Node.js可以用来做什么?', 136801: '05 | 课程实战项目介绍', 136803: '06 | 什么是技术预研?', 136804: '07 | Node.js开发环境安装', 136806: '08 | 第一个Node.js程序:石头剪刀布游戏', 136807: '09 | 模块:CommonJS规范', 136808: '10 | 模块:使用模块规范改造石头剪刀布游戏', 136809: '11 | 模块:npm', 141994: '12 | 模块:Node.js内置模块', 143517: '13 | 异步:非阻塞I/O', 143557: '14 | 异步:异步编程之callback', 143564: '15 | 异步:事件循环', 143644: '16 | 异步:异步编程之Promise', 146470: '17 | 异步:异步编程之async/await', 146569: '18 | HTTP:什么是HTTP服务器?', 146582: '19 | HTTP:简单实现一个HTTP服务器' } const server = net.createServer(socket => { // 监听客户端发送的消息 socket.on('data', buffer => { const lessonId = buffer.readInt32BE() setTimeout(() => { // 往客户端发送消息 socket.write(LESSON_DATA[lessonId]) }, 1000) }) }) server.listen(4000)
客户端代码:
const net = require('net') const socket = new net.Socket({}) const LESSON_IDS = [ '136797', '136798', '136799', '136800', '136801', '136803', '136804', '136806', '136807', '136808', '136809', '141994', '143517', '143557', '143564', '143644', '146470', '146569', '146582' ] socket.connect({ host: '127.0.0.1', port: 4000 }) let buffer = Buffer.alloc(4) buffer.writeInt32BE(LESSON_IDS[Math.floor(Math.random() * LESSON_IDS.length)]) // 往服务端发送消息 socket.write(buffer) // 监听从服务端传回的消息 socket.on('data', buffer => { console.log(buffer.toString()) // 获取到数据之后再次发送消息 buffer = Buffer.alloc(4) buffer.writeInt32BE(LESSON_IDS[Math.floor(Math.random() * LESSON_IDS.length)]) socket.write(buffer) })
以上半双工通信步骤如下:
这样在一个时间端之内,只有一个端往另一个端发送消息,这样就实现了半双工通信。那如何实现全双工通信呢,也就是在客户端往服务端发送消息的同时,服务端还没有消息返回给客户端之前,客户端又发送了一个消息给服务端。
先来看一个场景:
客户端发送了一个id1的请求,但是服务端还来不及返回,接着客户端又发送了一个id2的请求。
等了一个之后,服务端先把id2的结果返回了,然后再把id1的结果返回。
那如何结果匹配到对应的请求上呢?
如果按照时间顺序,那么id1的请求对应了id2的结果,因为id2是先返回的;id2的请求对应了id1的结果,这样就导致请求包和返回包错位的情况。
怎么办呢?
我们可以给请求包和返回包都带上序号,这样就能对应上。
客户端代码:
socket.on('data', buffer => { // 包序号 const seqBuffer = buffer.slice(0, 2) // 服务端返回的内容 const titleBuffer = buffer.slice(2) console.log(seqBuffer.readInt16BE(), titleBuffer.toString()) }) // 包序号 let seq = 0 function encode(index) { // 请求包的长度现在是6 = 2(包序号) + 4(课程id) buffer = Buffer.alloc(6) buffer.writeInt16BE(seq) buffer.writeInt32BE(LESSON_IDS[index], 2) seq++ return buffer } // 每50ms发送一次请求 setInterval(() => { id = Math.floor(Math.random() * LESSON_IDS.length) socket.write(encode(id)) }, 50)
服务端代码:
const server = net.createServer(socket => { socket.on('data', buffer => { // 把包序号取出 const seqBuffer = buffer.slice(0, 2) // 从第2个字节开始读取 const lessonId = buffer.readInt32BE(2) setTimeout(() => { const buffer = Buffer.concat([ seqBuffer, Buffer.from(LESSON_DATA[lessonId]) ]) socket.write(buffer) // 这里返回时间采用随机的,这样就不会按顺序返回,就可以测试错位的情况 }, 10 + Math.random() * 1000) }) })
如果我们这样发送请求:
for (let i = 0; i < 100; i++) { id = Math.floor(Math.random() * LESSON_IDS.length) socket.write(encode(id)) }
我们发现服务端接收到的信息如下:
<Buffer 00 00 00 02 16 64 00 01 00 02 16 68 00 02 00 02 31 1c 00 03 00 02 3c 96 00 04 00 02 16 68 00 05 00 02 16 5e 00 06 00 02 16 66 00 07 00 02 16 67 00 08 ... 550 more bytes>
这是因为TCP
自己做的一个优化,它会把所有的请求包拼接在一起,这样就会产生粘包的现象。
服务端需要把包进行拆分,拆分成100个小包。
那如何拆分呢?
首先客户端发送的数据包包括两部分:定长的包头和不定长的包体。
包头又分为两部分:包序号及包体的长度。只有知道包体的长度,才能知道从哪里进行分割。
let seq = 0 function encode(data) { // 正常情况下,这里应该是使用 protocol-buffers 来encode一段代表业务数据的数据包 // 为了不要混淆重点,这个例子比较简单,就直接把课程id转buffer发送 const body = Buffer.alloc(4); body.writeInt32BE(LESSON_IDS[data.id]); // 一般来说,一个rpc调用的数据包会分为定长的包头和不定长的包体两部分 // 包头的作用就是用来记载包的序号和包的长度,以实现全双工通信 const header = Buffer.alloc(6); // 包序号占2个字节,包体长度占4个字节,共6个字节 header.writeInt16BE(seq) header.writeInt32BE(body.length, 2); // 包头和包体拼起来发送 const buffer = Buffer.concat([header, body]) console.log(`包${seq}传输的课程id为${LESSON_IDS[data.id]}`); seq++; return buffer; } // 并发 for (let i = 0; i < 100; i++) { id = Math.floor(Math.random() * LESSON_IDS.length) socket.write(encode({ id })) }
服务端进行拆包
const server = net.createServer(socket => { let oldBuffer = null socket.on('data', buffer => { // 把上一次data事件使用残余的buffer接上来 if (oldBuffer) { buffer = Buffer.concat([oldBuffer, buffer]) } let packageLength = 0 // 只要还存在可以解成完整包的包长 while ((packageLength = checkComplete(buffer))) { // 确定包的长度后进行slice分割 const package = buffer.slice(0, packageLength) // 剩余的包利用循环继续分割 buffer = buffer.slice(packageLength) // 把这个包解成数据和seq const result = decode(package) // 计算得到要返回的结果,并write返回 socket.write(encode(LESSON_DATA[result.data], result.seq)) } // 把残余的buffer记下来 oldBuffer = buffer }) })
checkComplete
函数的作用来确定一个数据包的长度,然后进行分割:
function checkComplete(buffer) { // 如果包的长度小于6个字节说明只有包头,没有包体,那么直接返回0 if (buffer.length <= 6) { return 0 } // 读取包头的第二个字节,取出包体的长度 const bodyLength = buffer.readInt32BE(2) // 请求包包括包头(6个字节)和包体body return 6 + bodyLength }
decode
对包进行解密:
function decode(buffer) { // 读取包头 const header = buffer.slice(0, 6) const seq = header.readInt16BE() // 读取包体 // 正常情况下,这里应该是使用 protobuf 来decode一段代表业务数据的数据包 // 为了不要混淆重点,这个例子比较简单,就直接读一个Int32即可 const body = buffer.slice(6).readInt32BE() // 这里把seq和数据返回出去 return { seq, data: body } }
encode
把客户端想要的数据转化为二进制返回,这个包同样包括包头和包体,包头又包括包需要包序号和包体的长度。
function encode(data, seq) { // 正常情况下,这里应该是使用 protobuf 来encode一段代表业务数据的数据包 // 为了不要混淆重点,这个例子比较简单,就直接把课程标题转buffer返回 const body = Buffer.from(data) // 一般来说,一个rpc调用的数据包会分为定长的包头和不定长的包体两部分 // 包头的作用就是用来记载包的序号和包的长度,以实现全双工通信 const header = Buffer.alloc(6) header.writeInt16BE(seq) header.writeInt32BE(body.length, 2) const buffer = Buffer.concat([header, body]) return buffer }
当客户端收到服务端发送的包之后,同样也要进行拆包,因为所有的包同样都粘在一起了:
<Buffer 00 00 00 00 00 1d 30 36 20 7c 20 e4 bb 80 e4 b9 88 e6 98 af e6 8a 80 e6 9c af e9 a2 84 e7 a0 94 ef bc 9f 00 01 00 00 00 1d 30 36 20 7c 20 e4 bb 80 e4 ... 539 more bytes>
因此,客户端也需要拆包,拆包策略与服务端的拆包策略是一致的:
let oldBuffer = null socket.on('data', buffer => { // 把上一次data事件使用残余的buffer接上来 if (oldBuffer) { buffer = Buffer.concat([oldBuffer, buffer]) } let completeLength = 0 // 只要还存在可以解成完整包的包长 while ((completeLength = checkComplete(buffer))) { const package = buffer.slice(0, completeLength) buffer = buffer.slice(completeLength) // 把这个包解成数据和seq const result = decode(package) console.log(`包${result.seq},返回值是${result.data}`) } // 把残余的buffer记下来 oldBuffer = buffer })
到这里就实现了双全工通行,这样客户端和服务端随时都可以往对方发小消息了。
更多node相关知识,请访问:nodejs 教程!
The above is the detailed content of What is RPC? Let's talk about how to implement RPC communication in node. For more information, please follow other related articles on the PHP Chinese website!