Apache ShenYu
The gateway receives requests from the client, forwards them to the server, and returns the server results to the client. The gateway can record the details of each request,
The list includes: request time, request parameters, request path, response result, response status code, time consumption, upstream IP, exception information waiting.
The Logging-Kafka plugin is a plugin that records access logs and sends them to the Kafka cluster.
Core Module shenyu-plugin-logging-kafka
.
Core Class org.apache.shenyu.plugin.logging.kafka.LoggingKafkaPlugin
Core Claas org.apache.shenyu.plugin.logging.kafka.client.KafkaLogCollectClient
Full asynchronous collection and delivery of Logging
inside the Apache ShenYu
gateway
Logging platform by consuming the logs in the Kafka
cluster for repository, and then using Grafana
, Kibana
or other visualization platform to display
Logging-Kafka
dependency to the gateway's pom.xml
file.<!--shenyu logging-kafka plugin start--> <dependency> <groupId>org.apache.shenyu</groupId> <artifactId>shenyu-spring-boot-starter-plugin-logging-kafka</artifactId> <version>${project.version}</version> </dependency> <!--shenyu logging-kafka plugin end-->
shenyu-admin
--> Basic Configuration --> Plugin Management --> loggingKafka
, configure the kafka parameter and set it to on.config-item | type | description | remarks |
topic | String | Message Queue Topics | must |
namesrvAddr | String | Message queue nameserver address | must |
sampleRate | String | Sampling rate, range 0~1, 0: off, 0.01: acquisition 1%, 1: acquisition 100% | Optional, default 1, all collection |
compressAlg | String | Compression algorithm, no compression by default, currently supports LZ4 compression | Optional, no compression by default |
maxResponseBody | Ingeter | Maximum response size, above the threshold no response will be collected | Optional, default 512KB |
maxRequestBody | Ingeter | Maximum request body size, above the threshold no request body will be collected | Optional, default 512KB |
Except for topic, namesrvAddr, all others are optional, in most cases only these 3 items need to be configured. The default group id is “shenyu-access-logging” |
In addition sometimes a large gateway cluster corresponds to multiple applications (business), different applications (business) corresponds to different topics, related to isolation,
then you can configure different topics (optional) and sampling rate (optional) by selector, the meaning of the configuration items as shown in the table above.
The operation is shown below:
collect request info as follows
Field Name | Meaning | Description | Remarks |
---|---|---|---|
clientIp | Client IP | ||
timeLocal | Request time string, format: yyyy-MM-dd HH:mm:ss.SSS | ||
method | request method (different rpc type is not the same, http class for: get, post wait, rpc class for the interface name) | ||
requestHeader | Request header (json format) | ||
responseHeader | Response header (json format) | ||
queryParams | Request query parameters | ||
requestBody | Request Body (body of binary type will not be captured) | ||
requestUri | Request uri | ||
responseBody | Response body | ||
responseContentLength | Response body size | ||
rpcType | rpc type | ||
status | response status | ||
upstreamIp | Upstream (program providing the service) IP | ||
upstreamResponseTime | Time taken by the upstream (program providing the service) to respond to the request (ms ms) | ||
userAgent | Requested user agent | ||
host | The requested host | ||
module | Requested modules | ||
path | The requested path | ||
traceId | Requested Link Tracking ID | Need to access link tracking plugins, such as skywalking,zipkin |
Open the plugin and configure kafka, configure it as follows.
For detailed configuration of selectors and rules, please refer to: Selector and rule management。
In addition sometimes a large gateway cluster corresponds to multiple applications (business), different applications (business) corresponds to different topics, related to isolation,
then you can configure different topics (optional) and samplingf rate (optional) by selector, the meaning of the configuration items as shown in the table above.
The operation is shown below:
As each logging platform has differences, such as storage available clickhouse, ElasticSearch, etc., visualization has self-developed or open source Grafana, Kibana, etc..
Logging-Kafka plugin uses Kafka to decouple production and consumption, while outputting logs in json format, consumption and visualization require users to choose different technology stacks to achieve their own situation.=
shenyu-admin
--> BasicConfig --> Plugin --> loggingKafka
set Status disable.