kafka-logger
#
DescriptionThe kafka-logger
Plugin is used to push logs as JSON objects to Apache Kafka clusters. It works as a Kafka client driver for the ngx_lua Nginx module.
It might take some time to receive the log data. It will be automatically sent after the timer function in the batch processor expires.
#
AttributesName | Type | Required | Default | Valid values | Description |
---|---|---|---|---|---|
broker_list | object | True | Deprecated, use brokers instead. List of Kafka brokers. (nodes). | ||
brokers | array | True | List of Kafka brokers (nodes). | ||
brokers.host | string | True | The host of Kafka broker, e.g, 192.168.1.1 . | ||
brokers.port | integer | True | [0, 65535] | The port of Kafka broker | |
brokers.sasl_config | object | False | The sasl config of Kafka broker | ||
brokers.sasl_config.mechanism | string | False | "PLAIN" | ["PLAIN"] | The mechaism of sasl config |
brokers.sasl_config.user | string | True | The user of sasl_config. If sasl_config exists, it's required. | ||
brokers.sasl_config.password | string | True | The password of sasl_config. If sasl_config exists, it's required. | ||
kafka_topic | string | True | Target topic to push the logs for organisation. | ||
producer_type | string | False | async | ["async", "sync"] | Message sending mode of the producer. |
required_acks | integer | False | 1 | [1, -1] | Number of acknowledgements the leader needs to receive for the producer to consider the request complete. This controls the durability of the sent records. The attribute follows the same configuration as the Kafka acks attribute. required_acks cannot be 0. See Apache Kafka documentation for more. |
key | string | False | Key used for allocating partitions for messages. | ||
timeout | integer | False | 3 | [1,...] | Timeout for the upstream to send data. |
name | string | False | "kafka logger" | Unique identifier for the batch processor. If you use Prometheus to monitor APISIX metrics, the name is exported in apisix_batch_process_entries . | |
meta_format | enum | False | "default" | ["default","origin"] | Format to collect the request information. Setting to default collects the information in JSON format and origin collects the information with the original HTTP request. See examples below. |
log_format | object | False | Log format declared as key value pairs in JSON format. Values only support strings. APISIX or Nginx variables can be used by prefixing the string with $ . | ||
include_req_body | boolean | False | false | [false, true] | When set to true includes the request body in the log. If the request body is too big to be kept in the memory, it can't be logged due to Nginx's limitations. |
include_req_body_expr | array | False | Filter for when the include_req_body attribute is set to true . Request body is only logged when the expression set here evaluates to true . See lua-resty-expr for more. | ||
max_req_body_bytes | integer | False | 524288 | >=1 | Maximum request body allowed in bytes. Request bodies falling within this limit will be pushed to Kafka. If the size exceeds the configured value, the body will be truncated before being pushed to Kafka. |
include_resp_body | boolean | False | false | [false, true] | When set to true includes the response body in the log. |
include_resp_body_expr | array | False | Filter for when the include_resp_body attribute is set to true . Response body is only logged when the expression set here evaluates to true . See lua-resty-expr for more. | ||
max_resp_body_bytes | integer | False | 524288 | >=1 | Maximum response body allowed in bytes. Response bodies falling within this limit will be pushed to Kafka. If the size exceeds the configured value, the body will be truncated before being pushed to Kafka. |
cluster_name | integer | False | 1 | [0,...] | Name of the cluster. Used when there are two or more Kafka clusters. Only works if the producer_type attribute is set to async . |
producer_batch_num | integer | optional | 200 | [1,...] | batch_num parameter in lua-resty-kafka. The merge message and batch is send to the server. Unit is message count. |
producer_batch_size | integer | optional | 1048576 | [0,...] | batch_size parameter in lua-resty-kafka in bytes. |
producer_max_buffering | integer | optional | 50000 | [1,...] | max_buffering parameter in lua-resty-kafka representing maximum buffer size. Unit is message count. |
producer_time_linger | integer | optional | 1 | [1,...] | flush_time parameter in lua-resty-kafka in seconds. |
meta_refresh_interval | integer | optional | 30 | [1,...] | refresh_interval parameter in lua-resty-kafka specifies the time to auto refresh the metadata, in seconds. |
This Plugin supports using batch processors to aggregate and process entries (logs/data) in a batch. This avoids the need for frequently submitting the data. The batch processor submits data every 5
seconds or when the data in the queue reaches 1000
. See Batch Processor for more information or setting your custom configuration.
IMPORTANT
The data is first written to a buffer. When the buffer exceeds the batch_max_size
or buffer_duration
attribute, the data is sent to the Kafka server and the buffer is flushed.
If the process is successful, it will return true
and if it fails, returns nil
with a string with the "buffer overflow" error.
#
meta_format exampledefault
:{
"upstream": "127.0.0.1:1980",
"start_time": 1619414294760,
"client_ip": "127.0.0.1",
"service_id": "",
"route_id": "1",
"request": {
"querystring": {
"ab": "cd"
},
"size": 90,
"uri": "/hello?ab=cd",
"url": "http://localhost:1984/hello?ab=cd",
"headers": {
"host": "localhost",
"content-length": "6",
"connection": "close"
},
"body": "abcdef",
"method": "GET"
},
"response": {
"headers": {
"connection": "close",
"content-type": "text/plain; charset=utf-8",
"date": "Mon, 26 Apr 2021 05:18:14 GMT",
"server": "APISIX/2.5",
"transfer-encoding": "chunked"
},
"size": 190,
"status": 200
},
"server": {
"hostname": "localhost",
"version": "2.5"
},
"latency": 0
}origin
:GET /hello?ab=cd HTTP/1.1
host: localhost
content-length: 6
connection: close
abcdef
#
MetadataYou can also set the format of the logs by configuring the Plugin metadata. The following configurations are available:
Name | Type | Required | Default | Description |
---|---|---|---|---|
log_format | object | False | Log format declared as key value pairs in JSON format. Values only support strings. APISIX or Nginx variables can be used by prefixing the string with $ . |
IMPORTANT
Configuring the Plugin metadata is global in scope. This means that it will take effect on all Routes and Services which use the kafka-logger
Plugin.
The example below shows how you can configure through the Admin API:
note
You can fetch the admin_key
from config.yaml
and save to an environment variable with the following command:
admin_key=$(yq '.deployment.admin.admin_key[0].key' conf/config.yaml | sed 's/"//g')
curl http://127.0.0.1:9180/apisix/admin/plugin_metadata/kafka-logger -H "X-API-KEY: $admin_key" -X PUT -d '
{
"log_format": {
"host": "$host",
"@timestamp": "$time_iso8601",
"client_ip": "$remote_addr"
}
}'
With this configuration, your logs would be formatted as shown below:
{"host":"localhost","@timestamp":"2020-09-23T19:05:05-04:00","client_ip":"127.0.0.1","route_id":"1"}
{"host":"localhost","@timestamp":"2020-09-23T19:05:05-04:00","client_ip":"127.0.0.1","route_id":"1"}
#
Enable PluginThe example below shows how you can enable the kafka-logger
Plugin on a specific Route:
curl http://127.0.0.1:9180/apisix/admin/routes/5 -H "X-API-KEY: $admin_key" -X PUT -d '
{
"plugins": {
"kafka-logger": {
"brokers" : [
{
"host" :"127.0.0.1",
"port" : 9092
}
],
"kafka_topic" : "test2",
"key" : "key1",
"batch_max_size": 1,
"name": "kafka logger"
}
},
"upstream": {
"nodes": {
"127.0.0.1:1980": 1
},
"type": "roundrobin"
},
"uri": "/hello"
}'
This Plugin also supports pushing to more than one broker at a time. You can specify multiple brokers in the Plugin configuration as shown below:
"brokers" : [
{
"host" :"127.0.0.1",
"port" : 9092
},
{
"host" :"127.0.0.1",
"port" : 9093
}
],
#
Example usageNow, if you make a request to APISIX, it will be logged in your Kafka server:
curl -i http://127.0.0.1:9080/hello
#
Delete PluginTo remove the kafka-logger
Plugin, you can delete the corresponding JSON configuration from the Plugin configuration. APISIX will automatically reload and you do not have to restart for this to take effect.
curl http://127.0.0.1:9180/apisix/admin/routes/1 -H "X-API-KEY: $admin_key" -X PUT -d '
{
"methods": ["GET"],
"uri": "/hello",
"plugins": {},
"upstream": {
"type": "roundrobin",
"nodes": {
"127.0.0.1:1980": 1
}
}
}'