Create a Post
Showing results for 
Search instead for 
Did you mean: 

Log Exporter and Elasticsearch

There have been multiple posts requesting information about Check Point's options for integration with an ELK stack for logging. Many of the posts recommend using log exporter over LEA for exporting logs to that system. I have not seen a post that covers configuration of the ELK stack in order to use this exported data.  I put these behind spoiler tags as this is a fairly long winded post as I cover some of the basics before getting into the configuration options. 


Click to Expand


  • Log data within the "Discover" section of Kibana.



  • Basic Dashboard pic6.png




Click to Expand

Software Revisions

  • Check Point R80.30 (JHF 76)
  • Elasticsearch 7.4.0
  • Logstash 7.4.0
    • Logstash Plugins
      • logstash-input-udp
      • logstash-filter-kv
      • logstash-filter-mutate
      • logstash-output-elasticsearch
  • Kibana 7.4.0
  • Arch Linux



What is Elasticsearch?
From Elastic’s website:
     Elasticsearch is a distributed, RESTful search and analytics engine capable of solving a growing number of use cases. As the heart of the Elastic Stack, it centrally stores your data so you can discover the expected and uncover the unexpected.

Elasticsearch provides an API driven indexing solution for querying and storing large quantities of data. This can be used standalone or clustered solution. Elasticsearch also has additional deployment options providing real-time search integration with other database and storage platforms.

In Arch Linux, Elasticsearch configuration files are stored under /etc/elasticsearch. This guide uses the default settings within the elasticsearch.yml file. The default configuration file has reasonable defaults and listens on TCP port 9200 on localhost. The JAVA configuration is stored in the jvm.options file. This is to adjust memory requirements used for the Java processes running Elasticsearch.



What is Kibana?
From Elastic’s website:
     Kibana lets you visualize your Elasticsearch data and navigate the Elastic Stack…

Kibana is a customizable web interface that interacts with Elasticsearch in order to build dashboard, visualization, or search for stored data. In Arch Linux, the configuration folder is /etc/kibana. The kibana.yml file has a useful set of defaults that I recommend validating.

  • server.port: 5601
    o This is the service port the server will listen on for inbound HTTP(s) connections.
  • “localhost”
    o This will be the server name that kibana will use for this particular instance. Since this document only uses a single instance, the default name was used.
  • server.basePath: “/kibana”
    o This will be the URL path used when running behind a web service or reverse-proxy. In this example, the requests will be sent to http://host/kibana
  • server.rewriteBasePath: true
    o This forces the kibana process to rewrite outbound responses to include the basePath in the URL.



What is Logstash?

From Elastic’s website:
     Logstash is an open source, server-side data processing pipeline that ingests data from a multitude of sources simultaneously, transforms it, and then sends it to your favorite “stash.” (Ours is Elasticsearch, naturally.)

Logstash provides a method for receiving content, manipulating that content, and forwarding it on to a backend system for additional use. Logstash can manipulates these streams into acceptable formats for storage and indexing or additional processing of the content.

In Arch Linux, the configuration folder is /etc/logstash. The Arch Linux package starts the logstash process and reads the configuration files under /etc/logstash/conf.d. This allows logstash to use multiple input sources, filters, and outputs. For example, logstash could be configured as a syslog service and output specific content to an Elasticsearch cluster on the network. It could also have another configuration file that listens on port 5044 for separate data to export into a different Elasticsearch cluster.

Logstash Plugins

Depending on the Linux distribution, several logstash plugins might not be available from the distribution file repository. In order to add additional plugins, the use of git is required to import them from the Logstash plugin repository. The logstash-plugins repository are available online at and their installation process is documented on that site.



NGINX is a web and reverse proxy server. It also has other capabilities such as load balancing and web application firewall. The configuration for nginx is located in the /etc/nginx directory. Since NGINX can have a wide array of configurations, only the configuration used to allow access to Kibana is documented below. The configuration below accepts connections to /kibana and forwards them to the localhost:5601 service. For additional configuration options for NGINX, see

Snippet from /etc/nginx/nginx.conf
http {
   include mime.types;
   sendfile on;
   keepalive_timeout 65;

   ## This provides a reference to the Elasticsearch service listening on localhost:9200
   upstream kibana {
       keepalive 15;
server {
    listen 80;
    server_name _;
    location / {
        root /usr/share/nginx/html;
        index index.html index.htm;

    location /kibana {
        proxy_pass http://kibana;

    error_page 500 502 503 504 /50x.html;
    location = /50x.html {
        root /usr/share/nginx/html;


Check Point Log Exporter

Check Point Log Exporter documentation can be found in SK article 122323 on Support Center. The configuration used in this document is very simple. The logstash configuration file is expecting the syslog messages to use the Splunk format. The Splunk format provides field delineation using the | character. This provides an easy option for Logstash’s KV filter to use in order to split the log fields.

[Expert@vMgmt01:0]# cp_log_export add name lab01 target-server target-port 514 protocol udp format splunk read-mode semi-unified

Export settings for lab01 has been added successfully

To apply the changes run: cp_log_export restart name lab01



Logstash Configuration

There is a single configuration file defined for logstash in this example. Logstash will attempt to load all YML files in the configuration directory. Each file could have a different set of inputs, filters, and outputs.  The location of the YAML file might differ depending on the distribution.

In this example, there are three primary sections: input, filter, and output.

The input configuration tells the logstash process which plugins to run to receive content from external sources. Our example uses the UDP input plugin (logstash-input-udp) and is configured to act as a syslog service.

Logstash has many other options for input types and content codecs. Additional information can be found by accessing the logstash Input plugin documentation at

input {
    udp {
        port => 514
        codec => plain
        type => syslog


Logstash filters match, label, edit, and make decisions on content before it passes into Elasticsearch (ES). The filter section is where the patterns and labels are defined. In our example, logstash processes the syslog data through two separate filters.

The kv filter is used to automatically assign labels to the received logs. Since the messages sent via log_exporter are in a <key>=<value> format, the kv filter was chosen. This provides a simpler mechanism than using the grok or dissect plugins for assigning those values.

filter {

     if [type] == "syslog" {
        kv {
             allow_duplicate_values => false
             recursive => false
             field_split => "\|"


        mutate {
            # Fields beginning with underscore are not supported by ES/Kibana, so rename them.
rename => { "__nsons" => "nsons" }
             rename => { "__p_dport" => "p_dport" }
             rename => { "__pos" => "pos" }

             # Not necessary, just a field name preference.
             rename => { "originsicname" => "sicname" }

             # Example of removing specific fields
# remove_field => [ "connection_luuid", "loguid" ]

             # String substitution
# Strip the O\=.*$ and the ^CN= in the field.
gsub => [

                  "sicname", "CN\\=", "",
                  "sicname", ",O\\=.*$",""

The mutate plugin (logstash-filter-mutate) is used to manipulate the data as needed.  In the configuration above, the originsicname field is renamed to sicname. Additional fields can be dropped using the remove_field configuration.

The configuration can be validated using the “-t” parameter when launching logstash. Configuration errors will be displayed and include the line numbers for the error.

/usr/share/logstash/bin/logstash -t -f /etc/logstash/conf.d/checkpoint.yml


Kibana - Basic Configuration

The default settings for Kibana are used in this document, so there are no additional configuration steps necessary. The NGINX configuration used in the earlier section will pass the requests to http://hostname/kibana to the http://localhost:5601 service.

After opening the management page, Kibana needs to be configured to use the data contained within Elasticsearch.

  • Go to Management -> Index Patterns
  • pic1.png

     Select Create Pattern and add a pattern that will highlight the logstash indices and click next.


  • The next step will ask for the Time filter setting. Select the @timestamp default and click “Create Index Pattern”



  • Once this step has been completed, Kibana can now present available data from Elasticsearch.


12 Replies

Thanks for sharing 😁

It would be great if this could be added to the Log Exporter SK


One customer said us about this integration, we work with syslog object for sending logs through a  VPN to Elastic SIEM to the another partner. It works more easy that log exporter.

0 Kudos

Except you are missing a LOT of logs using this approach.
You are only getting basic firewall logs and nothing from other blades (IPS, App Control, URL Filtering, etc).

Nice work!



How do I send specific logs, Critical IPS logs to the elastic? only IPS.


Thancks help me

0 Kudos

Arch Linux? 🙂

@masher Masher:  Excellent walk through.  The LS pipeline snippit in the spoiler's Logstash Config snippit is most helpful.  I look forward to trying the splunk exporter with the kv filter.



Is there a way to force Checkpoint log exporter to include a timestamp on messages created using the the cef format? 



Logstash's "syslog" input supports the 'cef' codec.  This is the route we were originally exploring - and it did successfully parse the log messages from checkpoints cef format log exporter into key:value fields.  However, the logs created when you choose the 'cef' format of the Checkpoint log exporter do not (for the most part - read on) include a timestamp (which is pretty important for time series log data)! When we looked at the raw output from a handful of the other Checkpoint output formats options (syslog, generic, splunk for example) they all included timestamps with every message - it is only missing from the cef format.  Looking at the cef format field mappings  <> we see that start_time maps to start.  Unfortunately, only ~1/20 of the raw log messages (validated by using netcat on the Checkpoint manager itself) created by the cef exporter contained the 'start' field and there was nothing else that looked like a timestamp in the documents.  I bring up the validation methodology (we used nc -l -p 514 | tee file_name.pcap, after creating an exporter: this should be easy to recreate) so we can rule out that the logstash cef codec was somehow stripping the data.

I'm hoping we're missing something obvious and that someone can give us a quick and easy way to get timestamps out of the cef format log exporter.  If not, we'll follow your path of using splunk exporter with the kv filter (thanks again for that!).





0 Kudos

I'm not sure why there would be a difference in the how  the CEF exports sends the fields versus other log_exporter export formats. I do recall that I had some issues when attempting to use logstash's CEF input which is why I ended up with using kv.

It might be worth opening a new topic with some screenshots and examples.


@masher:  Thanks for the reply and feedback.  In our search for the path of least resistance, we'll likely follow the path you carved.  We'll shift to generic (or splunk) output, use the KV filter and see if that meets our needs. 


logstash's cef input codec seemed to function as expected.  That is to say, when netcat (running on the checkpoint manager) showed that there was a key/field, this key/field was reliably parsed by logstash's cef codec and shipped to elasticsearch.  

0 Kudos
Employee Alumnus
Employee Alumnus

Hi @satsujinsha 

May I ask you where exactly do you want to add time field? in the header of the log or in its body?

Time field is exported by default as part of the body of the log, look for the field 'rt'.


Please let me know about it, I will be happy to help here.




0 Kudos

@masher Have you been able to get log exporter to work with the new elastic-agent and built in Check Point Integration? The documentation on the Integration is lackluster and I've been unable to get the log exporter to work with the agent.



Unfortunately, I have not had an opportunity to play around with the new integration options with Elasticsearch. I have plans to redo this how-to to take advantage of newer options, but I have not had time to get to that yet.

0 Kudos


Epsum factorial non deposit quid pro quo hic escorol.

Upcoming Events

    CheckMates Events