User Guide

Feature Modules Installation

NetEye Core is the set of most commonly used functionalities offered by the platform, including monitoring, visualization (both dashboards and maps), configuration, reporting, and event handling.

However, the NetEye modular architecture supports the installation of additional Feature Modules that extend the NetEye Core functionalities. This separation allows to customize NetEye in order to address specific customer needs.

NetEye Modules

Feature Modules are fully fledged modules, whose functionalities are well defined and established, and need the installation of a group of packages.

In order to install a Feature Module which changes slightly depending on the module’s type (please consider following the advice on Safe Command Execution when you do this), and then follow the procedure for either a Single Node or Cluster Node.

NetEye Modules

Each of these modules has its own, distinct contract and requires NetEye Core. Also the SIEM module dependency from Log Manager has been removed, starting from the 4.12 release and now requires only NetEye Core.

Module

Requires

Yum group name

Log Manager

NetEye Core

neteye-logmanagement

SIEM

NetEye Core

neteye-siem

vSphereDB

NetEye Core

neteye-vmd

SLM

NetEye Core

neteye-slm

Asset

NetEye Core

neteye-asset

ntopng

NetEye Core

neteye-ntopng

Command Orchestrator

NetEye Core

neteye-cmd

Single Node

To install a NetEye Module, run the following command with the appropriate Yum group name from the table above:

# yum -y groupinstall <yum-group-name> --enablerepo=neteye

Once done, please follow the procedure needed to update a NetEye single instance, then the directions on section Refreshing the additional module, to complete the overall installation.

Cluster Node

Differently from the Single Node, the NetEye Module or Preview Software must be installed on every node of the cluster with the same commands described in the previous section. So, the command(s):

# yum -y groupinstall <yum-group-name> --enablerepo=neteye

must be run on each node of the cluster.

After the yum installation of the <feature_module> OR <preview_software>, these additional steps are needed:

  • Look for the template file having filepath with pattern /usr/share/neteye/cluster/templates/Services-<name>-*.conf.tpl (where <name> is the name of the <feature_module> OR <preview_software> you are installing, and the * is a wildcard for any string). If any such file does not exist, skip the following steps and go to the next section.

  • If, on the contrary, any such file exists, adapt it to the settings of your cluster, and save it to a file with the same name without the .tpl suffix.

  • Now, for each file saved in the previous step, create the cluster resource by executing the following command on one of the nodes of the cluster (replace <name> with the name of the <feature_module> OR <preview_software> you are installing, and the * with the string that completes the actual filename):

    # /usr/share/neteye/scripts/cluster/cluster_service_setup.pl -c /usr/share/neteye/cluster/templates/Services-<name>-*.conf``
    

    When the execution of the script above has finished, please perform the steps described the procedure to Update a NetEye Cluster.

Example: I want to install the ‘asset’ feature module on a NetEye cluster.

After performing yum groupinstall neteye-asset --enablerepo=neteye on each node of the cluster, on one node I find the following files with pattern /usr/share/neteye/cluster/templates/Services-asset-*.conf.tpl:

/usr/share/neteye/cluster/templates/Services-asset-glpi.conf.tpl
/usr/share/neteye/cluster/templates/Services-asset-ocsinventory-ocsreports.conf.tpl
/usr/share/neteye/cluster/templates/Services-asset-ocsinventory-server.conf.tpl

I adapt them to my cluster settings (adapt the ip_pre, cidr_netmask and check the drbd_minor and the drbd_port in this case) and save them in the files:

/usr/share/neteye/cluster/templates/Services-asset-glpi.conf
/usr/share/neteye/cluster/templates/Services-asset-ocsinventory-ocsreports.conf
/usr/share/neteye/cluster/templates/Services-asset-ocsinventory-server.conf

I create the cluster resources with the commands:

# /usr/share/neteye/scripts/cluster/cluster_service_setup.pl -c /usr/share/neteye/cluster/templates/Services-asset-glpi.conf
# /usr/share/neteye/scripts/cluster/cluster_service_setup.pl -c /usr/share/neteye/cluster/templates/Services-asset-ocsinventory-ocsreports.conf
# /usr/share/neteye/scripts/cluster/cluster_service_setup.pl -c /usr/share/neteye/cluster/templates/Services-asset-ocsinventory-server.conf

Finally, follow the procedure to Update a NetEye Cluster. To complete the overall installation, please follow the directions on section Refreshing the additional module.

Verify if a module is running correctly

After installing a Feature Module or Preview Software, you need to make sure that all services are running.

The commands to be used differ on a Single Node and on Cluster Installations.

NetEye Single Node Installation

The neteye status command outputs a list of the status of all NetEye services, similar to the following snippet:

DOWN [3] elastic-blockchain-proxy.service
DOWN [3] elasticsearch.service
DOWN [3] eventhandlerd.service
UP   [0] filebeat.service
UP   [0] grafana-server.service
UP   [0] httpd.service
DOWN [3] icinga2-master.service
UP   [0] influxdb.service
DOWN [3] kibana-logmanager.service
DOWN [0] lampod.service
UP   [0] logstash.service
UP   [0] mariadb.service
DOWN [3] nats-server.service
UP   [0] neteye-agent.service
UP   [0] nginx.service
UP   [0] nprobe.service
UP   [0] ntopng.service
UP   [0] redis.service
UP   [0] rh-php73-php-fpm.service
UP   [0] rsyslog-logmanager.service
UP   [0] slmd.service
UP   [0] smsd.service
UP   [0] snmptrapd.service
UP   [0] tornado.service
DOWN [3] tornado_email_collector.service
DOWN [0] tornado_icinga2_collector.service
DOWN [3] tornado_nats_json_collector.service
DOWN [3] tornado_webhook_collector.service

Note

Output may vary, depending on both installed modules and running services.

Suppose you have just install Tornado and all its collectors: they should be running, but are marked as DOWN. This means that something has gone wrong and you need to understand why. You can therefore check the dedicated troubleshooting section for directions.

NetEye Cluster Installation

On a cluster it is necessary to differentiate between clustered and non clustered services: Non clustered services, which for example include Elasticsearch, follow the same approach shown in the previous section and in case of issues, can be inspected with the same commands mentioned in the corresponding troubleshooting section.

Clustered services, on the contrary, require a different approach. Indeed, the neteye status, neteye start, and neteye stop commands can not be used, because they are not available on cluster.

Note

Clustered services are referred to as Resources. For example, a Tornado instance running on a NetEye single installation is a service, while a Tornado instance running on a NetEye cluster is a resource.

Therefore, to verify if resources are correctly running, use the pcs status command, which outputs the status of the cluster and all the resources, similarly to the following excerpt.

Cluster name: NetEye
Stack: corosync
Current DC: neteye01.local (version 1.1.23-1.el7_9.1-9acf116022) - partition with quorum
Last updated: Wed Jul 28 09:47:52 2021
Last change: Tue Jul 27 15:04:36 2021 by root via cibadmin on neteye02.local
2 nodes configured
74 resource instances configured
Online: [ neteye01.local neteye02.local ]
Full list of resources:
 cluster_ip    (ocf::heartbeat:IPaddr2):    Started neteye02.local
 Resource Group: tornado_rsyslog_collector_group
     tornado_rsyslog_collector_drbd_fs    (ocf::heartbeat:Filesystem):    Started neteye02.local
 Resource Group: tornado_group

In case a resource is not starting correctly, it will be listed at the end of the output (see snippet below) as Failed. You need to understand why it is not running: the dedicated cluster troubleshooting section features options that you can apply to find the root cause of the problem.

Failed Resource Actions:
* tornado_email_collector_monitor_30000 on neteye02.local 'not running' (7): call=414, status=complete, exitreason='',
    last-rc-change='Wed Jul 28 09:57:21 2021', queued=0ms, exec=0ms

Refreshing the additional module

If the procedure you followed above was successful, you can now refresh the additional module with these steps:

  • Refresh your browser window. This will ensure that the new module appears in the NetEye menu and all Javascript and CSS is reloaded properly.

  • Log out of NetEye and then log back in so that any permissions or roles required by the new module will take effect.