[syslog-ng] Connect syslog-ng to HDFS

Lee Keng Ket kengket at gmail.com
Thu Aug 23 01:53:30 UTC 2018


Hi, Gabor

After I change the lib to executable as reminded by Janos, now there is
different error. Any idea if this is Kerberos issue or still the Hadoop lib
issue?

The Hadoop that I'm using is Huawei FusionInsight. Hadoop HDFS version
should be 2.7.2.

This is the syslog-ng.conf for HDFS part:
destination d_hdfs {
        hdfs(client_lib_dir("/opt/hadoop/lib")
        hdfs_uri("hdfs://x.x.x.x:25000")
        kerberos-keytab-file("/etc/syslog.keytab")
        kerberos-principal("syslog at HADOOP.COM")
        hdfs_file("/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log")
        template(t_cgnat)
        hdfs-append-enabled(true)
        );
};

There are the messages I get when starting in debug:

[2018-08-23T09:40:09.210168] Running application hooks; hook='1'
[2018-08-23T09:40:09.210192] Running application hooks; hook='3'
[2018-08-23T09:40:09.210501] syslog-ng starting up; version='3.14.1'
[2018-08-23T09:40:09.213049] Worker thread started; driver='d_hdfs#0'
[2018-08-23T09:40:09.214922] Opening hdfs;
[2018-08-23T09:40:09.548286] field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess
with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of
successful kerberos logins and latency (milliseconds)]);
[2018-08-23T09:40:09.561345] field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure
with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of
failed kerberos logins and latency (milliseconds)]);
[2018-08-23T09:40:09.561608] field
org.apache.hadoop.metrics2.lib.MutableRate
org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with
annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
sampleName=Ops, always=false, type=DEFAULT, valueName=Time,
value=[GetGroups]);
[2018-08-23T09:40:09.562485] UgiMetrics, User and group related metrics;
[2018-08-23T09:40:09.604037] Failed to detect a valid hadoop home directory;
[2018-08-23T09:40:09.687386] setsid exited with exit code 0;
[2018-08-23T09:40:09.715804]  Creating new Groups object;
[2018-08-23T09:40:09.717743] Trying to load the custom-built native-hadoop
library...;
[2018-08-23T09:40:09.718065] Failed to load native-hadoop with error:
java.lang.UnsatisfiedLinkError: no hadoop in java.library.path;
[2018-08-23T09:40:09.718095] java.library.path=//usr/lib64/syslog-ng;
[2018-08-23T09:40:09.718117] Unable to load native-hadoop library for your
platform... using builtin-java classes where applicable;
[2018-08-23T09:40:09.718418] Falling back to shell based;
[2018-08-23T09:40:09.718997] Group mapping
impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping;
[2018-08-23T09:40:09.753615] Group mapping
impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback\;
cacheTimeout=300000\; warningDeltaMs=5000;
[2018-08-23T09:40:09.905060] hadoop login;
[2018-08-23T09:40:09.906493] hadoop login commit;
[2018-08-23T09:40:09.907300] using kerberos user:syslog at HADOOP.COM;
[2018-08-23T09:40:09.907333] Using user: "syslog at HADOOP.COM" with name
syslog at HADOOP.COM;
[2018-08-23T09:40:09.907592] User entry: "syslog at HADOOP.COM";
[2018-08-23T09:40:09.908005] Login successful for user syslog at HADOOP.COM
using keytab file syslog.keytab;
[2018-08-23T09:40:10.104386] dfs.client.use.legacy.blockreader.local =
false;
[2018-08-23T09:40:10.104436] dfs.client.read.shortcircuit = false;
[2018-08-23T09:40:10.104450] dfs.client.domain.socket.data.traffic = false;
[2018-08-23T09:40:10.104461] dfs.domain.socket.path = ;
[2018-08-23T09:40:10.121280] Sets
dfs.client.block.write.replace-datanode-on-failure.replication to 0;
[2018-08-23T09:40:10.144892] multipleLinearRandomRetry = null;
[2018-08-23T09:40:10.168901] rpcKind=RPC_PROTOCOL_BUFFER,
rpcRequestWrapperClass=class
org.apache.hadoop.ipc.ProtobufRpcEngine$RpcProtobufRequest,
rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker at 4379f1a4
;
[2018-08-23T09:40:10.173738] getting client out of cache:
org.apache.hadoop.ipc.Client at 5b52b909;
[2018-08-23T09:40:10.238466] finalize() called.;
[2018-08-23T09:40:10.238705] finalize() called.;
[2018-08-23T09:40:10.526021] Both short-circuit local reads and UNIX domain
socket are disabled.;
[2018-08-23T09:40:10.532037] DataTransferProtocol not using
SaslPropertiesResolver, no QOP found in configuration for
dfs.data.transfer.protection;
[2018-08-23T09:40:10.555581] The ping interval is 60000 ms.;
[2018-08-23T09:40:10.556336] Connecting to /x.x.x.x:25000;
[2018-08-23T09:40:10.572385] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:10.613823] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:10.723447] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:10.723514] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:10.730466] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:10.758296] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:10.759031] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:10.759531] Exception encountered while connecting to the
server : javax.security.sasl.SaslException: No common protection layer
between client and server;
[2018-08-23T09:40:14.446925] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:14.447824] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:14.447982] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:14.448001] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:14.449087] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:14.455070] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:14.455190] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:14.455476] Exception encountered while connecting to the
server : javax.security.sasl.SaslException: No common protection layer
between client and server;
[2018-08-23T09:40:17.206928] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:17.207978] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:17.208115] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:17.208133] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:17.208877] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:17.214382] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:17.214536] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:17.214845] Exception encountered while connecting to the
server : javax.security.sasl.SaslException: No common protection layer
between client and server;
[2018-08-23T09:40:17.535313] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:17.536419] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:17.536615] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:17.536634] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:17.537679] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:17.546375] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:17.546587] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:17.546963] Exception encountered while connecting to the
server : javax.security.sasl.SaslException: No common protection layer
between client and server;
[2018-08-23T09:40:21.891382] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:21.892476] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:21.909267] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:21.909306] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:21.910125] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:21.915324] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:21.915540] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:21.915889] Exception encountered while connecting to the
server : javax.security.sasl.SaslException: No common protection layer
between client and server;
[2018-08-23T09:40:26.095299] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
[2018-08-23T09:40:26.096234] Get kerberos info proto:interface
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
serverPrincipal=dfs.namenode.kerberos.principal);
[2018-08-23T09:40:26.096391] RPC Server's Kerberos principal name for
protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
hadoop.hadoop.com at HADOOP.COM;
[2018-08-23T09:40:26.096411] Creating SASL GSSAPI(KERBEROS)  client to
authenticate to service at hadoop.hadoop.com;
[2018-08-23T09:40:26.097116] Use KERBEROS authentication for protocol
ClientNamenodeProtocolPB;
[2018-08-23T09:40:26.101823] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:javax.security.sasl.SaslException: No common
protection layer between client and server;
[2018-08-23T09:40:26.101960] PrivilegedAction as:syslog at HADOOP.COM
(auth:KERBEROS)
from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
[2018-08-23T09:40:26.102283] Couldn't setup connection for syslog at HADOOP.COM
to /x.x.x.x:25000;
[2018-08-23T09:40:26.102365] PrivilegedActionException as:syslog at HADOOP.COM
(auth:KERBEROS) cause:java.io.IOException: Couldn't setup connection for
syslog at HADOOP.COM to /x.x.x.x:25000;
[2018-08-23T09:40:26.103844] closing ipc connection to /x.x.x.x:25000:
Couldn't setup connection for syslog at HADOOP.COM to /x.x.x.x:25000;
[2018-08-23T09:40:26.103994] IPC Client (1766482975) connection to
/x.x.x.x:25000 from syslog at HADOOP.COM: closed;

Thank you.


On Tue, Aug 21, 2018 at 6:48 PM Nagy, Gábor <gabor.nagy at oneidentity.com>
wrote:

> Hi,
>
> Can you tell ne what is the version of the hadoop lib you use with
> syslog-ng, please?
>
> Can you share your syslog-ng configuration, mainly the hdfs part, please?
>
> Regards,
> Gabor
>
> On Tue, Aug 21, 2018 at 4:15 AM Lee Keng Ket <kengket at gmail.com> wrote:
>
>> Hi, Gabor
>>
>> I have run it, seems like it stops at the HDFS side.
>>
>> [2018-08-21T10:07:51.212015] Worker thread started; driver='d_hdfs#0'
>> [2018-08-21T10:07:51.212499] Running application hooks; hook='1'
>> [2018-08-21T10:07:51.212516] Running application hooks; hook='3'
>> [2018-08-21T10:07:51.212595] syslog-ng starting up; version='3.14.1'
>> [2018-08-21T10:07:51.214113] Opening hdfs;
>> [2018-08-21T10:08:01.215622] Opening hdfs;
>> [2018-08-21T10:08:11.216050] Opening hdfs;
>> [2018-08-21T10:08:21.226340] Opening hdfs;
>> [2018-08-21T10:08:31.236589] Opening hdfs;
>> [2018-08-21T10:08:41.240623] Opening hdfs;
>> [2018-08-21T10:08:51.250879] Opening hdfs;
>> [2018-08-21T10:09:01.261172] Opening hdfs;
>> [2018-08-21T10:09:11.271410] Opening hdfs;
>> [2018-08-21T10:09:21.281685] Opening hdfs;
>> [2018-08-21T10:09:31.290765] Opening hdfs;
>> [2018-08-21T10:09:41.301098] Opening hdfs;
>> [2018-08-21T10:09:51.311362] Opening hdfs;
>> [2018-08-21T10:10:01.321152] Opening hdfs;
>> [2018-08-21T10:10:11.321818] Opening hdfs;
>> [2018-08-21T10:10:21.330114] Opening hdfs;
>> [2018-08-21T10:10:31.340413] Opening hdfs;
>> [2018-08-21T10:10:41.350654] Opening hdfs;
>> [2018-08-21T10:10:51.354016] Opening hdfs;
>> [2018-08-21T10:11:01.364267] Opening hdfs;
>> [2018-08-21T10:11:11.374516] Opening hdfs;
>> [2018-08-21T10:11:21.384761] Opening hdfs;
>> [2018-08-21T10:11:31.395017] Opening hdfs;
>> [2018-08-21T10:11:41.402256] Opening hdfs;
>> [2018-08-21T10:11:51.404097] Opening hdfs;
>> ^C[2018-08-21T10:11:59.672252] syslog-ng shutting down; version='3.14.1'
>> Exception in thread "" java.lang.NoClassDefFoundError:
>> org/apache/hadoop/conf/Configuration
>>         at
>> org.syslog_ng.hdfs.HdfsDestination.open(HdfsDestination.java:92)
>>         at org.syslog_ng.LogDestination.openProxy(LogDestination.java:65)
>> [2018-08-21T10:11:59.774895] Worker thread finished; driver='d_hdfs#0'
>> [2018-08-21T10:11:59.775384] Closing log transport fd; fd='13'
>> [2018-08-21T10:11:59.775508] Deinitialize hdfs destination;
>> [2018-08-21T10:11:59.776534] Java machine free;
>> [2018-08-21T10:11:59.778421] Running application hooks; hook='4'
>>
>> Any idea what to be checked further?
>>
>> Thank you.
>>
>> On Fri, Aug 17, 2018 at 4:45 PM Nagy, Gábor <gabor.nagy at oneidentity.com>
>> wrote:
>>
>>> Hello!
>>>
>>> In the statistics it can be seen that the log message is not sent to the
>>> HDFS server:
>>> dropped='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=0'
>>> processed='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1'
>>> queued='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1'
>>>
>>> Well, generally on write error there should be an exception that results
>>> in an error message.
>>>
>>> You should try debugging it either in running syslog-ng in foreground
>>> (-F option), forwarding internal logs to stderr (-e) and with debug mode
>>> (-dv) on.
>>> Or in service mode use the internal() source in your config and connect
>>> it to a destination (e.g. file()) which you prefer.
>>>
>>> You could turn on debug messages on java side too using jvm_options() in
>>> syslog-ng config and configuring the log4j logging service, e.g.:
>>> options {
>>>
>>> jvm_options("-Dlog4j.configuration=file:/etc/hadoop/log4j.properties
>>> -Dlog4j.debug=true");
>>> };
>>>
>>> Regards,
>>> Gabor
>>>
>>> On Fri, Aug 17, 2018 at 10:34 AM Czanik, Péter <peter.czanik at balabit.com>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> As https://copr.fedorainfracloud.org/coprs/czanik/syslog-ng314/ also
>>>> writes: "Java is enabled, but JAR dependencies are not provided in package,
>>>> except for Elasticsearch http mode." The syslog-ng-java-deps.noarch
>>>> contains build time dependencies. Probably I should rename the package to
>>>> syslog-ng-java-build-deps...
>>>>
>>>> Check the documentation at
>>>> https://www.syslog-ng.com/technical-documents/doc/syslog-ng-open-source-edition/3.16/administration-guide/33#TOPIC-956506
>>>> on how to download and configure HDFS related JAR dependencies.
>>>>
>>>> Bye,
>>>>
>>>> Peter Czanik (CzP) <peter.czanik at balabit.com>
>>>> Balabit / syslog-ng upstream
>>>> https://syslog-ng.com/community/
>>>> https://twitter.com/PCzanik
>>>>
>>>> On Fri, Aug 17, 2018 at 10:22 AM, Lee Keng Ket <kengket at gmail.com>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I'm trying to connect syslog-ng 3.14.1 to HDFS to store the syslog
>>>>> messages. The syslog-ng can start without error, and it's able to write
>>>>> into local file. However, the log is not written to the HDFS. As there is
>>>>> no single error, I'm not sure how I should troubleshoot on this.
>>>>>
>>>>> I have installed the syslog-ng from this repo,
>>>>> https://copr.fedorainfracloud.org/coprs/czanik/syslog-ng314/repo/epel-7/czanik-syslog-ng314-epel-7.repo
>>>>>
>>>>> Installed Packages
>>>>> syslog-ng.x86_64
>>>>> 3.14.1-4.el7.centos
>>>>> @czanik-syslog-ng314
>>>>> syslog-ng-java.x86_64
>>>>> 3.14.1-4.el7.centos
>>>>> @czanik-syslog-ng314
>>>>> syslog-ng-java-deps.noarch
>>>>> 1.0-2
>>>>> @czanik-syslog-ng314
>>>>>
>>>>> This is the message from /var/log/message:
>>>>> Log statistics; processed='src.internal(s_sys#0)=1',
>>>>> stamp='src.internal(s_sys#0)=1534491834',
>>>>> processed='destination(d_spol)=0', processed='destination(d_mlal)=0',
>>>>> processed='center(received)=2', processed='destination(d_mesg)=1',
>>>>> processed='destination(d_mail)=0', processed='destination(d_auth)=0',
>>>>> processed='destination(d_cron)=0', processed='destination(d_hdfs)=1',
>>>>> processed='center(queued)=3', queued='global(scratch_buffers_count)=0',
>>>>> processed='source(remote_log)=1',
>>>>> dropped='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=0',
>>>>> processed='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1',
>>>>> queued='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1',
>>>>> processed='global(payload_reallocs)=0',
>>>>> processed='src.journald(journal)=0', stamp='src.journald(journal)=0',
>>>>> processed='global(sdata_updates)=0',
>>>>> queued='global(scratch_buffers_bytes)=0',
>>>>> processed='destination(d_boot)=0', processed='destination(d_kern)=0',
>>>>> processed='source(s_sys)=1', processed='destination(remote)=1',
>>>>> processed='global(internal_queue_length)=0',
>>>>> processed='global(msg_clones)=0'
>>>>>
>>>>> Anyone has any idea how should I proceed the troubleshooting?
>>>>>
>>>>>
>>>>> ______________________________________________________________________________
>>>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>>>> Documentation:
>>>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>>>
>>>>>
>>>>>
>>>>
>>>> ______________________________________________________________________________
>>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>>> Documentation:
>>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>>
>>>>
>>> ______________________________________________________________________________
>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>> Documentation:
>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>
>>>
>> ______________________________________________________________________________
>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>> Documentation:
>> http://www.balabit.com/support/documentation/?product=syslog-ng
>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>
>>
> ______________________________________________________________________________
> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
> Documentation:
> http://www.balabit.com/support/documentation/?product=syslog-ng
> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.balabit.hu/pipermail/syslog-ng/attachments/20180823/c4ec9d93/attachment-0001.html>


More information about the syslog-ng mailing list