[syslog-ng] Connect syslog-ng to HDFS

Nagy, Gábor gabor.nagy at oneidentity.com
Fri Aug 24 14:29:05 UTC 2018


Hi,

As I was searching for answer what i saw was that "SaslException: No common
protection layer between client and server;" can occur
if the configuration differs on client-server side.

You should configure the "hadoop.rpc.protection" value in core-site.xml on
the client side to the same one as on the server side.
To use that xml with hdfs, you should put the filename in syslog-ng config
in the hdfs destination block:
destination d_hdfs {
  ...
  hdfs-resources("/path/to/xmlconfig/core-site.xml");
};

Regards,
Gabor

On Thu, Aug 23, 2018 at 3:53 AM Lee Keng Ket <kengket at gmail.com> wrote:

> Hi, Gabor
>
> After I change the lib to executable as reminded by Janos, now there is
> different error. Any idea if this is Kerberos issue or still the Hadoop lib
> issue?
>
> The Hadoop that I'm using is Huawei FusionInsight. Hadoop HDFS version
> should be 2.7.2.
>
> This is the syslog-ng.conf for HDFS part:
> destination d_hdfs {
>         hdfs(client_lib_dir("/opt/hadoop/lib")
>         hdfs_uri("hdfs://x.x.x.x:25000")
>         kerberos-keytab-file("/etc/syslog.keytab")
>         kerberos-principal("syslog at HADOOP.COM")
>         hdfs_file("/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log")
>         template(t_cgnat)
>         hdfs-append-enabled(true)
>         );
> };
>
> There are the messages I get when starting in debug:
>
> [2018-08-23T09:40:09.210168] Running application hooks; hook='1'
> [2018-08-23T09:40:09.210192] Running application hooks; hook='3'
> [2018-08-23T09:40:09.210501] syslog-ng starting up; version='3.14.1'
> [2018-08-23T09:40:09.213049] Worker thread started; driver='d_hdfs#0'
> [2018-08-23T09:40:09.214922] Opening hdfs;
> [2018-08-23T09:40:09.548286] field
> org.apache.hadoop.metrics2.lib.MutableRate
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess
> with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
> sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of
> successful kerberos logins and latency (milliseconds)]);
> [2018-08-23T09:40:09.561345] field
> org.apache.hadoop.metrics2.lib.MutableRate
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure
> with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
> sampleName=Ops, always=false, type=DEFAULT, valueName=Time, value=[Rate of
> failed kerberos logins and latency (milliseconds)]);
> [2018-08-23T09:40:09.561608] field
> org.apache.hadoop.metrics2.lib.MutableRate
> org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with
> annotation @org.apache.hadoop.metrics2.annotation.Metric(about=,
> sampleName=Ops, always=false, type=DEFAULT, valueName=Time,
> value=[GetGroups]);
> [2018-08-23T09:40:09.562485] UgiMetrics, User and group related metrics;
> [2018-08-23T09:40:09.604037] Failed to detect a valid hadoop home
> directory;
> [2018-08-23T09:40:09.687386] setsid exited with exit code 0;
> [2018-08-23T09:40:09.715804]  Creating new Groups object;
> [2018-08-23T09:40:09.717743] Trying to load the custom-built native-hadoop
> library...;
> [2018-08-23T09:40:09.718065] Failed to load native-hadoop with error:
> java.lang.UnsatisfiedLinkError: no hadoop in java.library.path;
> [2018-08-23T09:40:09.718095] java.library.path=//usr/lib64/syslog-ng;
> [2018-08-23T09:40:09.718117] Unable to load native-hadoop library for your
> platform... using builtin-java classes where applicable;
> [2018-08-23T09:40:09.718418] Falling back to shell based;
> [2018-08-23T09:40:09.718997] Group mapping
> impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping;
> [2018-08-23T09:40:09.753615] Group mapping
> impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback\;
> cacheTimeout=300000\; warningDeltaMs=5000;
> [2018-08-23T09:40:09.905060] hadoop login;
> [2018-08-23T09:40:09.906493] hadoop login commit;
> [2018-08-23T09:40:09.907300] using kerberos user:syslog at HADOOP.COM;
> [2018-08-23T09:40:09.907333] Using user: "syslog at HADOOP.COM" with name
> syslog at HADOOP.COM;
> [2018-08-23T09:40:09.907592] User entry: "syslog at HADOOP.COM";
> [2018-08-23T09:40:09.908005] Login successful for user syslog at HADOOP.COM
> using keytab file syslog.keytab;
> [2018-08-23T09:40:10.104386] dfs.client.use.legacy.blockreader.local =
> false;
> [2018-08-23T09:40:10.104436] dfs.client.read.shortcircuit = false;
> [2018-08-23T09:40:10.104450] dfs.client.domain.socket.data.traffic = false;
> [2018-08-23T09:40:10.104461] dfs.domain.socket.path = ;
> [2018-08-23T09:40:10.121280] Sets
> dfs.client.block.write.replace-datanode-on-failure.replication to 0;
> [2018-08-23T09:40:10.144892] multipleLinearRandomRetry = null;
> [2018-08-23T09:40:10.168901] rpcKind=RPC_PROTOCOL_BUFFER,
> rpcRequestWrapperClass=class
> org.apache.hadoop.ipc.ProtobufRpcEngine$RpcProtobufRequest,
> rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker at 4379f1a4
> ;
> [2018-08-23T09:40:10.173738] getting client out of cache:
> org.apache.hadoop.ipc.Client at 5b52b909;
> [2018-08-23T09:40:10.238466] finalize() called.;
> [2018-08-23T09:40:10.238705] finalize() called.;
> [2018-08-23T09:40:10.526021] Both short-circuit local reads and UNIX
> domain socket are disabled.;
> [2018-08-23T09:40:10.532037] DataTransferProtocol not using
> SaslPropertiesResolver, no QOP found in configuration for
> dfs.data.transfer.protection;
> [2018-08-23T09:40:10.555581] The ping interval is 60000 ms.;
> [2018-08-23T09:40:10.556336] Connecting to /x.x.x.x:25000;
> [2018-08-23T09:40:10.572385] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:10.613823] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:10.723447] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:10.723514] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:10.730466] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:10.758296] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:10.759031] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:10.759531] Exception encountered while connecting to the
> server : javax.security.sasl.SaslException: No common protection layer
> between client and server;
> [2018-08-23T09:40:14.446925] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:14.447824] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:14.447982] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:14.448001] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:14.449087] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:14.455070] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:14.455190] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:14.455476] Exception encountered while connecting to the
> server : javax.security.sasl.SaslException: No common protection layer
> between client and server;
> [2018-08-23T09:40:17.206928] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:17.207978] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:17.208115] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:17.208133] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:17.208877] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:17.214382] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:17.214536] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:17.214845] Exception encountered while connecting to the
> server : javax.security.sasl.SaslException: No common protection layer
> between client and server;
> [2018-08-23T09:40:17.535313] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:17.536419] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:17.536615] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:17.536634] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:17.537679] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:17.546375] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:17.546587] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:17.546963] Exception encountered while connecting to the
> server : javax.security.sasl.SaslException: No common protection layer
> between client and server;
> [2018-08-23T09:40:21.891382] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:21.892476] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:21.909267] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:21.909306] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:21.910125] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:21.915324] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:21.915540] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:21.915889] Exception encountered while connecting to the
> server : javax.security.sasl.SaslException: No common protection layer
> between client and server;
> [2018-08-23T09:40:26.095299] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:771);
> [2018-08-23T09:40:26.096234] Get kerberos info proto:interface
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB
> info:@org.apache.hadoop.security.KerberosInfo(clientPrincipal=,
> serverPrincipal=dfs.namenode.kerberos.principal);
> [2018-08-23T09:40:26.096391] RPC Server's Kerberos principal name for
> protocol=org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB is hdfs/
> hadoop.hadoop.com at HADOOP.COM;
> [2018-08-23T09:40:26.096411] Creating SASL GSSAPI(KERBEROS)  client to
> authenticate to service at hadoop.hadoop.com;
> [2018-08-23T09:40:26.097116] Use KERBEROS authentication for protocol
> ClientNamenodeProtocolPB;
> [2018-08-23T09:40:26.101823] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS)
> cause:javax.security.sasl.SaslException: No common protection layer between
> client and server;
> [2018-08-23T09:40:26.101960] PrivilegedAction as:syslog at HADOOP.COM
> (auth:KERBEROS)
> from:org.apache.hadoop.ipc.Client$Connection.handleSaslConnectionFailure(Client.java:694);
> [2018-08-23T09:40:26.102283] Couldn't setup connection for
> syslog at HADOOP.COM to /x.x.x.x:25000;
> [2018-08-23T09:40:26.102365] PrivilegedActionException
> as:syslog at HADOOP.COM (auth:KERBEROS) cause:java.io.IOException: Couldn't
> setup connection for syslog at HADOOP.COM to /x.x.x.x:25000;
> [2018-08-23T09:40:26.103844] closing ipc connection to /x.x.x.x:25000:
> Couldn't setup connection for syslog at HADOOP.COM to /x.x.x.x:25000;
> [2018-08-23T09:40:26.103994] IPC Client (1766482975) connection to
> /x.x.x.x:25000 from syslog at HADOOP.COM: closed;
>
> Thank you.
>
>
> On Tue, Aug 21, 2018 at 6:48 PM Nagy, Gábor <gabor.nagy at oneidentity.com>
> wrote:
>
>> Hi,
>>
>> Can you tell ne what is the version of the hadoop lib you use with
>> syslog-ng, please?
>>
>> Can you share your syslog-ng configuration, mainly the hdfs part, please?
>>
>> Regards,
>> Gabor
>>
>> On Tue, Aug 21, 2018 at 4:15 AM Lee Keng Ket <kengket at gmail.com> wrote:
>>
>>> Hi, Gabor
>>>
>>> I have run it, seems like it stops at the HDFS side.
>>>
>>> [2018-08-21T10:07:51.212015] Worker thread started; driver='d_hdfs#0'
>>> [2018-08-21T10:07:51.212499] Running application hooks; hook='1'
>>> [2018-08-21T10:07:51.212516] Running application hooks; hook='3'
>>> [2018-08-21T10:07:51.212595] syslog-ng starting up; version='3.14.1'
>>> [2018-08-21T10:07:51.214113] Opening hdfs;
>>> [2018-08-21T10:08:01.215622] Opening hdfs;
>>> [2018-08-21T10:08:11.216050] Opening hdfs;
>>> [2018-08-21T10:08:21.226340] Opening hdfs;
>>> [2018-08-21T10:08:31.236589] Opening hdfs;
>>> [2018-08-21T10:08:41.240623] Opening hdfs;
>>> [2018-08-21T10:08:51.250879] Opening hdfs;
>>> [2018-08-21T10:09:01.261172] Opening hdfs;
>>> [2018-08-21T10:09:11.271410] Opening hdfs;
>>> [2018-08-21T10:09:21.281685] Opening hdfs;
>>> [2018-08-21T10:09:31.290765] Opening hdfs;
>>> [2018-08-21T10:09:41.301098] Opening hdfs;
>>> [2018-08-21T10:09:51.311362] Opening hdfs;
>>> [2018-08-21T10:10:01.321152] Opening hdfs;
>>> [2018-08-21T10:10:11.321818] Opening hdfs;
>>> [2018-08-21T10:10:21.330114] Opening hdfs;
>>> [2018-08-21T10:10:31.340413] Opening hdfs;
>>> [2018-08-21T10:10:41.350654] Opening hdfs;
>>> [2018-08-21T10:10:51.354016] Opening hdfs;
>>> [2018-08-21T10:11:01.364267] Opening hdfs;
>>> [2018-08-21T10:11:11.374516] Opening hdfs;
>>> [2018-08-21T10:11:21.384761] Opening hdfs;
>>> [2018-08-21T10:11:31.395017] Opening hdfs;
>>> [2018-08-21T10:11:41.402256] Opening hdfs;
>>> [2018-08-21T10:11:51.404097] Opening hdfs;
>>> ^C[2018-08-21T10:11:59.672252] syslog-ng shutting down; version='3.14.1'
>>> Exception in thread "" java.lang.NoClassDefFoundError:
>>> org/apache/hadoop/conf/Configuration
>>>         at
>>> org.syslog_ng.hdfs.HdfsDestination.open(HdfsDestination.java:92)
>>>         at org.syslog_ng.LogDestination.openProxy(LogDestination.java:65)
>>> [2018-08-21T10:11:59.774895] Worker thread finished; driver='d_hdfs#0'
>>> [2018-08-21T10:11:59.775384] Closing log transport fd; fd='13'
>>> [2018-08-21T10:11:59.775508] Deinitialize hdfs destination;
>>> [2018-08-21T10:11:59.776534] Java machine free;
>>> [2018-08-21T10:11:59.778421] Running application hooks; hook='4'
>>>
>>> Any idea what to be checked further?
>>>
>>> Thank you.
>>>
>>> On Fri, Aug 17, 2018 at 4:45 PM Nagy, Gábor <gabor.nagy at oneidentity.com>
>>> wrote:
>>>
>>>> Hello!
>>>>
>>>> In the statistics it can be seen that the log message is not sent to
>>>> the HDFS server:
>>>> dropped='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=0'
>>>> processed='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1'
>>>> queued='dst.java(d_hdfs#0 java_dst hdfs hdfs://x.x.x.x:25000
>>>> /user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1'
>>>>
>>>> Well, generally on write error there should be an exception that
>>>> results in an error message.
>>>>
>>>> You should try debugging it either in running syslog-ng in foreground
>>>> (-F option), forwarding internal logs to stderr (-e) and with debug mode
>>>> (-dv) on.
>>>> Or in service mode use the internal() source in your config and connect
>>>> it to a destination (e.g. file()) which you prefer.
>>>>
>>>> You could turn on debug messages on java side too using jvm_options()
>>>> in syslog-ng config and configuring the log4j logging service, e.g.:
>>>> options {
>>>>
>>>> jvm_options("-Dlog4j.configuration=file:/etc/hadoop/log4j.properties
>>>> -Dlog4j.debug=true");
>>>> };
>>>>
>>>> Regards,
>>>> Gabor
>>>>
>>>> On Fri, Aug 17, 2018 at 10:34 AM Czanik, Péter <
>>>> peter.czanik at balabit.com> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> As https://copr.fedorainfracloud.org/coprs/czanik/syslog-ng314/ also
>>>>> writes: "Java is enabled, but JAR dependencies are not provided in package,
>>>>> except for Elasticsearch http mode." The syslog-ng-java-deps.noarch
>>>>> contains build time dependencies. Probably I should rename the package to
>>>>> syslog-ng-java-build-deps...
>>>>>
>>>>> Check the documentation at
>>>>> https://www.syslog-ng.com/technical-documents/doc/syslog-ng-open-source-edition/3.16/administration-guide/33#TOPIC-956506
>>>>> on how to download and configure HDFS related JAR dependencies.
>>>>>
>>>>> Bye,
>>>>>
>>>>> Peter Czanik (CzP) <peter.czanik at balabit.com>
>>>>> Balabit / syslog-ng upstream
>>>>> https://syslog-ng.com/community/
>>>>> https://twitter.com/PCzanik
>>>>>
>>>>> On Fri, Aug 17, 2018 at 10:22 AM, Lee Keng Ket <kengket at gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I'm trying to connect syslog-ng 3.14.1 to HDFS to store the syslog
>>>>>> messages. The syslog-ng can start without error, and it's able to write
>>>>>> into local file. However, the log is not written to the HDFS. As there is
>>>>>> no single error, I'm not sure how I should troubleshoot on this.
>>>>>>
>>>>>> I have installed the syslog-ng from this repo,
>>>>>> https://copr.fedorainfracloud.org/coprs/czanik/syslog-ng314/repo/epel-7/czanik-syslog-ng314-epel-7.repo
>>>>>>
>>>>>> Installed Packages
>>>>>> syslog-ng.x86_64
>>>>>> 3.14.1-4.el7.centos
>>>>>> @czanik-syslog-ng314
>>>>>> syslog-ng-java.x86_64
>>>>>> 3.14.1-4.el7.centos
>>>>>> @czanik-syslog-ng314
>>>>>> syslog-ng-java-deps.noarch
>>>>>> 1.0-2
>>>>>> @czanik-syslog-ng314
>>>>>>
>>>>>> This is the message from /var/log/message:
>>>>>> Log statistics; processed='src.internal(s_sys#0)=1',
>>>>>> stamp='src.internal(s_sys#0)=1534491834',
>>>>>> processed='destination(d_spol)=0', processed='destination(d_mlal)=0',
>>>>>> processed='center(received)=2', processed='destination(d_mesg)=1',
>>>>>> processed='destination(d_mail)=0', processed='destination(d_auth)=0',
>>>>>> processed='destination(d_cron)=0', processed='destination(d_hdfs)=1',
>>>>>> processed='center(queued)=3', queued='global(scratch_buffers_count)=0',
>>>>>> processed='source(remote_log)=1',
>>>>>> dropped='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=0',
>>>>>> processed='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1',
>>>>>> queued='dst.java(d_hdfs#0,java_dst,hdfs,hdfs://x.x.x.x:25000,/user/syslog/$HOST-$DAY-$MONTH-$YEAR.log)=1',
>>>>>> processed='global(payload_reallocs)=0',
>>>>>> processed='src.journald(journal)=0', stamp='src.journald(journal)=0',
>>>>>> processed='global(sdata_updates)=0',
>>>>>> queued='global(scratch_buffers_bytes)=0',
>>>>>> processed='destination(d_boot)=0', processed='destination(d_kern)=0',
>>>>>> processed='source(s_sys)=1', processed='destination(remote)=1',
>>>>>> processed='global(internal_queue_length)=0',
>>>>>> processed='global(msg_clones)=0'
>>>>>>
>>>>>> Anyone has any idea how should I proceed the troubleshooting?
>>>>>>
>>>>>>
>>>>>> ______________________________________________________________________________
>>>>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>>>>> Documentation:
>>>>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>>>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> ______________________________________________________________________________
>>>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>>>> Documentation:
>>>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>>>
>>>>>
>>>> ______________________________________________________________________________
>>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>>> Documentation:
>>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>>
>>>>
>>> ______________________________________________________________________________
>>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>>> Documentation:
>>> http://www.balabit.com/support/documentation/?product=syslog-ng
>>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>>
>>>
>> ______________________________________________________________________________
>> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
>> Documentation:
>> http://www.balabit.com/support/documentation/?product=syslog-ng
>> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>>
>>
> ______________________________________________________________________________
> Member info: https://lists.balabit.hu/mailman/listinfo/syslog-ng
> Documentation:
> http://www.balabit.com/support/documentation/?product=syslog-ng
> FAQ: http://www.balabit.com/wiki/syslog-ng-faq
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.balabit.hu/pipermail/syslog-ng/attachments/20180824/c1ac64b2/attachment-0001.html>


More information about the syslog-ng mailing list