You require an activation key to use these libraries. Visit http://www.baboonbones.com/#activation to obtain a non-expiring key
The purpose of this project is to create a logging framework to allow developers to as seamlessly as possible integrate Splunk best practice logging semantics into their code and easily send their log events to Splunk. There are also custom handler/appender implementations and config examples for the most prevalent Java logging frameworks in play.
This framework contains :
If you want to use UDP to send events to Splunk , then Log4j 1.x and Logback already have Syslog Appenders. Log4j 2 has a UDP Appender and Syslog Appender. And of course you can still use any File appenders and have the file monitored by a Splunk Universal Forwarder.
I generally recommend using the raw TCP or HEC handlers/appenders I have provided , they perform the best, and have features coded into them for auto connection re-establishment and configurable buffering of log events which will get flushed upon reconnection.
Log4j 2 and Log4j 1.x are very distinct from one another. Logback was actually the "new version" of Log4j 1.x , and then Log4J 2 attempted to improve upon Logback. This rather convoluted family tree has essentially transpired with 3 different logging frameworks in play, each with different characteristics. Log4j 1.x still has a very large legacy usage base in enterprise software therefore warrants addressing with its own custom appenders and example configurations.
I always advocate the best practice of using a Splunk Universal Forwarder(UF) monitoring local files wherever possible. Not only do you get the features inherent in the UF, but you get the added resiliency of the persistence of files. However, there are going to be situations where, for whatever reason(technical or bureaucratic), that a UF can not be deployed.In this case, Splunk Java Logging can be used to forward events to Splunk. Furthermore, in either scenario, you can still utilize the SplunkLogEvent class to construct your log events in best practice semantic format.
The HTTP REST ,Raw TCP and HEC handler/appenders have autonomous socket reconnection logic in case of connection failures. There is also internal event queuing that is loosely modelled off Splunk's outputs.conf for Universal Forwarders. You can set these propertys : * maxQueueSize : defaults to 500KB , format [integer|integer[KB|MB|GB]] * dropEventsOnQueueFull : defaults to false , format [ true | false]
And you can use a parallel File appender if you absolutely need disk persistence.
If you want "data cloning" functionality, then you can leverage the logging configuration and have (n) different appender definitions for your various target Indexers.
If you wish to have load balancing of your log events, then configure your logging appenders to send to a Splunk Universal Forwarder acting as a load balancing intermediary before you Indexer Cluster.
Log4J 2 has a Failover appender you can use : http://logging.apache.org/log4j/2.x/manual/appenders.html#FailoverAppender There is an example in config/log4j2.xml
Log4J 2 has a Routing appender you can use : http://logging.apache.org/log4j/2.x/manual/appenders.html#RoutingAppender
Log4j and Logback are thread safe.
If you haven't already installed Splunk, download it here: http://www.splunk.com/download.