This is a Splunk Add-On for receiving data via a number of different data protocols.
Yes we do. And by all means use those. But if you want to perform some custom data handling and pre-processing of the received data before it gets indexed (above and beyond what you can accomplish using Splunk conf files) , then this Modular Input presents another option for you.
Furthermore , this Modular Input also implements several other protocols for sending data to Splunk.
This Modular Input utilizes VERTX.IO version 2.1.4 under the hood.http://vertx.io/vertx2/manual.html#what-is-vertx.
This framework provides for an implementation that is :
The way in which the Modular Input processes the received raw data is entirely pluggable with custom implementations should you wish.
This allows you to :
To do this you code a Vertx "Verticle" to handle the received data. http://vertx.io/vertx2/manual.html#verticle
These data handlers can be written in numerous JVM languages. http://vertx.io/vertx2/manual.html#polyglot
You then place the handler in the $SPLUNKHOME/etc/apps/protocolta/bin/datahandlers directory.
On the Splunk config screen for the Modular Input there is a field where you can then specify the name of this handler to be applied.
If you don't need a custom handler then the default handler com.splunk.modinput.protocolverticle.DefaultHandlerVerticle will be used.
To get started , you can refer to the default handler examples in the datahandlers directory.
Note : experimental Nashorn support is included for js and coffee (requires Java 8). To use the Nashorn JS/Coffee engine rather than the default Rhino engine , then edit protocolta/bin/vertxconf/langs.properties
This is provisioned using your own Java Keystore that you can create using the keytool utility that is part of the JDK.
Refer to http://vertx.io/vertx2/coremanualjava.html#ssl-servers
Client certificate based authentication can be enabled for the TLS/SSL channels you setup.
Any required Vertx modules , such as various language modules for the polyglot functionality (JS , Scala , Groovy etc...) will be dynamically downloaded from online repositorys and installed in your protocolta/bin/vertxmodules directory.
You can edit your repository locations in protocolta/bin/vertxconf/repos.txt
Due to the nature of the async/event driven/non blocking architecture , the out of the box default settings may just well suffice for you.
But there are some other parameters that you can tune to take more advantage of your underlying computing resource(ie: cpu cores) available to you.
These are the "serververticleinstances" and "handlerverticleinstances" params.
Refer to http://vertx.io/vertx2/coremanualjava.html#specifying-number-of-instances for an explanation of how increasing the number of instances may help you.
You can also tune the TCP accept queue settings (also requires OS tweaks) , particularly if you are receiving lots of connections within a short time span.
Refer to http://vertx.io/vertx2/manual.html#improving-connection-time
By default data will be output to STDOUT in Modular Input Stream XML format.
However you can bypass this if you wish and declare that data is output to a Splunk TCP port or via Splunk's HTTP Event Collector.
Settings -> Data Inputs -> Protocol Data Inputsto add a new Input stanza via the UI
inputs.conffile should be placed in a
localdirectory under an App or User context.
You require an activation key to use this App. Visit http://www.baboonbones.com/#activation to obtain a non-expiring key
Any log entries/errors will get written to $SPLUNK_HOME/var/log/splunk/splunkd.log
These are also searchable in Splunk :
index=_internal error protocol.py
The default heap maximum is 256MB. If you require a larger heap, then you can alter this in $SPLUNKHOME/etc/apps/protocolta/bin/protocol.py on line 95
You can declare custom JVM System Properties when setting up new input stanzas. Note : these JVM System Properties will apply to the entire JVM context and all stanzas you have setup
This project was initiated by Damien Dallimore , email@example.com