This plugin enables Drill to query Splunk.
Option | Default | Description | Since |
---|---|---|---|
type | (none) | Set to "splunk" to use this plugin | 1.19 |
username | null | Splunk username to be used by Drill | 1.19 |
password | null | Splunk password to be used by Drill | 1.19 |
scheme | https | The scheme with which to access the Splunk host | 2.0 |
hostname | localhost | Splunk host to be queried by Drill | 1.19 |
port | 8089 | TCP port over which Drill will connect to Splunk | 1.19 |
earliestTime | null | Global earliest record timestamp default | 1.19 |
latestTime | null | Global latest record timestamp default | 1.19 |
app | null | The application context of the service1 | 1.21 |
owner | null | The owner context of the service1 | 1.21 |
token | null | A Splunk authentication token to use for the session2 | 1.21 |
cookie | null | A valid login cookie | 1.21 |
validateCertificates | true | Whether the Splunk client will validates the server's SSL cert | 1.21 |
validateHostname | true | Whether the Splunk client will validate the server's host name | 1.22 |
maxColumns | 1024 | The maximum number of columns Drill will accept from Splunk | 1.22 |
maxCacheSize | 10000 | The size (in bytes) of Splunk's schema cache. | 1.22 |
cacheExpiration | 1024 | The number of minutes for Drill to persist the schema cache. | 1.22 |
To connect Drill to Splunk, create a new storage plugin with the following configuration:
{
"type":"splunk",
"enabled": false,
"username": "admin",
"password": "changeme",
"hostname": "localhost",
"port": 8089,
"earliestTime": "-14d",
"latestTime": "now",
"reconnectRetries": 3
}
To successfully connect, Splunk uses port 8089
for interfaces. This port must be open for Drill to query Splunk.
Sometimes Splunk has issue in connection to it:
splunk/splunk-sdk-java#62
To bypass it by Drill please specify "reconnectRetries": 3. It allows you to retry the connection several times.
The Splunk plugin supports user translation. Simply set the authMode
parameter to USER_TRANSLATION
and use either the plain or vault credential provider for credentials.
For every query that you send to Splunk from Drill, Drill will have to pull schema information from Splunk. If you have a lot of indexes, this process can cause slow planning time. To improve planning time, you can configure Drill to cache the index names so that it does not need to make additional calls to Splunk.
There are two configuration parameters for the schema caching: maxCacheSize
and cacheExpiration
. The maxCacheSize defaults to 10k bytes and the cacheExpiration
defaults to 1024 minutes. To disable schema caching simply set the cacheExpiration
parameter to a value less than zero.
Splunk's primary use case is analyzing event logs with a timestamp. As such, data is indexed by the timestamp, with the most recent data being indexed first. By default, Splunk will sort the data in reverse chronological order. Large Splunk installations will put older data into buckets of hot, warm and cold storage with the "cold" storage on the slowest and cheapest disks.
With this understood, it is very important to put time boundaries on your Splunk queries. The Drill plugin allows you to set default values in the configuration such that every query you run will be bounded by these boundaries. Alternatively, you can set the time boundaries at query time. In either case, you will achieve the best performance when you are asking Splunk for the smallest amount of data possible.
Drill treats Splunk indexes as tables. Splunk's access model does not restrict to the catalog, but does restrict access to the actual data. It is therefore possible that you can
see the names of indexes to which you do not have access. You can view the list of available indexes with a SHOW TABLES IN splunk
query.
apache drill> SHOW TABLES IN splunk;
+--------------+----------------+
| TABLE_SCHEMA | TABLE_NAME |
+--------------+----------------+
| splunk | summary |
| splunk | splunklogger |
| splunk | _configtracker |
| splunk | _thefishbucket |
| splunk | _audit |
| splunk | _internal |
| splunk | _introspection |
| splunk | main |
| splunk | history |
| splunk | _telemetry |
+--------------+----------------+
9 rows selected (0.304 seconds)
To query Splunk from Drill, use the following format:
SELECT <fields>
FROM splunk.<index>
When you learn to query Splunk via their interface, the first thing you learn is to bound your queries so that they are looking at the shortest time span possible. When using Drill to query Splunk, it is advisable to do the same thing, and Drill offers two ways to accomplish this: via the configuration and at query time.
The easiest way to bound your query is to do so at querytime via special filters in the WHERE
clause. There are two special fields, earliestTime
and latestTime
which can
be set to bound the query. If they are not set, the query will be bounded to the defaults set in the configuration.
You can use any of the time formats specified in the Splunk documentation here: https://docs.splunk.com/Documentation/Splunk/8.0.3/SearchReference/SearchTimeModifiers
So if you wanted to see your data for the last 15 minutes, you could execute the following query:
SELECT <fields>
FROM splunk.<index>
WHERE earliestTime='-15m' AND latestTime='now'
The variables set in a query override the defaults from the configuration.
Splunk does not have sophisticated data types and unfortunately does not provide metadata from its query results. With the exception of the fields below, Drill will interpret
all fields as VARCHAR
and hence you will have to convert them to the appropriate data type at query time.
_indextime
_time
date_hour
date_mday
date_minute
date_second
date_year
linecount
Splunk has two different types of nested data which roughly map to Drill's LIST
and MAP
data types. Unfortunately, there is no easy way to identify whether a field is a
nested field at querytime as Splunk does not provide any metadata and therefore all fields are treated as VARCHAR
.
However, Drill does have built in functions to easily convert Splunk multifields into Drill LIST
and MAP
data types. For a LIST, simply use the
SPLIT(<field>, ' ')
function to split the field into a LIST
.
MAP
data types are rendered as JSON in Splunk. Fortunately JSON can easily be parsed into a Drill Map by using the convert_fromJSON()
function. The query below
demonstrates how to convert a JSON column into a Drill MAP
.
SELECT convert_fromJSON(_raw)
FROM splunk.spl
WHERE spl = '| makeresults
| eval _raw="{\"pc\":{\"label\":\"PC\",\"count\":24,\"peak24\":12},\"ps3\":
{\"label\":\"PS3\",\"count\":51,\"peak24\":10},\"xbox\":
{\"label\":\"XBOX360\",\"count\":40,\"peak24\":11},\"xone\":
{\"label\":\"XBOXONE\",\"count\":105,\"peak24\":99},\"ps4\":
{\"label\":\"PS4\",\"count\":200,\"peak24\":80}}"'
When you execute a query in Drill for Splunk, the fields you select are pushed down to Splunk. Therefore, it will always be more efficient to explicitly specify fields to push
down to Splunk rather than using SELECT *
queries.
There are several fields which can be included in a Drill query
spl
: If you just want to send an SPL query to Splunk, this will do that.earliestTime
: Overrides theearliestTime
setting in the configuration.latestTime
: Overrides thelatestTime
setting in the configuration.
Due to the nature of Splunk indexes, data will always be returned in reverse chronological order. Thus, sorting is not necessary if that is the desired order.
There is a special table called spl
which you can use to send arbitrary queries to Splunk. If you use this table, you must include a query in the spl
filter as shown below:
SELECT *
FROM splunk.spl
WHERE spl='<your SPL query'
As of Drill version 2.0 you can write and append data to Splunk. All fields are sent as strings, to include complex objects.
You can create a new index with a CREATE TABLE splunk.new_index AS ...
query. Likewise, you can append to an existing index with an INSERT INTO splunk.index SELECT...
query.
This plugin includes a series of unit tests in the src/test/
directory, however there are a few tests for which you will need an active Splunk installation to run them.
Simply follow the instructions below to test Splunk with Drill.
From Splunk's website, simply download and install the free version here: https://www.splunk.com/en_us/download/splunk-enterprise.html
Once you've downloaded Splunk, start it up, and make sure everything is working properly.
Next, go here: https://docs.splunk.com/Documentation/Splunk/7.0.3/SearchTutorial/Systemrequirements and download the dummy datasets that Splunk provides. Once you've downloaded this data, have Splunk index this data and you're ready to go from the Splunk end.
- At present, Drill will not interpret Splunk multifields as anything other than a String. If there is interest, this feature can be implemented.
Footnotes
-
See this Splunk documentation for more information. ↩ ↩2
-
See this Splunk documentation for more information. ↩