# Copyright 2011-2014 Splunk, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"): you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # # The purpose of this module is to provide a friendlier domain interface to # various Splunk endpoints. The approach here is to leverage the binding # layer to capture endpoint context and provide objects and methods that # offer simplified access their corresponding endpoints. The design avoids # caching resource state. From the perspective of this module, the 'policy' # for caching resource state belongs in the application or a higher level # framework, and its the purpose of this module to provide simplified # access to that resource state. # # A side note, the objects below that provide helper methods for updating eg: # Entity state, are written so that they may be used in a fluent style. # """The **splunklib.client** module provides a Pythonic interface to the `Splunk REST API `_, allowing you programmatically access Splunk's resources. **splunklib.client** wraps a Pythonic layer around the wire-level binding of the **splunklib.binding** module. The core of the library is the :class:`Service` class, which encapsulates a connection to the server, and provides access to the various aspects of Splunk's functionality, which are exposed via the REST API. Typically you connect to a running Splunk instance with the :func:`connect` function:: import splunklib.client as client service = client.connect(host='localhost', port=8089, username='admin', password='...') assert isinstance(service, client.Service) :class:`Service` objects have fields for the various Splunk resources (such as apps, jobs, saved searches, inputs, and indexes). All of these fields are :class:`Collection` objects:: appcollection = service.apps my_app = appcollection.create('my_app') my_app = appcollection['my_app'] appcollection.delete('my_app') The individual elements of the collection, in this case *applications*, are subclasses of :class:`Entity`. An ``Entity`` object has fields for its attributes, and methods that are specific to each kind of entity. For example:: print my_app['author'] # Or: print my_app.author my_app.package() # Creates a compressed package of this application """ import datetime import json import urllib import logging from time import sleep from datetime import datetime, timedelta import socket import contextlib from binding import Context, HTTPError, AuthenticationError, namespace, UrlEncoded, _encode from data import record import data __all__ = [ "connect", "NotSupportedError", "OperationError", "IncomparableException", "Service", "namespace" ] PATH_APPS = "apps/local/" PATH_CAPABILITIES = "authorization/capabilities/" PATH_CONF = "configs/conf-%s/" PATH_PROPERTIES = "properties/" PATH_DEPLOYMENT_CLIENTS = "deployment/client/" PATH_DEPLOYMENT_TENANTS = "deployment/tenants/" PATH_DEPLOYMENT_SERVERS = "deployment/server/" PATH_DEPLOYMENT_SERVERCLASSES = "deployment/serverclass/" PATH_EVENT_TYPES = "saved/eventtypes/" PATH_FIRED_ALERTS = "alerts/fired_alerts/" PATH_INDEXES = "data/indexes/" PATH_INPUTS = "data/inputs/" PATH_JOBS = "search/jobs/" PATH_LOGGER = "server/logger/" PATH_MESSAGES = "messages/" PATH_MODULAR_INPUTS = "data/modular-inputs" PATH_ROLES = "authorization/roles/" PATH_SAVED_SEARCHES = "saved/searches/" PATH_STANZA = "configs/conf-%s/%s" # (file, stanza) PATH_USERS = "authentication/users/" PATH_RECEIVERS_STREAM = "receivers/stream" PATH_RECEIVERS_SIMPLE = "receivers/simple" XNAMEF_ATOM = "{http://www.w3.org/2005/Atom}%s" XNAME_ENTRY = XNAMEF_ATOM % "entry" XNAME_CONTENT = XNAMEF_ATOM % "content" MATCH_ENTRY_CONTENT = "%s/%s/*" % (XNAME_ENTRY, XNAME_CONTENT) class IllegalOperationException(Exception): """Thrown when an operation is not possible on the Splunk instance that a :class:`Service` object is connected to.""" pass class IncomparableException(Exception): """Thrown when trying to compare objects (using ``==``, ``<``, ``>``, and so on) of a type that doesn't support it.""" pass class AmbiguousReferenceException(ValueError): """Thrown when the name used to fetch an entity matches more than one entity.""" pass class InvalidNameException(Exception): """Thrown when the specified name contains characters that are not allowed in Splunk entity names.""" pass class NoSuchCapability(Exception): """Thrown when the capability that has been referred to doesn't exist.""" pass class OperationError(Exception): """Raised for a failed operation, such as a time out.""" pass class NotSupportedError(Exception): """Raised for operations that are not supported on a given object.""" pass def _trailing(template, *targets): """Substring of *template* following all *targets*. **Example**:: template = "this is a test of the bunnies." _trailing(template, "is", "est", "the") == " bunnies" Each target is matched successively in the string, and the string remaining after the last target is returned. If one of the targets fails to match, a ValueError is raised. :param template: Template to extract a trailing string from. :type template: ``string`` :param targets: Strings to successively match in *template*. :type targets: list of ``string``s :return: Trailing string after all targets are matched. :rtype: ``string`` :raises ValueError: Raised when one of the targets does not match. """ s = template for t in targets: n = s.find(t) if n == -1: raise ValueError("Target " + t + " not found in template.") s = s[n + len(t):] return s # Filter the given state content record according to the given arg list. def _filter_content(content, *args): if len(args) > 0: return record((k, content[k]) for k in args) return record((k, v) for k, v in content.iteritems() if k not in ['eai:acl', 'eai:attributes', 'type']) # Construct a resource path from the given base path + resource name def _path(base, name): if not base.endswith('/'): base = base + '/' return base + name # Load an atom record from the body of the given response def _load_atom(response, match=None): return data.load(response.body.read(), match) # Load an array of atom entries from the body of the given response def _load_atom_entries(response): r = _load_atom(response) if 'feed' in r: # Need this to handle a random case in the REST API if r.feed.get('totalResults') == 0: return [] entries = r.feed.get('entry', None) if entries is None: return None return entries if isinstance(entries, list) else [entries] # Unlike most other endpoints, the jobs endpoint does not return # its state wrapped in another element, but at the top level. # For example, in XML, it returns ... instead of # .... else: entries = r.get('entry', None) if entries is None: return None return entries if isinstance(entries, list) else [entries] # Load the sid from the body of the given response def _load_sid(response): return _load_atom(response).response.sid # Parse the given atom entry record into a generic entity state record def _parse_atom_entry(entry): title = entry.get('title', None) elink = entry.get('link', []) elink = elink if isinstance(elink, list) else [elink] links = record((link.rel, link.href) for link in elink) # Retrieve entity content values content = entry.get('content', {}) # Host entry metadata metadata = _parse_atom_metadata(content) # Filter some of the noise out of the content record content = record((k, v) for k, v in content.iteritems() if k not in ['eai:acl', 'eai:attributes', 'type']) return record({ 'title': title, 'links': links, 'access': metadata.access, 'fields': metadata.fields, 'content': content }) # Parse the metadata fields out of the given atom entry content record def _parse_atom_metadata(content): # Hoist access metadata access = content.get('eai:acl', None) # Hoist content metadata (and cleanup some naming) attributes = content.get('eai:attributes', {}) fields = record({ 'required': attributes.get('requiredFields', []), 'optional': attributes.get('optionalFields', []), 'wildcard': attributes.get('wildcardFields', [])}) return record({'access': access, 'fields': fields}) # kwargs: scheme, host, port, app, owner, username, password def connect(**kwargs): """This function connects and logs in to a Splunk instance. This function is a shorthand for :meth:`Service.login`. The ``connect`` function makes one round trip to the server (for logging in). :param host: The host name (the default is "localhost"). :type host: ``string`` :param port: The port number (the default is 8089). :type port: ``integer`` :param scheme: The scheme for accessing the service (the default is "https"). :type scheme: "https" or "http" :param `owner`: The owner context of the namespace (optional). :type owner: ``string`` :param `app`: The app context of the namespace (optional). :type app: ``string`` :param sharing: The sharing mode for the namespace (the default is "user"). :type sharing: "global", "system", "app", or "user" :param `token`: The current session token (optional). Session tokens can be shared across multiple service instances. :type token: ``string`` :param autologin: When ``True``, automatically tries to log in again if the session terminates. :type autologin: ``boolean`` :param `username`: The Splunk account username, which is used to authenticate the Splunk instance. :type username: ``string`` :param `password`: The password for the Splunk account. :type password: ``string`` :return: An initialized :class:`Service` connection. **Example**:: import splunklib.client as client s = client.connect(...) a = s.apps["my_app"] ... """ return Service(**kwargs).login() # In preparation for adding Storm support, we added an # intermediary class between Service and Context. Storm's # API is not going to be the same as enterprise Splunk's # API, so we will derive both Service (for enterprise Splunk) # and StormService for (Splunk Storm) from _BaseService, and # put any shared behavior on it. class _BaseService(Context): pass class Service(_BaseService): """A Pythonic binding to Splunk instances. A :class:`Service` represents a binding to a Splunk instance on an HTTP or HTTPS port. It handles the details of authentication, wire formats, and wraps the REST API endpoints into something more Pythonic. All of the low-level operations on the instance from :class:`splunklib.binding.Context` are also available in case you need to do something beyond what is provided by this class. After creating a ``Service`` object, you must call its :meth:`login` method before you can issue requests to Splunk. Alternately, use the :func:`connect` function to create an already authenticated :class:`Service` object, or provide a session token when creating the :class:`Service` object explicitly (the same token may be shared by multiple :class:`Service` objects). :param host: The host name (the default is "localhost"). :type host: ``string`` :param port: The port number (the default is 8089). :type port: ``integer`` :param scheme: The scheme for accessing the service (the default is "https"). :type scheme: "https" or "http" :param `owner`: The owner context of the namespace (optional; use "-" for wildcard). :type owner: ``string`` :param `app`: The app context of the namespace (optional; use "-" for wildcard). :type app: ``string`` :param `token`: The current session token (optional). Session tokens can be shared across multiple service instances. :type token: ``string`` :param `username`: The Splunk account username, which is used to authenticate the Splunk instance. :type username: ``string`` :param `password`: The password, which is used to authenticate the Splunk instance. :type password: ``string`` :return: A :class:`Service` instance. **Example**:: import splunklib.client as client s = client.Service(username="boris", password="natasha", ...) s.login() # Or equivalently s = client.connect(username="boris", password="natasha") # Or if you already have a session token s = client.Service(token="atg232342aa34324a") """ def __init__(self, **kwargs): super(Service, self).__init__(**kwargs) self._splunk_version = None @property def apps(self): """Returns the collection of applications that are installed on this instance of Splunk. :return: A :class:`Collection` of :class:`Application` entities. """ return Collection(self, PATH_APPS, item=Application) @property def confs(self): """Returns the collection of configuration files for this Splunk instance. :return: A :class:`Configurations` collection of :class:`ConfigurationFile` entities. """ return Configurations(self) @property def capabilities(self): """Returns the list of system capabilities. :return: A ``list`` of capabilities. """ response = self.get(PATH_CAPABILITIES) return _load_atom(response, MATCH_ENTRY_CONTENT).capabilities @property def event_types(self): """Returns the collection of event types defined in this Splunk instance. :return: An :class:`Entity` containing the event types. """ return Collection(self, PATH_EVENT_TYPES) @property def fired_alerts(self): """Returns the collection of alerts that have been fired on the Splunk instance, grouped by saved search. :return: A :class:`Collection` of :class:`AlertGroup` entities. """ return Collection(self, PATH_FIRED_ALERTS, item=AlertGroup) @property def indexes(self): """Returns the collection of indexes for this Splunk instance. :return: An :class:`Indexes` collection of :class:`Index` entities. """ return Indexes(self, PATH_INDEXES, item=Index) @property def info(self): """Returns the information about this instance of Splunk. :return: The system information, as key-value pairs. :rtype: ``dict`` """ response = self.get("server/info") return _filter_content(_load_atom(response, MATCH_ENTRY_CONTENT)) @property def inputs(self): """Returns the collection of inputs configured on this Splunk instance. :return: An :class:`Inputs` collection of :class:`Input` entities. """ return Inputs(self) @property def jobs(self): """Returns the collection of current search jobs. :return: A :class:`Jobs` collection of :class:`Job` entities. """ return Jobs(self) @property def loggers(self): """Returns the collection of logging level categories and their status. :return: A :class:`Loggers` collection of logging levels. """ return Loggers(self) @property def messages(self): """Returns the collection of service messages. :return: A :class:`Collection` of :class:`Message` entities. """ return Collection(self, PATH_MESSAGES, item=Message) @property def modular_input_kinds(self): """Returns the collection of the modular input kinds on this Splunk instance. :return: A :class:`ReadOnlyCollection` of :class:`ModularInputKind` entities. """ if self.splunk_version >= (5,): return ReadOnlyCollection(self, PATH_MODULAR_INPUTS, item=ModularInputKind) else: raise IllegalOperationException("Modular inputs are not supported before Splunk version 5.") # kwargs: enable_lookups, reload_macros, parse_only, output_mode def parse(self, query, **kwargs): """Parses a search query and returns a semantic map of the search. :param query: The search query to parse. :type query: ``string`` :param kwargs: Arguments to pass to the ``search/parser`` endpoint (optional). Valid arguments are: * "enable_lookups" (``boolean``): If ``True``, performs reverse lookups to expand the search expression. * "output_mode" (``string``): The output format (XML or JSON). * "parse_only" (``boolean``): If ``True``, disables the expansion of search due to evaluation of subsearches, time term expansion, lookups, tags, eventtypes, and sourcetype alias. * "reload_macros" (``boolean``): If ``True``, reloads macro definitions from macros.conf. :type kwargs: ``dict`` :return: A semantic map of the parsed search query. """ return self.get("search/parser", q=query, **kwargs) def restart(self, timeout=None): """Restarts this Splunk instance. The service is unavailable until it has successfully restarted. If a *timeout* value is specified, ``restart`` blocks until the service resumes or the timeout period has been exceeded. Otherwise, ``restart`` returns immediately. :param timeout: A timeout period, in seconds. :type timeout: ``integer`` """ result = self.post("server/control/restart") if timeout is None: return result start = datetime.now() diff = timedelta(seconds=10) while datetime.now() - start < diff: try: self.login() # Has the server gone down yet? sleep(0.3) except Exception: break # Server is down. Move on. start = datetime.now() diff = timedelta(seconds=timeout) while datetime.now() - start < diff: try: self.login() # Awake yet? return result except Exception, e: sleep(2) raise Exception, "Operation timed out." @property def restart_required(self): """Indicates whether splunkd is in a state that requires a restart. :return: A ``boolean`` that indicates whether a restart is required. """ response = self.get("messages").body.read() messages = data.load(response)['feed'] if 'entry' not in messages: result = False else: if isinstance(messages['entry'], dict): titles = [messages['entry']['title']] else: titles = [x['title'] for x in messages['entry']] result = 'restart_required' in titles return result @property def roles(self): """Returns the collection of user roles. :return: A :class:`Roles` collection of :class:`Role` entities. """ return Roles(self) def search(self, query, **kwargs): """Runs a search using a search query and any optional arguments you provide, and returns a `Job` object representing the search. :param query: A search query. :type query: ``string`` :param kwargs: Arguments for the search (optional): * "output_mode" (``string``): Specifies the output format of the results. * "earliest_time" (``string``): Specifies the earliest time in the time range to search. The time string can be a UTC time (with fractional seconds), a relative time specifier (to now), or a formatted time string. * "latest_time" (``string``): Specifies the latest time in the time range to search. The time string can be a UTC time (with fractional seconds), a relative time specifier (to now), or a formatted time string. * "rf" (``string``): Specifies one or more fields to add to the search. :type kwargs: ``dict`` :rtype: class:`Job` :returns: An object representing the created job. """ return self.jobs.create(query, **kwargs) @property def saved_searches(self): """Returns the collection of saved searches. :return: A :class:`SavedSearches` collection of :class:`SavedSearch` entities. """ return SavedSearches(self) @property def settings(self): """Returns the configuration settings for this instance of Splunk. :return: A :class:`Settings` object containing configuration settings. """ return Settings(self) @property def splunk_version(self): """Returns the version of the splunkd instance this object is attached to. The version is returned as a tuple of the version components as integers (for example, `(4,3,3)` or `(5,)`). :return: A ``tuple`` of ``integers``. """ if self._splunk_version is None: self._splunk_version = tuple([int(p) for p in self.info['version'].split('.')]) return self._splunk_version @property def users(self): """Returns the collection of users. :return: A :class:`Users` collection of :class:`User` entities. """ return Users(self) class Endpoint(object): """This class represents individual Splunk resources in the Splunk REST API. An ``Endpoint`` object represents a URI, such as ``/services/saved/searches``. This class provides the common functionality of :class:`Collection` and :class:`Entity` (essentially HTTP GET and POST methods). """ def __init__(self, service, path): self.service = service self.path = path if path.endswith('/') else path + '/' def get(self, path_segment="", owner=None, app=None, sharing=None, **query): """Performs a GET operation on the path segment relative to this endpoint. This method is named to match the HTTP method. This method makes at least one roundtrip to the server, one additional round trip for each 303 status returned, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. If *owner*, *app*, and *sharing* are omitted, this method takes a default namespace from the :class:`Service` object for this :class:`Endpoint`. All other keyword arguments are included in the URL as query parameters. :raises AuthenticationError: Raised when the ``Service`` is not logged in. :raises HTTPError: Raised when an error in the request occurs. :param path_segment: A path segment relative to this endpoint. :type path_segment: ``string`` :param owner: The owner context of the namespace (optional). :type owner: ``string`` :param app: The app context of the namespace (optional). :type app: ``string`` :param sharing: The sharing mode for the namespace (optional). :type sharing: "global", "system", "app", or "user" :param query: All other keyword arguments, which are used as query parameters. :type query: ``string`` :return: The response from the server. :rtype: ``dict`` with keys ``body``, ``headers``, ``reason``, and ``status`` **Example**:: import splunklib.client s = client.service(...) apps = s.apps apps.get() == \\ {'body': ...a response reader object..., 'headers': [('content-length', '26208'), ('expires', 'Fri, 30 Oct 1998 00:00:00 GMT'), ('server', 'Splunkd'), ('connection', 'close'), ('cache-control', 'no-store, max-age=0, must-revalidate, no-cache'), ('date', 'Fri, 11 May 2012 16:30:35 GMT'), ('content-type', 'text/xml; charset=utf-8')], 'reason': 'OK', 'status': 200} apps.get('nonexistant/path') # raises HTTPError s.logout() apps.get() # raises AuthenticationError """ # self.path to the Endpoint is relative in the SDK, so passing # owner, app, sharing, etc. along will produce the correct # namespace in the final request. if path_segment.startswith('/'): path = path_segment else: path = self.service._abspath(self.path + path_segment, owner=owner, app=app, sharing=sharing) # ^-- This was "%s%s" % (self.path, path_segment). # That doesn't work, because self.path may be UrlEncoded. return self.service.get(path, owner=owner, app=app, sharing=sharing, **query) def post(self, path_segment="", owner=None, app=None, sharing=None, **query): """Performs a POST operation on the path segment relative to this endpoint. This method is named to match the HTTP method. This method makes at least one roundtrip to the server, one additional round trip for each 303 status returned, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. If *owner*, *app*, and *sharing* are omitted, this method takes a default namespace from the :class:`Service` object for this :class:`Endpoint`. All other keyword arguments are included in the URL as query parameters. :raises AuthenticationError: Raised when the ``Service`` is not logged in. :raises HTTPError: Raised when an error in the request occurs. :param path_segment: A path segment relative to this endpoint. :type path_segment: ``string`` :param owner: The owner context of the namespace (optional). :type owner: ``string`` :param app: The app context of the namespace (optional). :type app: ``string`` :param sharing: The sharing mode of the namespace (optional). :type sharing: ``string`` :param query: All other keyword arguments, which are used as query parameters. :type query: ``string`` :return: The response from the server. :rtype: ``dict`` with keys ``body``, ``headers``, ``reason``, and ``status`` **Example**:: import splunklib.client s = client.service(...) apps = s.apps apps.post(name='boris') == \\ {'body': ...a response reader object..., 'headers': [('content-length', '2908'), ('expires', 'Fri, 30 Oct 1998 00:00:00 GMT'), ('server', 'Splunkd'), ('connection', 'close'), ('cache-control', 'no-store, max-age=0, must-revalidate, no-cache'), ('date', 'Fri, 11 May 2012 18:34:50 GMT'), ('content-type', 'text/xml; charset=utf-8')], 'reason': 'Created', 'status': 201} apps.get('nonexistant/path') # raises HTTPError s.logout() apps.get() # raises AuthenticationError """ if path_segment.startswith('/'): path = path_segment else: path = self.service._abspath(self.path + path_segment, owner=owner, app=app, sharing=sharing) return self.service.post(path, owner=owner, app=app, sharing=sharing, **query) # kwargs: path, app, owner, sharing, state class Entity(Endpoint): """This class is a base class for Splunk entities in the REST API, such as saved searches, jobs, indexes, and inputs. ``Entity`` provides the majority of functionality required by entities. Subclasses only implement the special cases for individual entities. For example for deployment serverclasses, the subclass makes whitelists and blacklists into Python lists. An ``Entity`` is addressed like a dictionary, with a few extensions, so the following all work:: ent['email.action'] ent['disabled'] ent['whitelist'] Many endpoints have values that share a prefix, such as ``email.to``, ``email.action``, and ``email.subject``. You can extract the whole fields, or use the key ``email`` to get a dictionary of all the subelements. That is, ``ent['email']`` returns a dictionary with the keys ``to``, ``action``, ``subject``, and so on. If there are multiple levels of dots, each level is made into a subdictionary, so ``email.body.salutation`` can be accessed at ``ent['email']['body']['salutation']`` or ``ent['email.body.salutation']``. You can also access the fields as though they were the fields of a Python object, as in:: ent.email.action ent.disabled ent.whitelist However, because some of the field names are not valid Python identifiers, the dictionary-like syntax is preferrable. The state of an :class:`Entity` object is cached, so accessing a field does not contact the server. If you think the values on the server have changed, call the :meth:`Entity.refresh` method. """ # Not every endpoint in the API is an Entity or a Collection. For # example, a saved search at saved/searches/{name} has an additional # method saved/searches/{name}/scheduled_times, but this isn't an # entity in its own right. In these cases, subclasses should # implement a method that uses the get and post methods inherited # from Endpoint, calls the _load_atom function (it's elsewhere in # client.py, but not a method of any object) to read the # information, and returns the extracted data in a Pythonesque form. # # The primary use of subclasses of Entity is to handle specially # named fields in the Entity. If you only need to provide a default # value for an optional field, subclass Entity and define a # dictionary ``defaults``. For instance,:: # # class Hypothetical(Entity): # defaults = {'anOptionalField': 'foo', # 'anotherField': 'bar'} # # If you have to do more than provide a default, such as rename or # actually process values, then define a new method with the # ``@property`` decorator. # # class Hypothetical(Entity): # @property # def foobar(self): # return self.content['foo'] + "-" + self.content["bar"] # Subclasses can override defaults the default values for # optional fields. See above. defaults = {} def __init__(self, service, path, **kwargs): Endpoint.__init__(self, service, path) self._state = None if not kwargs.get('skip_refresh', False): self.refresh(kwargs.get('state', None)) # "Prefresh" def __contains__(self, item): try: self[item] return True except KeyError: return False def __eq__(self, other): """Raises IncomparableException. Since Entity objects are snapshots of times on the server, no simple definition of equality will suffice beyond instance equality, and instance equality leads to strange situations such as:: import splunklib.client as client c = client.connect(...) saved_searches = c.saved_searches x = saved_searches['asearch'] but then ``x != saved_searches['asearch']``. whether or not there was a change on the server. Rather than try to do something fancy, we simple declare that equality is undefined for Entities. Makes no roundtrips to the server. """ raise IncomparableException( "Equality is undefined for objects of class %s" % \ self.__class__.__name__) def __getattr__(self, key): # Called when an attribute was not found by the normal method. In this # case we try to find it in self.content and then self.defaults. if key in self.state.content: return self.state.content[key] elif key in self.defaults: return self.defaults[key] else: raise AttributeError(key) def __getitem__(self, key): # getattr attempts to find a field on the object in the normal way, # then calls __getattr__ if it cannot. return getattr(self, key) # Load the Atom entry record from the given response - this is a method # because the "entry" record varies slightly by entity and this allows # for a subclass to override and handle any special cases. def _load_atom_entry(self, response): elem = _load_atom(response, XNAME_ENTRY) if isinstance(elem, list): raise AmbiguousReferenceException("Fetch from server returned multiple entries for name %s." % self.name) else: return elem.entry # Load the entity state record from the given response def _load_state(self, response): entry = self._load_atom_entry(response) return _parse_atom_entry(entry) def _run_action(self, path_segment, **kwargs): """Run a method and return the content Record from the returned XML. A method is a relative path from an Entity that is not itself an Entity. _run_action assumes that the returned XML is an Atom field containing one Entry, and the contents of Entry is what should be the return value. This is right in enough cases to make this method useful. """ response = self.get(path_segment, **kwargs) data = self._load_atom_entry(response) rec = _parse_atom_entry(data) return rec.content def _proper_namespace(self, owner=None, app=None, sharing=None): """Produce a namespace sans wildcards for use in entity requests. This method tries to fill in the fields of the namespace which are `None` or wildcard (`'-'`) from the entity's namespace. If that fails, it uses the service's namespace. :param owner: :param app: :param sharing: :return: """ if owner is None and app is None and sharing is None: # No namespace provided if self._state is not None and 'access' in self._state: return (self._state.access.owner, self._state.access.app, self._state.access.sharing) else: return (self.service.namespace['owner'], self.service.namespace['app'], self.service.namespace['sharing']) else: return (owner,app,sharing) def delete(self): owner, app, sharing = self._proper_namespace() return self.service.delete(self.path, owner=owner, app=app, sharing=sharing) def get(self, path_segment="", owner=None, app=None, sharing=None, **query): owner, app, sharing = self._proper_namespace(owner, app, sharing) return super(Entity, self).get(path_segment, owner=owner, app=app, sharing=sharing, **query) def post(self, path_segment="", owner=None, app=None, sharing=None, **query): owner, app, sharing = self._proper_namespace(owner, app, sharing) return super(Entity, self).post(path_segment, owner=owner, app=app, sharing=sharing, **query) def refresh(self, state=None): """Refreshes the state of this entity. If *state* is provided, load it as the new state for this entity. Otherwise, make a roundtrip to the server (by calling the :meth:`read` method of ``self``) to fetch an updated state, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. :param state: Entity-specific arguments (optional). :type state: ``dict`` :raises EntityDeletedException: Raised if the entity no longer exists on the server. **Example**:: import splunklib.client as client s = client.connect(...) search = s.apps['search'] search.refresh() """ if state is not None: self._state = state else: self._state = self.read(self.get()) return self @property def access(self): """Returns the access metadata for this entity. :return: A :class:`splunklib.data.Record` object with three keys: ``owner``, ``app``, and ``sharing``. """ return self.state.access @property def content(self): """Returns the contents of the entity. :return: A ``dict`` containing values. """ return self.state.content def disable(self): """Disables the entity at this endpoint.""" self.post("disable") if self.service.restart_required: self.service.restart(120) return self def enable(self): """Enables the entity at this endpoint.""" self.post("enable") return self @property def fields(self): """Returns the content metadata for this entity. :return: A :class:`splunklib.data.Record` object with three keys: ``required``, ``optional``, and ``wildcard``. """ return self.state.fields @property def links(self): """Returns a dictionary of related resources. :return: A ``dict`` with keys and corresponding URLs. """ return self.state.links @property def name(self): """Returns the entity name. :return: The entity name. :rtype: ``string`` """ return self.state.title def read(self, response): """ Reads the current state of the entity from the server. """ results = self._load_state(response) # In lower layers of the SDK, we end up trying to URL encode # text to be dispatched via HTTP. However, these links are already # URL encoded when they arrive, and we need to mark them as such. unquoted_links = dict([(k, UrlEncoded(v, skip_encode=True)) for k,v in results['links'].iteritems()]) results['links'] = unquoted_links return results def reload(self): """Reloads the entity.""" self.post("_reload") return self @property def state(self): """Returns the entity's state record. :return: A ``dict`` containing fields and metadata for the entity. """ if self._state is None: self.refresh() return self._state def update(self, **kwargs): """Updates the server with any changes you've made to the current entity along with any additional arguments you specify. **Note**: You cannot update the ``name`` field of an entity. Many of the fields in the REST API are not valid Python identifiers, which means you cannot pass them as keyword arguments. That is, Python will fail to parse the following:: # This fails x.update(check-new=False, email.to='boris@utopia.net') However, you can always explicitly use a dictionary to pass such keys:: # This works x.update(**{'check-new': False, 'email.to': 'boris@utopia.net'}) :param kwargs: Additional entity-specific arguments (optional). :type kwargs: ``dict`` :return: The entity this method is called on. :rtype: class:`Entity` """ # The peculiarity in question: the REST API creates a new # Entity if we pass name in the dictionary, instead of the # expected behavior of updating this Entity. Therefore we # check for 'name' in kwargs and throw an error if it is # there. if 'name' in kwargs: raise IllegalOperationException('Cannot update the name of an Entity via the REST API.') self.post(**kwargs) return self class ReadOnlyCollection(Endpoint): """This class represents a read-only collection of entities in the Splunk instance. """ def __init__(self, service, path, item=Entity): Endpoint.__init__(self, service, path) self.item = item # Item accessor self.null_count = -1 def __contains__(self, name): """Is there at least one entry called *name* in this collection? Makes a single roundtrip to the server, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. """ try: self[name] return True except KeyError: return False except AmbiguousReferenceException: return True def __getitem__(self, key): """Fetch an item named *key* from this collection. A name is not a unique identifier in a collection. The unique identifier is a name plus a namespace. For example, there can be a saved search named ``'mysearch'`` with sharing ``'app'`` in application ``'search'``, and another with sharing ``'user'`` with owner ``'boris'`` and application ``'search'``. If the ``Collection`` is attached to a ``Service`` that has ``'-'`` (wildcard) as user and app in its namespace, then both of these may be visible under the same name. Where there is no conflict, ``__getitem__`` will fetch the entity given just the name. If there is a conflict and you pass just a name, it will raise a ``ValueError``. In that case, add the namespace as a second argument. This function makes a single roundtrip to the server, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. :param key: The name to fetch, or a tuple (name, namespace). :return: An :class:`Entity` object. :raises KeyError: Raised if *key* does not exist. :raises ValueError: Raised if no namespace is specified and *key* does not refer to a unique name. *Example*:: s = client.connect(...) saved_searches = s.saved_searches x1 = saved_searches.create( 'mysearch', 'search * | head 1', owner='admin', app='search', sharing='app') x2 = saved_searches.create( 'mysearch', 'search * | head 1', owner='admin', app='search', sharing='user') # Raises ValueError: saved_searches['mysearch'] # Fetches x1 saved_searches[ 'mysearch', client.namespace(sharing='app', app='search')] # Fetches x2 saved_searches[ 'mysearch', client.namespace(sharing='user', owner='boris', app='search')] """ try: if isinstance(key, tuple) and len(key) == 2: # x[a,b] is translated to x.__getitem__( (a,b) ), so we # have to extract values out. key, ns = key response = self.get(key, owner=ns.owner, app=ns.app) else: response = self.get(key) entries = self._load_list(response) if len(entries) > 1: raise AmbiguousReferenceException("Found multiple entities named '%s'; please specify a namespace." % key) elif len(entries) == 0: raise KeyError(key) else: return entries[0] except HTTPError as he: if he.status == 404: # No entity matching key and namespace. raise KeyError(key) else: raise def __iter__(self, **kwargs): """Iterate over the entities in the collection. :param kwargs: Additional arguments. :type kwargs: ``dict`` :rtype: iterator over entities. Implemented to give Collection a listish interface. This function always makes a roundtrip to the server, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. **Example**:: import splunklib.client as client c = client.connect(...) saved_searches = c.saved_searches for entity in saved_searches: print "Saved search named %s" % entity.name """ for item in self.iter(**kwargs): yield item def __len__(self): """Enable ``len(...)`` for ``Collection`` objects. Implemented for consistency with a listish interface. No further failure modes beyond those possible for any method on an Endpoint. This function always makes a round trip to the server, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. **Example**:: import splunklib.client as client c = client.connect(...) saved_searches = c.saved_searches n = len(saved_searches) """ return len(self.list()) def _entity_path(self, state): """Calculate the path to an entity to be returned. *state* should be the dictionary returned by :func:`_parse_atom_entry`. :func:`_entity_path` extracts the link to this entity from *state*, and strips all the namespace prefixes from it to leave only the relative path of the entity itself, sans namespace. :rtype: ``string`` :return: an absolute path """ # This has been factored out so that it can be easily # overloaded by Configurations, which has to switch its # entities' endpoints from its own properties/ to configs/. raw_path = urllib.unquote(state.links.alternate) if 'servicesNS/' in raw_path: return _trailing(raw_path, 'servicesNS/', '/', '/') elif 'services/' in raw_path: return _trailing(raw_path, 'services/') else: return raw_path def _load_list(self, response): """Converts *response* to a list of entities. *response* is assumed to be a :class:`Record` containing an HTTP response, of the form:: {'status': 200, 'headers': [('content-length', '232642'), ('expires', 'Fri, 30 Oct 1998 00:00:00 GMT'), ('server', 'Splunkd'), ('connection', 'close'), ('cache-control', 'no-store, max-age=0, must-revalidate, no-cache'), ('date', 'Tue, 29 May 2012 15:27:08 GMT'), ('content-type', 'text/xml; charset=utf-8')], 'reason': 'OK', 'body': ...a stream implementing .read()...} The ``'body'`` key refers to a stream containing an Atom feed, that is, an XML document with a toplevel element ````, and within that element one or more ```` elements. """ # Some subclasses of Collection have to override this because # splunkd returns something that doesn't match # . entries = _load_atom_entries(response) if entries is None: return [] entities = [] for entry in entries: state = _parse_atom_entry(entry) entity = self.item( self.service, self._entity_path(state), state=state) entities.append(entity) return entities def itemmeta(self): """Returns metadata for members of the collection. Makes a single roundtrip to the server, plus two more at most if the ``autologin`` field of :func:`connect` is set to ``True``. :return: A :class:`splunklib.data.Record` object containing the metadata. **Example**:: import splunklib.client as client import pprint s = client.connect(...) pprint.pprint(s.apps.itemmeta()) {'access': {'app': 'search', 'can_change_perms': '1', 'can_list': '1', 'can_share_app': '1', 'can_share_global': '1', 'can_share_user': '1', 'can_write': '1', 'modifiable': '1', 'owner': 'admin', 'perms': {'read': ['*'], 'write': ['admin']}, 'removable': '0', 'sharing': 'user'}, 'fields': {'optional': ['author', 'configured', 'description', 'label', 'manageable', 'template', 'visible'], 'required': ['name'], 'wildcard': []}} """ response = self.get("_new") content = _load_atom(response, MATCH_ENTRY_CONTENT) return _parse_atom_metadata(content) def iter(self, offset=0, count=None, pagesize=None, **kwargs): """Iterates over the collection. This method is equivalent to the :meth:`list` method, but it returns an iterator and can load a certain number of entities at a time from the server. :param offset: The index of the first entity to return (optional). :type offset: ``integer`` :param count: The maximum number of entities to return (optional). :type count: ``integer`` :param pagesize: The number of entities to load (optional). :type pagesize: ``integer`` :param kwargs: Additional arguments (optional): - "search" (``string``): The search query to filter responses. - "sort_dir" (``string``): The direction to sort returned items: "asc" or "desc". - "sort_key" (``string``): The field to use for sorting (optional). - "sort_mode" (``string``): The collating sequence for sorting returned items: "auto", "alpha", "alpha_case", or "num". :type kwargs: ``dict`` **Example**:: import splunklib.client as client s = client.connect(...) for saved_search in s.saved_searches.iter(pagesize=10): # Loads 10 saved searches at a time from the # server. ... """ assert pagesize is None or pagesize > 0 if count is None: count = self.null_count fetched = 0 while count == self.null_count or fetched < count: response = self.get(count=pagesize or count, offset=offset, **kwargs) items = self._load_list(response) N = len(items) fetched += N for item in items: yield item if pagesize is None or N < pagesize: break offset += N logging.debug("pagesize=%d, fetched=%d, offset=%d, N=%d, kwargs=%s", pagesize, fetched, offset, N, kwargs) # kwargs: count, offset, search, sort_dir, sort_key, sort_mode def list(self, count=None, **kwargs): """Retrieves a list of entities in this collection. The entire collection is loaded at once and is returned as a list. This function makes a single roundtrip to the server, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. There is no caching--every call makes at least one round trip. :param count: The maximum number of entities to return (optional). :type count: ``integer`` :param kwargs: Additional arguments (optional): - "offset" (``integer``): The offset of the first item to return. - "search" (``string``): The search query to filter responses. - "sort_dir" (``string``): The direction to sort returned items: "asc" or "desc". - "sort_key" (``string``): The field to use for sorting (optional). - "sort_mode" (``string``): The collating sequence for sorting returned items: "auto", "alpha", "alpha_case", or "num". :type kwargs: ``dict`` :return: A ``list`` of entities. """ # response = self.get(count=count, **kwargs) # return self._load_list(response) return list(self.iter(count=count, **kwargs)) class Collection(ReadOnlyCollection): """A collection of entities. Splunk provides a number of different collections of distinct entity types: applications, saved searches, fired alerts, and a number of others. Each particular type is available separately from the Splunk instance, and the entities of that type are returned in a :class:`Collection`. The interface for :class:`Collection` does not quite match either ``list`` or ``dict`` in Python, because there are enough semantic mismatches with either to make its behavior surprising. A unique element in a :class:`Collection` is defined by a string giving its name plus namespace (although the namespace is optional if the name is unique). **Example**:: import splunklib.client as client service = client.connect(...) mycollection = service.saved_searches mysearch = mycollection['my_search', client.namespace(owner='boris', app='natasha', sharing='user')] # Or if there is only one search visible named 'my_search' mysearch = mycollection['my_search'] Similarly, ``name`` in ``mycollection`` works as you might expect (though you cannot currently pass a namespace to the ``in`` operator), as does ``len(mycollection)``. However, as an aggregate, :class:`Collection` behaves more like a list. If you iterate over a :class:`Collection`, you get an iterator over the entities, not the names and namespaces. **Example**:: for entity in mycollection: assert isinstance(entity, client.Entity) Use the :meth:`create` and :meth:`delete` methods to create and delete entities in this collection. To view the access control list and other metadata of the collection, use the :meth:`ReadOnlyCollection.itemmeta` method. :class:`Collection` does no caching. Each call makes at least one round trip to the server to fetch data. """ def create(self, name, **params): """Creates a new entity in this collection. This function makes either one or two roundtrips to the server, depending on the type of entities in this collection, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :param name: The name of the entity to create. :type name: ``string`` :param namespace: A namespace, as created by the :func:`splunklib.binding.namespace` function (optional). You can also set ``owner``, ``app``, and ``sharing`` in ``params``. :type namespace: A :class:`splunklib.data.Record` object with keys ``owner``, ``app``, and ``sharing``. :param params: Additional entity-specific arguments (optional). :type params: ``dict`` :return: The new entity. :rtype: A subclass of :class:`Entity`, chosen by :meth:`Collection.self.item`. **Example**:: import splunklib.client as client s = client.connect(...) applications = s.apps new_app = applications.create("my_fake_app") """ if not isinstance(name, basestring): raise InvalidNameException("%s is not a valid name for an entity." % name) if 'namespace' in params: namespace = params.pop('namespace') params['owner'] = namespace.owner params['app'] = namespace.app params['sharing'] = namespace.sharing response = self.post(name=name, **params) atom = _load_atom(response, XNAME_ENTRY) if atom is None: # This endpoint doesn't return the content of the new # item. We have to go fetch it ourselves. return self[name] else: entry = atom.entry state = _parse_atom_entry(entry) entity = self.item( self.service, self._entity_path(state), state=state) return entity def delete(self, name, **params): """Deletes a specified entity from the collection. :param name: The name of the entity to delete. :type name: ``string`` :return: The collection. :rtype: ``self`` This method is implemented for consistency with the REST API's DELETE method. If there is no *name* entity on the server, a ``KeyError`` is thrown. This function always makes a roundtrip to the server. **Example**:: import splunklib.client as client c = client.connect(...) saved_searches = c.saved_searches saved_searches.create('my_saved_search', 'search * | head 1') assert 'my_saved_search' in saved_searches saved_searches.delete('my_saved_search') assert 'my_saved_search' not in saved_searches """ if 'namespace' in params: namespace = params.pop('namespace') params['owner'] = namespace.owner params['app'] = namespace.app params['sharing'] = namespace.sharing try: self.service.delete(_path(self.path, name), **params) except HTTPError as he: # An HTTPError with status code 404 means that the entity # has already been deleted, and we reraise it as a # KeyError. if he.status == 404: raise KeyError("No such entity %s" % name) else: raise return self class ConfigurationFile(Collection): """This class contains all of the stanzas from one configuration file. """ # __init__'s arguments must match those of an Entity, not a # Collection, since it is being created as the elements of a # Configurations, which is a Collection subclass. def __init__(self, service, path, **kwargs): Collection.__init__(self, service, path, item=Stanza) self.name = kwargs['state']['title'] class Configurations(Collection): """This class provides access to the configuration files from this Splunk instance. Retrieve this collection using :meth:`Service.confs`. Splunk's configuration is divided into files, and each file into stanzas. This collection is unusual in that the values in it are themselves collections of :class:`ConfigurationFile` objects. """ def __init__(self, service): Collection.__init__(self, service, PATH_PROPERTIES, item=ConfigurationFile) if self.service.namespace.owner == '-' or self.service.namespace.app == '-': raise ValueError("Configurations cannot have wildcards in namespace.") def __getitem__(self, key): # The superclass implementation is designed for collections that contain # entities. This collection (Configurations) contains collections # (ConfigurationFile). # # The configurations endpoint returns multiple entities when we ask for a single file. # This screws up the default implementation of __getitem__ from Collection, which thinks # that multiple entities means a name collision, so we have to override it here. try: response = self.get(key) return ConfigurationFile(self.service, PATH_CONF % key, state={'title': key}) except HTTPError as he: if he.status == 404: # No entity matching key raise KeyError(key) else: raise def __contains__(self, key): # configs/conf-{name} never returns a 404. We have to post to properties/{name} # in order to find out if a configuration exists. try: response = self.get(key) return True except HTTPError as he: if he.status == 404: # No entity matching key return False else: raise def create(self, name): """ Creates a configuration file named *name*. If there is already a configuration file with that name, the existing file is returned. :param name: The name of the configuration file. :type name: ``string`` :return: The :class:`ConfigurationFile` object. """ # This has to be overridden to handle the plumbing of creating # a ConfigurationFile (which is a Collection) instead of some # Entity. if not isinstance(name, basestring): raise ValueError("Invalid name: %s" % repr(name)) response = self.post(__conf=name) if response.status == 303: return self[name] elif response.status == 201: return ConfigurationFile(self.service, PATH_CONF % name, item=Stanza, state={'title': name}) else: raise ValueError("Unexpected status code %s returned from creating a stanza" % response.status) def delete(self, key): """Raises `IllegalOperationException`.""" raise IllegalOperationException("Cannot delete configuration files from the REST API.") def _entity_path(self, state): # Overridden to make all the ConfigurationFile objects # returned refer to the configs/ path instead of the # properties/ path used by Configrations. return PATH_CONF % state['title'] class Stanza(Entity): """This class contains a single configuration stanza.""" def submit(self, stanza): """Adds keys to the current configuration stanza as a dictionary of key-value pairs. :param stanza: A dictionary of key-value pairs for the stanza. :type stanza: ``dict`` :return: The :class:`Stanza` object. """ body = _encode(**stanza) self.service.post(self.path, body=body) return self def __len__(self): # The stanza endpoint returns all the keys at the same level in the XML as the eai information # and 'disabled', so to get an accurate length, we have to filter those out and have just # the stanza keys. return len([x for x in self._state.content.keys() if not x.startswith('eai') and x != 'disabled']) class AlertGroup(Entity): """This class represents a group of fired alerts for a saved search. Access it using the :meth:`alerts` property.""" def __init__(self, service, path, **kwargs): Entity.__init__(self, service, path, **kwargs) def __len__(self): return self.count @property def alerts(self): """Returns a collection of triggered alerts. :return: A :class:`Collection` of triggered alerts. """ return Collection(self.service, self.path) @property def count(self): """Returns the count of triggered alerts. :return: The triggered alert count. :rtype: ``integer`` """ return int(self.content.get('triggered_alert_count', 0)) class Indexes(Collection): """This class contains the collection of indexes in this Splunk instance. Retrieve this collection using :meth:`Service.indexes`. """ def get_default(self): """ Returns the name of the default index. :return: The name of the default index. """ index = self['_audit'] return index['defaultDatabase'] def delete(self, name): """ Deletes a given index. **Note**: This method is only supported in Splunk 5.0 and later. :param name: The name of the index to delete. :type name: ``string`` """ if self.service.splunk_version >= (5,): Collection.delete(self, name) else: raise IllegalOperationException("Deleting indexes via the REST API is " "not supported before Splunk version 5.") class Index(Entity): """This class represents an index and provides different operations, such as cleaning the index, writing to the index, and so forth.""" def __init__(self, service, path, **kwargs): Entity.__init__(self, service, path, **kwargs) def attach(self, host=None, source=None, sourcetype=None): """Opens a stream (a writable socket) for writing events to the index. :param host: The host value for events written to the stream. :type host: ``string`` :param source: The source value for events written to the stream. :type source: ``string`` :param sourcetype: The sourcetype value for events written to the stream. :type sourcetype: ``string`` :return: A writable socket. """ args = { 'index': self.name } if host is not None: args['host'] = host if source is not None: args['source'] = source if sourcetype is not None: args['sourcetype'] = sourcetype path = UrlEncoded(PATH_RECEIVERS_STREAM + "?" + urllib.urlencode(args), skip_encode=True) # Since we need to stream to the index connection, we have to keep # the connection open and use the Splunk extension headers to note # the input mode sock = self.service.connect() headers = ["POST %s HTTP/1.1\r\n" % self.service._abspath(path), "Host: %s:%s\r\n" % (self.service.host, int(self.service.port)), "Accept-Encoding: identity\r\n", "Authorization: %s\r\n" % self.service.token, "X-Splunk-Input-Mode: Streaming\r\n", "\r\n"] for h in headers: sock.write(h) return sock @contextlib.contextmanager def attached_socket(self, *args, **kwargs): """Opens a raw socket in a ``with`` block to write data to Splunk. The arguments are identical to those for :meth:`attach`. The socket is automatically closed at the end of the ``with`` block, even if an exception is raised in the block. :param host: The host value for events written to the stream. :type host: ``string`` :param source: The source value for events written to the stream. :type source: ``string`` :param sourcetype: The sourcetype value for events written to the stream. :type sourcetype: ``string`` :returns: Nothing. **Example**:: import splunklib.client as client s = client.connect(...) index = s.indexes['some_index'] with index.attached_socket(sourcetype='test') as sock: sock.send('Test event\\r\\n') """ try: sock = self.attach(*args, **kwargs) yield sock finally: sock.shutdown(socket.SHUT_RDWR) sock.close() def clean(self, timeout=60): """Deletes the contents of the index. This method blocks until the index is empty, because it needs to restore values at the end of the operation. :param timeout: The time-out period for the operation, in seconds (the default is 60). :type timeout: ``integer`` :return: The :class:`Index`. """ self.refresh() tds = self['maxTotalDataSizeMB'] ftp = self['frozenTimePeriodInSecs'] was_disabled_initially = self.disabled try: if (not was_disabled_initially and \ self.service.splunk_version < (5,)): # Need to disable the index first on Splunk 4.x, # but it doesn't work to disable it on 5.0. self.disable() self.update(maxTotalDataSizeMB=1, frozenTimePeriodInSecs=1) self.roll_hot_buckets() # Wait until event count goes to 0. start = datetime.now() diff = timedelta(seconds=timeout) while self.content.totalEventCount != '0' and datetime.now() < start+diff: sleep(1) self.refresh() if self.content.totalEventCount != '0': raise OperationError, "Cleaning index %s took longer than %s seconds; timing out." %\ (self.name, timeout) finally: # Restore original values self.update(maxTotalDataSizeMB=tds, frozenTimePeriodInSecs=ftp) if (not was_disabled_initially and \ self.service.splunk_version < (5,)): # Re-enable the index if it was originally enabled and we messed with it. self.enable() return self def roll_hot_buckets(self): """Performs rolling hot buckets for this index. :return: The :class:`Index`. """ self.post("roll-hot-buckets") return self def submit(self, event, host=None, source=None, sourcetype=None): """Submits a single event to the index using ``HTTP POST``. :param event: The event to submit. :type event: ``string`` :param `host`: The host value of the event. :type host: ``string`` :param `source`: The source value of the event. :type source: ``string`` :param `sourcetype`: The sourcetype value of the event. :type sourcetype: ``string`` :return: The :class:`Index`. """ args = { 'index': self.name } if host is not None: args['host'] = host if source is not None: args['source'] = source if sourcetype is not None: args['sourcetype'] = sourcetype # The reason we use service.request directly rather than POST # is that we are not sending a POST request encoded using # x-www-form-urlencoded (as we do not have a key=value body), # because we aren't really sending a "form". self.service.post(PATH_RECEIVERS_SIMPLE, body=event, **args) return self # kwargs: host, host_regex, host_segment, rename-source, sourcetype def upload(self, filename, **kwargs): """Uploads a file for immediate indexing. **Note**: The file must be locally accessible from the server. :param filename: The name of the file to upload. The file can be a plain, compressed, or archived file. :type filename: ``string`` :param kwargs: Additional arguments (optional). For more about the available parameters, see `Index parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The :class:`Index`. """ kwargs['index'] = self.name path = 'data/inputs/oneshot' self.service.post(path, name=filename, **kwargs) return self class Input(Entity): """This class represents a Splunk input. This class is the base for all typed input classes and is also used when the client does not recognize an input kind. """ def __init__(self, service, path, kind=None, **kwargs): # kind can be omitted (in which case it is inferred from the path) # Otherwise, valid values are the paths from data/inputs ("udp", # "monitor", "tcp/raw"), or two special cases: "tcp" (which is "tcp/raw") # and "splunktcp" (which is "tcp/cooked"). Entity.__init__(self, service, path, **kwargs) if kind is None: path_segments = path.split('/') i = path_segments.index('inputs') + 1 if path_segments[i] == 'tcp': self.kind = path_segments[i] + '/' + path_segments[i+1] else: self.kind = path_segments[i] else: self.kind = kind # Handle old input kind names. if self.kind == 'tcp': self.kind = 'tcp/raw' if self.kind == 'splunktcp': self.kind = 'tcp/cooked' def update(self, **kwargs): """Updates the server with any changes you've made to the current input along with any additional arguments you specify. :param kwargs: Additional arguments (optional). For more about the available parameters, see `Input parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The input this method was called on. :rtype: class:`Input` """ # UDP and TCP inputs require special handling due to their restrictToHost # field. For all other inputs kinds, we can dispatch to the superclass method. if self.kind not in ['tcp', 'splunktcp', 'tcp/raw', 'tcp/cooked', 'udp']: return super(Input, self).update(**kwargs) else: # The behavior of restrictToHost is inconsistent across input kinds and versions of Splunk. # In Splunk 4.x, the name of the entity is only the port, independent of the value of # restrictToHost. In Splunk 5.0 this changed so the name will be of the form :. # In 5.0 and 5.0.1, if you don't supply the restrictToHost value on every update, it will # remove the host restriction from the input. As of 5.0.2 you simply can't change restrictToHost # on an existing input. # The logic to handle all these cases: # - Throw an exception if the user tries to set restrictToHost on an existing input # for *any* version of Splunk. # - Set the existing restrictToHost value on the update args internally so we don't # cause it to change in Splunk 5.0 and 5.0.1. to_update = kwargs.copy() if 'restrictToHost' in kwargs: raise IllegalOperationException("Cannot set restrictToHost on an existing input with the SDK.") elif 'restrictToHost' in self._state.content: to_update['restrictToHost'] = self._state.content['restrictToHost'] # Do the actual update operation. return super(Input, self).update(**to_update) # Inputs is a "kinded" collection, which is a heterogenous collection where # each item is tagged with a kind, that provides a single merged view of all # input kinds. class Inputs(Collection): """This class represents a collection of inputs. The collection is heterogeneous and each member of the collection contains a *kind* property that indicates the specific type of input. Retrieve this collection using :meth:`Service.inputs`.""" def __init__(self, service, kindmap=None): Collection.__init__(self, service, PATH_INPUTS, item=Input) def __getitem__(self, key): if isinstance(key, tuple) and len(key) == 2: # Fetch a single kind key, kind = key try: response = self.get(self.kindpath(kind) + "/" + key) entries = self._load_list(response) if len(entries) > 1: raise AmbiguousReferenceException("Found multiple inputs of kind %s named %s." % (kind, key)) elif len(entries) == 0: raise KeyError((kind,key)) else: return entries[0] except HTTPError as he: if he.status == 404: # No entity matching kind and key raise KeyError((kind,key)) else: raise else: # Iterate over all the kinds looking for matches. kind = None candidate = None for kind in self.kinds: try: response = self.get(kind + "/" + key) entries = self._load_list(response) if len(entries) > 1: raise AmbiguousReferenceException("Found multiple inputs of kind %s named %s." % (kind, key)) elif len(entries) == 0: pass else: if candidate is not None: # Already found at least one candidate raise AmbiguousReferenceException("Found multiple inputs named %s, please specify a kind" % key) candidate = entries[0] except HTTPError as he: if he.status == 404: pass # Just carry on to the next kind. else: raise if candidate is None: raise KeyError(key) # Never found a match else: return candidate def __contains__(self, key): if isinstance(key, tuple) and len(key) == 2: # If we specify a kind, this will shortcut properly try: self.__getitem__(key) return True except KeyError: return False else: # Without a kind, we want to minimize the number of round trips to the server, so we # reimplement some of the behavior of __getitem__ in order to be able to stop searching # on the first hit. for kind in self.kinds: try: response = self.get(self.kindpath(kind) + "/" + key) entries = self._load_list(response) if len(entries) > 0: return True else: pass except HTTPError as he: if he.status == 404: pass # Just carry on to the next kind. else: raise return False def create(self, name, kind, **kwargs): """Creates an input of a specific kind in this collection, with any arguments you specify. :param `name`: The input name. :type name: ``string`` :param `kind`: The kind of input: - "ad": Active Directory - "monitor": Files and directories - "registry": Windows Registry - "script": Scripts - "splunktcp": TCP, processed - "tcp": TCP, unprocessed - "udp": UDP - "win-event-log-collections": Windows event log - "win-perfmon": Performance monitoring - "win-wmi-collections": WMI :type kind: ``string`` :param `kwargs`: Additional arguments (optional). For more about the available parameters, see `Input parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The new :class:`Input`. """ kindpath = self.kindpath(kind) self.post(kindpath, name=name, **kwargs) # If we created an input with restrictToHost set, then # its path will be :, not just , # and we have to adjust accordingly. path = _path( self.path + kindpath, '%s:%s' % (kwargs['restrictToHost'], name) \ if kwargs.has_key('restrictToHost') else name ) return Input(self.service, path, kind) def delete(self, name, kind=None): """Removes an input from the collection. :param `kind`: The kind of input: - "ad": Active Directory - "monitor": Files and directories - "registry": Windows Registry - "script": Scripts - "splunktcp": TCP, processed - "tcp": TCP, unprocessed - "udp": UDP - "win-event-log-collections": Windows event log - "win-perfmon": Performance monitoring - "win-wmi-collections": WMI :type kind: ``string`` :param name: The name of the input to remove. :type name: ``string`` :return: The :class:`Inputs` collection. """ if kind is None: self.service.delete(self[name].path) else: self.service.delete(self[name, kind].path) return self def itemmeta(self, kind): """Returns metadata for the members of a given kind. :param `kind`: The kind of input: - "ad": Active Directory - "monitor": Files and directories - "registry": Windows Registry - "script": Scripts - "splunktcp": TCP, processed - "tcp": TCP, unprocessed - "udp": UDP - "win-event-log-collections": Windows event log - "win-perfmon": Performance monitoring - "win-wmi-collections": WMI :type kind: ``string`` :return: The metadata. :rtype: class:``splunklib.data.Record`` """ response = self.get("%s/_new" % self._kindmap[kind]) content = _load_atom(response, MATCH_ENTRY_CONTENT) return _parse_atom_metadata(content) def _get_kind_list(self, subpath=None): if subpath is None: subpath = [] kinds = [] response = self.get('/'.join(subpath)) content = _load_atom_entries(response) for entry in content: this_subpath = subpath + [entry.title] # The "all" endpoint doesn't work yet. # The "tcp/ssl" endpoint is not a real input collection. if entry.title == 'all' or this_subpath == ['tcp','ssl']: continue elif 'create' in [x.rel for x in entry.link]: path = '/'.join(subpath + [entry.title]) kinds.append(path) else: subkinds = self._get_kind_list(subpath + [entry.title]) kinds.extend(subkinds) return kinds @property def kinds(self): """Returns the input kinds on this Splunk instance. :return: The list of input kinds. :rtype: ``list`` """ return self._get_kind_list() def kindpath(self, kind): """Returns a path to the resources for a given input kind. :param `kind`: The kind of input: - "ad": Active Directory - "monitor": Files and directories - "registry": Windows Registry - "script": Scripts - "splunktcp": TCP, processed - "tcp": TCP, unprocessed - "udp": UDP - "win-event-log-collections": Windows event log - "win-perfmon": Performance monitoring - "win-wmi-collections": WMI :type kind: ``string`` :return: The relative endpoint path. :rtype: ``string`` """ if kind in self.kinds: return kind # Special cases elif kind == 'tcp': return 'tcp/raw' elif kind == 'splunktcp': return 'tcp/cooked' else: raise ValueError("No such kind on server: %s" % kind) def list(self, *kinds, **kwargs): """Returns a list of inputs that are in the :class:`Inputs` collection. You can also filter by one or more input kinds. This function iterates over all possible inputs, regardless of any arguments you specify. Because the :class:`Inputs` collection is the union of all the inputs of each kind, this method implements parameters such as "count", "search", and so on at the Python level once all the data has been fetched. The exception is when you specify a single input kind, and then this method makes a single request with the usual semantics for parameters. :param kinds: The input kinds to return (optional). - "ad": Active Directory - "monitor": Files and directories - "registry": Windows Registry - "script": Scripts - "splunktcp": TCP, processed - "tcp": TCP, unprocessed - "udp": UDP - "win-event-log-collections": Windows event log - "win-perfmon": Performance monitoring - "win-wmi-collections": WMI :type kinds: ``string`` :param kwargs: Additional arguments (optional): - "count" (``integer``): The maximum number of items to return. - "offset" (``integer``): The offset of the first item to return. - "search" (``string``): The search query to filter responses. - "sort_dir" (``string``): The direction to sort returned items: "asc" or "desc". - "sort_key" (``string``): The field to use for sorting (optional). - "sort_mode" (``string``): The collating sequence for sorting returned items: "auto", "alpha", "alpha_case", or "num". :type kwargs: ``dict`` :return: A list of input kinds. :rtype: ``list`` """ if len(kinds) == 0: kinds = self.kinds if len(kinds) == 1: kind = kinds[0] logging.debug("Inputs.list taking short circuit branch for single kind.") path = self.kindpath(kind) logging.debug("Path for inputs: %s", path) try: response = self.get(path, **kwargs) except HTTPError, he: if he.status == 404: # No inputs of this kind return [] entities = [] entries = _load_atom_entries(response) if entries is None: return [] # No inputs in a collection comes back with no feed or entry in the XML for entry in entries: state = _parse_atom_entry(entry) # Unquote the URL, since all URL encoded in the SDK # should be of type UrlEncoded, and all str should not # be URL encoded. path = urllib.unquote(state.links.alternate) entity = Input(self.service, path, kind, state=state) entities.append(entity) return entities search = kwargs.get('search', '*') entities = [] for kind in kinds: response = None try: response = self.get(self.kindpath(kind), search=search) except HTTPError as e: if e.status == 404: continue # No inputs of this kind else: raise entries = _load_atom_entries(response) if entries is None: continue # No inputs to process for entry in entries: state = _parse_atom_entry(entry) # Unquote the URL, since all URL encoded in the SDK # should be of type UrlEncoded, and all str should not # be URL encoded. path = urllib.unquote(state.links.alternate) entity = Input(self.service, path, kind, state=state) entities.append(entity) if 'offset' in kwargs: entities = entities[kwargs['offset']:] if 'count' in kwargs: entities = entities[:kwargs['count']] if kwargs.get('sort_mode', None) == 'alpha': sort_field = kwargs.get('sort_field', 'name') if sort_field == 'name': f = lambda x: x.name.lower() else: f = lambda x: x[sort_field].lower() entities = sorted(entities, key=f) if kwargs.get('sort_mode', None) == 'alpha_case': sort_field = kwargs.get('sort_field', 'name') if sort_field == 'name': f = lambda x: x.name else: f = lambda x: x[sort_field] entities = sorted(entities, key=f) if kwargs.get('sort_dir', 'asc') == 'desc': entities = list(reversed(entities)) return entities def __iter__(self, **kwargs): for item in self.iter(**kwargs): yield item def iter(self, **kwargs): """ Iterates over the collection of inputs. :param kwargs: Additional arguments (optional): - "count" (``integer``): The maximum number of items to return. - "offset" (``integer``): The offset of the first item to return. - "search" (``string``): The search query to filter responses. - "sort_dir" (``string``): The direction to sort returned items: "asc" or "desc". - "sort_key" (``string``): The field to use for sorting (optional). - "sort_mode" (``string``): The collating sequence for sorting returned items: "auto", "alpha", "alpha_case", or "num". :type kwargs: ``dict`` """ for item in self.list(**kwargs): yield item def oneshot(self, path, **kwargs): """ Creates a oneshot data input, which is an upload of a single file for one-time indexing. :param path: The path and filename. :type path: ``string`` :param kwargs: Additional arguments (optional). For more about the available parameters, see `Input parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` """ self.post('oneshot', name=path, **kwargs) class Job(Entity): """This class represents a search job.""" def __init__(self, service, sid, **kwargs): path = PATH_JOBS + sid Entity.__init__(self, service, path, skip_refresh=True, **kwargs) self.sid = sid # The Job entry record is returned at the root of the response def _load_atom_entry(self, response): return _load_atom(response).entry def cancel(self): """Stops the current search and deletes the results cache. :return: The :class:`Job`. """ try: self.post("control", action="cancel") except HTTPError as he: if he.status == 404: # The job has already been cancelled, so # cancelling it twice is a nop. pass else: raise return self def disable_preview(self): """Disables preview for this job. :return: The :class:`Job`. """ self.post("control", action="disablepreview") return self def enable_preview(self): """Enables preview for this job. **Note**: Enabling preview might slow search considerably. :return: The :class:`Job`. """ self.post("control", action="enablepreview") return self def events(self, **kwargs): """Returns a streaming handle to this job's events. :param kwargs: Additional parameters (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/events `_ in the REST API documentation. :type kwargs: ``dict`` :return: The ``InputStream`` IO handle to this job's events. """ kwargs['segmentation'] = kwargs.get('segmentation', 'none') return self.get("events", **kwargs).body def finalize(self): """Stops the job and provides intermediate results for retrieval. :return: The :class:`Job`. """ self.post("control", action="finalize") return self def is_done(self): """Indicates whether this job finished running. :return: ``True`` if the job is done, ``False`` if not. :rtype: ``boolean`` """ if not self.is_ready(): return False done = (self._state.content['isDone'] == '1') return done def is_ready(self): """Indicates whether this job is ready for querying. :return: ``True`` if the job is ready, ``False`` if not. :rtype: ``boolean`` """ response = self.get() if response.status == 204: return False self._state = self.read(response) ready = self._state.content['dispatchState'] not in ['QUEUED', 'PARSING'] return ready @property def name(self): """Returns the name of the search job, which is the search ID (SID). :return: The search ID. :rtype: ``string`` """ return self.sid def pause(self): """Suspends the current search. :return: The :class:`Job`. """ self.post("control", action="pause") return self def results(self, **query_params): """Returns a streaming handle to this job's search results. To get a nice, Pythonic iterator, pass the handle to :class:`splunklib.results.ResultsReader`, as in:: import splunklib.client as client import splunklib.results as results from time import sleep service = client.connect(...) job = service.jobs.create("search * | head 5") while not job.is_done(): sleep(.2) rr = results.ResultsReader(job.results()) for result in rr: if isinstance(result, results.Message): # Diagnostic messages may be returned in the results print '%s: %s' % (result.type, result.message) elif isinstance(result, dict): # Normal events are returned as dicts print result assert rr.is_preview == False Results are not available until the job has finished. If called on an unfinished job, the result is an empty event set. This method makes a single roundtrip to the server, plus at most two additional round trips if the ``autologin`` field of :func:`connect` is set to ``True``. :param query_params: Additional parameters (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/results `_. :type query_params: ``dict`` :return: The ``InputStream`` IO handle to this job's results. """ query_params['segmentation'] = query_params.get('segmentation', 'none') return self.get("results", **query_params).body def preview(self, **query_params): """Returns a streaming handle to this job's preview search results. Unlike :class:`splunklib.results.ResultsReader`, which requires a job to be finished to return any results, the ``preview`` method returns any results that have been generated so far, whether the job is running or not. The returned search results are the raw data from the server. Pass the handle returned to :class:`splunklib.results.ResultsReader` to get a nice, Pythonic iterator over objects, as in:: import splunklib.client as client import splunklib.results as results service = client.connect(...) job = service.jobs.create("search * | head 5") rr = results.ResultsReader(job.preview()) for result in rr: if isinstance(result, results.Message): # Diagnostic messages may be returned in the results print '%s: %s' % (result.type, result.message) elif isinstance(result, dict): # Normal events are returned as dicts print result if rr.is_preview: print "Preview of a running search job." else: print "Job is finished. Results are final." This method makes one roundtrip to the server, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :param query_params: Additional parameters (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/results_preview `_ in the REST API documentation. :type query_params: ``dict`` :return: The ``InputStream`` IO handle to this job's preview results. """ query_params['segmentation'] = query_params.get('segmentation', 'none') return self.get("results_preview", **query_params).body def searchlog(self, **kwargs): """Returns a streaming handle to this job's search log. :param `kwargs`: Additional parameters (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/search.log `_ in the REST API documentation. :type kwargs: ``dict`` :return: The ``InputStream`` IO handle to this job's search log. """ return self.get("search.log", **kwargs).body def set_priority(self, value): """Sets this job's search priority in the range of 0-10. Higher numbers indicate higher priority. Unless splunkd is running as *root*, you can only decrease the priority of a running job. :param `value`: The search priority. :type value: ``integer`` :return: The :class:`Job`. """ self.post('control', action="setpriority", priority=value) return self def summary(self, **kwargs): """Returns a streaming handle to this job's summary. :param `kwargs`: Additional parameters (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/summary `_ in the REST API documentation. :type kwargs: ``dict`` :return: The ``InputStream`` IO handle to this job's summary. """ return self.get("summary", **kwargs).body def timeline(self, **kwargs): """Returns a streaming handle to this job's timeline results. :param `kwargs`: Additional timeline arguments (optional). For a list of valid parameters, see `GET search/jobs/{search_id}/timeline `_ in the REST API documentation. :type kwargs: ``dict`` :return: The ``InputStream`` IO handle to this job's timeline. """ return self.get("timeline", **kwargs).body def touch(self): """Extends the expiration time of the search to the current time (now) plus the time-to-live (ttl) value. :return: The :class:`Job`. """ self.post("control", action="touch") return self def set_ttl(self, value): """Set the job's time-to-live (ttl) value, which is the time before the search job expires and is still available. :param `value`: The ttl value, in seconds. :type value: ``integer`` :return: The :class:`Job`. """ self.post("control", action="setttl", ttl=value) return self def unpause(self): """Resumes the current search, if paused. :return: The :class:`Job`. """ self.post("control", action="unpause") return self class Jobs(Collection): """This class represents a collection of search jobs. Retrieve this collection using :meth:`Service.jobs`.""" def __init__(self, service): Collection.__init__(self, service, PATH_JOBS, item=Job) # The count value to say list all the contents of this # Collection is 0, not -1 as it is on most. self.null_count = 0 def _load_list(self, response): # Overridden because Job takes a sid instead of a path. entries = _load_atom_entries(response) if entries is None: return [] entities = [] for entry in entries: state = _parse_atom_entry(entry) entity = self.item( self.service, entry['content']['sid'], state=state) entities.append(entity) return entities def create(self, query, **kwargs): """ Creates a search using a search query and any additional parameters you provide. :param query: The search query. :type query: ``string`` :param kwargs: Additiona parameters (optional). For a list of available parameters, see `Search job parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The :class:`Job`. """ if kwargs.get("exec_mode", None) == "oneshot": raise TypeError("Cannot specify exec_mode=oneshot; use the oneshot method instead.") response = self.post(search=query, **kwargs) sid = _load_sid(response) return Job(self.service, sid) def export(self, query, **params): """Runs a search and immediately starts streaming preview events. This method returns a streaming handle to this job's events as an XML document from the server. To parse this stream into usable Python objects, pass the handle to :class:`splunklib.results.ResultsReader`:: import splunklib.client as client import splunklib.results as results service = client.connect(...) rr = results.ResultsReader(service.jobs.export("search * | head 5")) for result in rr: if isinstance(result, results.Message): # Diagnostic messages may be returned in the results print '%s: %s' % (result.type, result.message) elif isinstance(result, dict): # Normal events are returned as dicts print result assert rr.is_preview == False Running an export search is more efficient as it streams the results directly to you, rather than having to write them out to disk and make them available later. As soon as results are ready, you will receive them. The ``export`` method makes a single roundtrip to the server (as opposed to two for :meth:`create` followed by :meth:`preview`), plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :raises `ValueError`: Raised for invalid queries. :param query: The search query. :type query: ``string`` :param params: Additional arguments (optional). For a list of valid parameters, see `GET search/jobs/export `_ in the REST API documentation. :type params: ``dict`` :return: The ``InputStream`` IO handle to raw XML returned from the server. """ if "exec_mode" in params: raise TypeError("Cannot specify an exec_mode to export.") params['segmentation'] = params.get('segmentation', 'none') return self.post(path_segment="export", search=query, **params).body def itemmeta(self): """There is no metadata available for class:``Jobs``. Any call to this method raises a class:``NotSupportedError``. :raises: class:``NotSupportedError`` """ raise NotSupportedError() def oneshot(self, query, **params): """Run a oneshot search and returns a streaming handle to the results. The ``InputStream`` object streams XML fragments from the server. To parse this stream into usable Python objects, pass the handle to :class:`splunklib.results.ResultsReader`:: import splunklib.client as client import splunklib.results as results service = client.connect(...) rr = results.ResultsReader(service.jobs.oneshot("search * | head 5")) for result in rr: if isinstance(result, results.Message): # Diagnostic messages may be returned in the results print '%s: %s' % (result.type, result.message) elif isinstance(result, dict): # Normal events are returned as dicts print result assert rr.is_preview == False The ``oneshot`` method makes a single roundtrip to the server (as opposed to two for :meth:`create` followed by :meth:`results`), plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :raises ValueError: Raised for invalid queries. :param query: The search query. :type query: ``string`` :param params: Additional arguments (optional): - "output_mode": Specifies the output format of the results (XML, JSON, or CSV). - "earliest_time": Specifies the earliest time in the time range to search. The time string can be a UTC time (with fractional seconds), a relative time specifier (to now), or a formatted time string. - "latest_time": Specifies the latest time in the time range to search. The time string can be a UTC time (with fractional seconds), a relative time specifier (to now), or a formatted time string. - "rf": Specifies one or more fields to add to the search. :type params: ``dict`` :return: The ``InputStream`` IO handle to raw XML returned from the server. """ if "exec_mode" in params: raise TypeError("Cannot specify an exec_mode to oneshot.") params['segmentation'] = params.get('segmentation', 'none') return self.post(search=query, exec_mode="oneshot", **params).body class Loggers(Collection): """This class represents a collection of service logging categories. Retrieve this collection using :meth:`Service.loggers`.""" def __init__(self, service): Collection.__init__(self, service, PATH_LOGGER) def itemmeta(self): """There is no metadata available for class:``Loggers``. Any call to this method raises a class:``NotSupportedError``. :raises: class:``NotSupportedError`` """ raise NotSupportedError() class Message(Entity): def __init__(self, service, path, **kwargs): Entity.__init__(self, service, path, **kwargs) @property def value(self): """Returns the message value. :return: The :class:`Loggers` collection. """ return self[self.name] class ModularInputKind(Entity): """This class contains the different types of modular inputs. Retrieve this collection using :meth:`Service.modular_input_kinds`. """ def __contains__(self, name): args = self.state.content['endpoints']['args'] if name in args: return True else: return Entity.__contains__(self, name) def __getitem__(self, name): args = self.state.content['endpoint']['args'] if name in args: return args['item'] else: return Entity.__getitem__(self, name) @property def arguments(self): """A dictionary of all the arguments supported by this modular input kind. The keys in the dictionary are the names of the arguments. The values are another dictionary giving the metadata about that argument. The possible keys in that dictionary are ``"title"``, ``"description"``, ``"required_on_create``", ``"required_on_edit"``, ``"data_type"``. Each value is a string. It should be one of ``"true"`` or ``"false"`` for ``"required_on_create"`` and ``"required_on_edit"``, and one of ``"boolean"``, ``"string"``, or ``"number``" for ``"data_type"``. :return: A dictionary describing the arguments this modular input kind takes. :rtype: ``dict`` """ return self.state.content['endpoint']['args'] def update(self, **kwargs): """Raises an error. Modular input kinds are read only.""" raise IllegalOperationException("Modular input kinds cannot be updated via the REST API.") class SavedSearch(Entity): """This class represents a saved search.""" def __init__(self, service, path, **kwargs): Entity.__init__(self, service, path, **kwargs) def acknowledge(self): """Acknowledges the suppression of alerts from this saved search and resumes alerting. :return: The :class:`SavedSearch`. """ self.post("acknowledge") return self @property def alert_count(self): """Returns the number of alerts fired by this saved search. :return: The number of alerts fired by this saved search. :rtype: ``integer`` """ return int(self._state.content.get('triggered_alert_count', 0)) def dispatch(self, **kwargs): """Runs the saved search and returns the resulting search job. :param `kwargs`: Additional dispatch arguments (optional). For details, see the `POST saved/searches/{name}/dispatch `_ endpoint in the REST API documentation. :type kwargs: ``dict`` :return: The :class:`Job`. """ response = self.post("dispatch", **kwargs) sid = _load_sid(response) return Job(self.service, sid) @property def fired_alerts(self): """Returns the collection of fired alerts (a fired alert group) corresponding to this saved search's alerts. :raises IllegalOperationException: Raised when the search is not scheduled. :return: A collection of fired alerts. :rtype: :class:`AlertGroup` """ if self['is_scheduled'] == '0': raise IllegalOperationException('Unscheduled saved searches have no alerts.') c = Collection( self.service, self.service._abspath(PATH_FIRED_ALERTS + self.name, owner=self._state.access.owner, app=self._state.access.app, sharing=self._state.access.sharing), item=AlertGroup) return c def history(self): """Returns a list of search jobs corresponding to this saved search. :return: A list of :class:`Job` objects. """ response = self.get("history") entries = _load_atom_entries(response) if entries is None: return [] jobs = [] for entry in entries: job = Job(self.service, entry.title) jobs.append(job) return jobs def update(self, search=None, **kwargs): """Updates the server with any changes you've made to the current saved search along with any additional arguments you specify. :param `search`: The search query (optional). :type search: ``string`` :param `kwargs`: Additional arguments (optional). For a list of available parameters, see `Saved search parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The :class:`SavedSearch`. """ # Updates to a saved search *require* that the search string be # passed, so we pass the current search string if a value wasn't # provided by the caller. if search is None: search = self.content.search Entity.update(self, search=search, **kwargs) return self def scheduled_times(self, earliest_time='now', latest_time='+1h'): """Returns the times when this search is scheduled to run. By default this method returns the times in the next hour. For different time ranges, set *earliest_time* and *latest_time*. For example, for all times in the last day use "earliest_time=-1d" and "latest_time=now". :param earliest_time: The earliest time. :type earliest_time: ``string`` :param latest_time: The latest time. :type latest_time: ``string`` :return: The list of search times. """ response = self.get("scheduled_times", earliest_time=earliest_time, latest_time=latest_time) data = self._load_atom_entry(response) rec = _parse_atom_entry(data) times = [datetime.fromtimestamp(int(t)) for t in rec.content.scheduled_times] return times def suppress(self, expiration): """Skips any scheduled runs of this search in the next *expiration* number of seconds. :param expiration: The expiration period, in seconds. :type expiration: ``integer`` :return: The :class:`SavedSearch`. """ self.post("suppress", expiration=expiration) return self @property def suppressed(self): """Returns the number of seconds that this search is blocked from running (possibly 0). :return: The number of seconds. :rtype: ``integer`` """ r = self._run_action("suppress") if r.suppressed == "1": return int(r.expiration) else: return 0 def unsuppress(self): """Cancels suppression and makes this search run as scheduled. :return: The :class:`SavedSearch`. """ self.post("suppress", expiration="0") return self class SavedSearches(Collection): """This class represents a collection of saved searches. Retrieve this collection using :meth:`Service.saved_searches`.""" def __init__(self, service): Collection.__init__( self, service, PATH_SAVED_SEARCHES, item=SavedSearch) def create(self, name, search, **kwargs): """ Creates a saved search. :param name: The name for the saved search. :type name: ``string`` :param search: The search query. :type search: ``string`` :param kwargs: Additional arguments (optional). For a list of available parameters, see `Saved search parameters `_ on Splunk Developer Portal. :type kwargs: ``dict`` :return: The :class:`SavedSearches` collection. """ return Collection.create(self, name, search=search, **kwargs) class Settings(Entity): """This class represents configuration settings for a Splunk service. Retrieve this collection using :meth:`Service.settings`.""" def __init__(self, service, **kwargs): Entity.__init__(self, service, "server/settings", **kwargs) # Updates on the settings endpoint are POSTed to server/settings/settings. def update(self, **kwargs): """Updates the settings on the server using the arguments you provide. :param kwargs: Additional arguments. For a list of valid arguments, see `POST server/settings/{name} `_ in the REST API documentation. :type kwargs: ``dict`` :return: The :class:`Settings` collection. """ self.service.post("server/settings/settings", **kwargs) return self class User(Entity): """This class represents a Splunk user. """ @property def role_entities(self): """Returns a list of roles assigned to this user. :return: The list of roles. :rtype: ``list`` """ return [self.service.roles[name] for name in self.content.roles] # Splunk automatically lowercases new user names so we need to match that # behavior here to ensure that the subsequent member lookup works correctly. class Users(Collection): """This class represents the collection of Splunk users for this instance of Splunk. Retrieve this collection using :meth:`Service.users`. """ def __init__(self, service): Collection.__init__(self, service, PATH_USERS, item=User) def __getitem__(self, key): return Collection.__getitem__(self, key.lower()) def __contains__(self, name): return Collection.__contains__(self, name.lower()) def create(self, username, password, roles, **params): """Creates a new user. This function makes two roundtrips to the server, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :param username: The username. :type username: ``string`` :param password: The password. :type password: ``string`` :param roles: A single role or list of roles for the user. :type roles: ``string`` or ``list`` :param params: Additional arguments (optional). For a list of available parameters, see `User authentication parameters `_ on Splunk Developer Portal. :type params: ``dict`` :return: The new user. :rtype: :class:`User` **Example**:: import splunklib.client as client c = client.connect(...) users = c.users boris = users.create("boris", "securepassword", roles="user") hilda = users.create("hilda", "anotherpassword", roles=["user","power"]) """ if not isinstance(username, basestring): raise ValueError("Invalid username: %s" % str(username)) username = username.lower() self.post(name=username, password=password, roles=roles, **params) # splunkd doesn't return the user in the POST response body, # so we have to make a second round trip to fetch it. response = self.get(username) entry = _load_atom(response, XNAME_ENTRY).entry state = _parse_atom_entry(entry) entity = self.item( self.service, urllib.unquote(state.links.alternate), state=state) return entity def delete(self, name): """ Deletes the user and returns the resulting collection of users. :param name: The name of the user to delete. :type name: ``string`` :return: :rtype: :class:`Users` """ return Collection.delete(self, name.lower()) class Role(Entity): """This class represents a user role. """ def grant(self, *capabilities_to_grant): """Grants additional capabilities to this role. :param capabilities_to_grant: Zero or more capabilities to grant this role. For a list of capabilities, see `Capabilities `_ on Splunk Developer Portal. :type capabilities_to_grant: ``string`` or ``list`` :return: The :class:`Role`. **Example**:: service = client.connect(...) role = service.roles['somerole'] role.grant('change_own_password', 'search') """ possible_capabilities = self.service.capabilities for capability in capabilities_to_grant: if capability not in possible_capabilities: raise NoSuchCapability(capability) new_capabilities = self['capabilities'] + list(capabilities_to_grant) self.post(capabilities=new_capabilities) return self def revoke(self, *capabilities_to_revoke): """Revokes zero or more capabilities from this role. :param capabilities_to_revoke: Zero or more capabilities to grant this role. For a list of capabilities, see `Capabilities `_ on Splunk Developer Portal. :type capabilities_to_revoke: ``string`` or ``list`` :return: The :class:`Role`. **Example**:: service = client.connect(...) role = service.roles['somerole'] role.revoke('change_own_password', 'search') """ possible_capabilities = self.service.capabilities for capability in capabilities_to_revoke: if capability not in possible_capabilities: raise NoSuchCapability(capability) old_capabilities = self['capabilities'] new_capabilities = [] for c in old_capabilities: if c not in capabilities_to_revoke: new_capabilities.append(c) if new_capabilities == []: new_capabilities = '' # Empty lists don't get passed in the body, so we have to force an empty argument. self.post(capabilities=new_capabilities) return self class Roles(Collection): """This class represents the collection of roles in the Splunk instance. Retrieve this collection using :meth:`Service.roles`.""" def __init__(self, service): return Collection.__init__(self, service, PATH_ROLES, item=Role) def __getitem__(self, key): return Collection.__getitem__(self, key.lower()) def __contains__(self, name): return Collection.__contains__(self, name.lower()) def create(self, name, **params): """Creates a new role. This function makes two roundtrips to the server, plus at most two more if the ``autologin`` field of :func:`connect` is set to ``True``. :param name: Name for the role. :type name: ``string`` :param params: Additional arguments (optional). For a list of available parameters, see `Roles parameters `_ on Splunk Developer Portal. :type params: ``dict`` :return: The new role. :rtype: :class:`Role` **Example**:: import splunklib.client as client c = client.connect(...) roles = c.roles paltry = roles.create("paltry", imported_roles="user", defaultApp="search") """ if not isinstance(name, basestring): raise ValueError("Invalid role name: %s" % str(name)) name = name.lower() self.post(name=name, **params) # splunkd doesn't return the user in the POST response body, # so we have to make a second round trip to fetch it. response = self.get(name) entry = _load_atom(response, XNAME_ENTRY).entry state = _parse_atom_entry(entry) entity = self.item( self.service, urllib.unquote(state.links.alternate), state=state) return entity def delete(self, name): """ Deletes the role and returns the resulting collection of roles. :param name: The name of the role to delete. :type name: ``string`` :rtype: The :class:`Roles` """ return Collection.delete(self, name.lower()) class Application(Entity): """Represents a locally-installed Splunk app.""" @property def setupInfo(self): """Returns the setup information for the app. :return: The setup information. """ return self.content.get('eai:setup', None) def package(self): """ Creates a compressed package of the app for archiving.""" return self._run_action("package") def updateInfo(self): """Returns any update information that is available for the app.""" return self._run_action("update")