The Salt integration tests come with a number of classes and methods which allow for components to be easily tested. These classes are generally inherited from and provide specific methods for hooking into the running integration test environment created by the integration tests.
It is noteworthy that since integration tests validate against a running environment that they are generally the preferred means to write tests.
The integration system is all located under tests/integration
in the Salt
source tree. Each directory within tests/integration
corresponds to a
directory in Salt's tree structure. For example, the integration tests for the
test.py
Salt module that is located in salt/modules
should also be
named test.py
and reside in tests/integration/modules
.
This guide assumes that your Salt development environment is already configured and that you have a basic understanding of contributing to the Salt codebase. If you're unfamiliar with either of these topics, please refer to the Installing Salt for Development and the Contributing pages, respectively.
This documentation also assumes that you have an understanding of how to run Salt's test suite, including running the test subsections, and running a single integration test file, class, or individual test.
Integration tests should be written to the following specifications.
Since integration tests are used to validate against a running Salt environment, integration tests should be written with the Salt components, and their various interactions, in mind.
Isolate testing functionality. Don't rely on the pass or failure of other, separate tests.
Individual tests should test against a single behavior.
Since it occasionally takes some effort to "set up" an individual test, it may be necessary to call several functions within a single test. However, be sure that once the work has been done to set up a test, make sure you are clear about the functionality that is being tested.
Test names and docstrings should indicate what functionality is being tested.
Test functions are named test_<fcn>_<test-name>
where <fcn>
is the
function being tested and <test-name>
describes the behavior being tested.
In order for integration tests to get picked up during a run of the test suite,
each individual test must be prepended with the test_
naming syntax, as
described above.
If a function does not start with test_
, then the function acts as a "normal"
function and is not considered a testing function. It will not be included in the
test run or testing output.
There are two special functions that can be utilized in the integration side of
Salt's test suite: setUp
and tearDown
. While these functions are not
required in all test files, there are many examples in Salt's integration
test suite illustrating the broad usefulness of each function.
The setUp
function is used to set up any repetitive or useful tasks that the
tests in a test class need before running. For example, any of the mac_*
integration tests should only run on macOS machines. The setUp
function can
be used to test for the presence of the Darwin
kernel. If the Darwin
kernel is not present, then the test should be skipped.
def setUp(self):
"""
Sets up test requirements
"""
os_grain = self.run_function("grains.item", ["kernel"])
if os_grain["kernel"] not in "Darwin":
self.skipTest("Test not applicable to '{kernel}' kernel".format(**os_grain))
The setUp
function can be used for many things. The above code snippet is
only one example. Another example might be to ensure that a particular setting
is present before running tests that would require the setting.
The tearDown
function is used to clean up after any tests. This function is
useful for restoring any settings that might have been changed during the test
run.
Note
The setUp
and tearDown
functions run before and after each test
in the test class that the setUp
and tearDown
functions are defined.
Be sure to read the Destructive vs Non-Destructive Tests section when
using any kind of destructive functions that might alter the system running the
test suite in either the setUp
or tearDown
function definitions.
The test functions within a test class do not run in the order they were defined, but instead run in lexicographical order.
Note that if any setUp
or tearDown
functions are defined in the class,
those functions will run before (for setUp
) or after (for tearDown
) each
test case.
The integration classes are located in tests/integration/__init__.py
and
can be extended therein. There are four classes available to extend:
Used to define executions run via the master to minions and to call single modules and states. The available testing functions are:
Run a single salt function and condition the return down to match the behavior of the raw function call. This will run the command and only return the results from a single minion to verify.
Run the state.single command and return the state return structure.
Run a single salt function on the 'minion' target and condition the return down to match the behavior of the raw function call.
Shell out to the scripts which ship with Salt. The testing functions are:
Execute salt-cp. Pass in the argument string as it would be passed on the command line.
Execute salt-call, pass in the argument string as it would be passed on the command line.
Execute the salt-cloud command. Pass in the argument string as it would be passed on the command line.
Execute the salt-key command. Pass in the argument string as it would be passed on the command line.
Execute the salt-run command. Pass in the argument string as it would be passed on the command line.
Execute the runner function the and return the return data and output in a dict
Execute the salt command. Pass in the argument string as it would be passed on the command line.
Execute a salt script with the given argument string.
Execute the salt-ssh. Pass in the argument string as it would be passed on the command line.
Used to execute remote commands via salt-ssh. The available methods are as follows:
Run a single salt function via salt-ssh and condition the return down to match the behavior of the raw function call. This will run the command and only return the results from a single minion to verify.
Used to execute remote commands via a syndic and is only used to verify the capabilities of the Salt Syndic. The available methods are as follows:
Run a single salt function and condition the return down to match the behavior of the raw function call. This will run the command and only return the results from a single minion to verify.
The following sections define simple integration tests present in Salt's integration test suite for each type of testing class.
Import the integration module, this module is already added to the python path
by the test execution. Inherit from the integration.ModuleCase
class.
Now the workhorse method run_function
can be used to test a module:
import os
from tests.support.case import ModuleCase
class TestModuleTest(ModuleCase):
"""
Validate the test module
"""
def test_ping(self):
"""
test.ping
"""
self.assertTrue(self.run_function("test.ping"))
def test_echo(self):
"""
test.echo
"""
self.assertEqual(self.run_function("test.echo", ["text"]), "text")
The fist example illustrates the testing master issuing a test.ping
call
to a testing minion. The test asserts that the minion returned with a True
value to the master from the test.ping
call.
The second example similarly verifies that the minion executed the
test.echo
command with the text
argument. The assertEqual
call
maintains that the minion ran the function and returned the data as expected
to the master.
Validating the shell commands can be done via shell tests:
import sys
import shutil
import tempfile
from tests.support.case import ShellCase
class KeyTest(ShellCase):
"""
Test salt-key script
"""
_call_binary_ = "salt-key"
def test_list(self):
"""
test salt-key -L
"""
data = self.run_key("-L")
expect = [
"Unaccepted Keys:",
"Accepted Keys:",
"minion",
"sub_minion",
"Rejected:",
"",
]
self.assertEqual(data, expect)
This example verifies that the salt-key
command executes and returns as
expected by making use of the run_key
method.
Testing salt-ssh functionality can be done using the SSHCase test class:
from tests.support.case import SSHCase
class SSHGrainsTest(SSHCase):
"""
Test salt-ssh grains functionality
Depend on proper environment set by integration.SSHCase class
"""
def test_grains_id(self):
"""
Test salt-ssh grains id work for localhost.
"""
cmd = self.run_function("grains.get", ["id"])
self.assertEqual(cmd, "localhost")
The fundamentially asynchronous nature of Salt makes testing the event system a challenge.
The SaltMinionEventAssertsMixin
provides a facility for testing that events were received
on a minion event bus.
import salt.utils.event
from tests.support.mixins import SaltEventAssertsMixin
class TestEvent(SaltEventAssertsMixin):
"""
Example test of firing an event and receiving it
"""
def test_event(self):
e = salt.utils.event.get_event(
"minion", sock_dir=self.minion_opts["sock_dir"], opts=self.minion_opts
)
e.fire_event({"a": "b"}, "/test_event")
self.assertMinionEventReceived({"a": "b"})
Testing Salt's Syndic can be done via the SyndicCase test class:
from tests.support.case import SyndicCase
class TestSyndic(SyndicCase):
"""
Validate the syndic interface by testing the test module
"""
def test_ping(self):
"""
test.ping
"""
self.assertTrue(self.run_function("test.ping"))
This example verifies that a test.ping
command is issued from the testing
master, is passed through to the testing syndic, down to the minion, and back
up again by using the run_function
located with in the SyndicCase
test
class.
Since using Salt largely involves configuring states, editing files, and changing
system data, the integration test suite contains a directory named files
to
aid in testing functions that require files. Various Salt integration tests use
these example files to test against instead of altering system files and data.
Each directory within tests/integration/files
contain files that accomplish
different tasks, based on the needs of the integration tests using those files.
For example, tests/integration/files/ssh
is used to bootstrap the test runner
for salt-ssh testing, while tests/integration/files/pillar
contains files
storing data needed to test various pillar functions.
The tests/integration/files
directory also includes an integration state tree.
The integration state tree can be found at tests/integration/files/file/base
.
The following example demonstrates how integration files can be used with ModuleCase to test states:
# Import python libs
from __future__ import absolute_import
import os
import shutil
# Import Salt Testing libs
from tests.support.case import ModuleCase
from tests.support.paths import FILES, TMP
from tests.support.mixins import SaltReturnAssertsMixin
# Import salt libs
import salt.utils.files
HFILE = os.path.join(TMP, "hosts")
class HostTest(ModuleCase, SaltReturnAssertsMixin):
"""
Validate the host state
"""
def setUp(self):
shutil.copyfile(os.path.join(FILES, "hosts"), HFILE)
super(HostTest, self).setUp()
def tearDown(self):
if os.path.exists(HFILE):
os.remove(HFILE)
super(HostTest, self).tearDown()
def test_present(self):
"""
host.present
"""
name = "spam.bacon"
ip = "10.10.10.10"
ret = self.run_state("host.present", name=name, ip=ip)
self.assertSaltTrueReturn(ret)
with salt.utils.files.fopen(HFILE) as fp_:
output = fp_.read()
self.assertIn("{0}\t\t{1}".format(ip, name), output)
To access the integration files, a variable named FILES
points to the
tests/integration/files
directory. This is where the referenced
host.present
sls file resides.
In addition to the static files in the integration state tree, the location
TMP
can also be used to store temporary files that the test system
will clean up when the execution finishes.
Since Salt is used to change the settings and behavior of systems, one testing approach is to run tests that make actual changes to the underlying system. This is where the concept of destructive integration tests comes into play. Tests can be written to alter the system they are running on. This capability is what fills in the gap needed to properly test aspects of system management like package installation.
Any test that changes the underlying system in any way, such as creating or
deleting users, installing packages, or changing permissions should include the
@destructive
decorator to signal system changes and should be written with
care. System changes executed within a destructive test should also be restored
once the related tests have completed. For example, if a new user is created to
test a module, the same user should be removed after the test is completed to
maintain system integrity.
To write a destructive test, import, and use the destructiveTest decorator for the test method:
from tests.support.case import ModuleCase
from tests.support.helpers import destructiveTest, skip_if_not_root
class DestructiveExampleModuleTest(ModuleCase):
"""
Demonstrate a destructive test
"""
@destructiveTest
@skip_if_not_root
def test_user_not_present(self):
"""
This is a DESTRUCTIVE TEST it creates a new user on the minion.
And then destroys that user.
"""
ret = self.run_state("user.present", name="salt_test")
self.assertSaltTrueReturn(ret)
ret = self.run_state("user.absent", name="salt_test")
self.assertSaltTrueReturn(ret)
Cloud provider integration tests are used to assess Salt-Cloud's ability to create and destroy cloud instances for various supported cloud providers. Cloud provider tests inherit from the ShellCase Integration Class.
Any new cloud provider test files should be added to the tests/integration/cloud/providers/
directory. Each cloud provider test file also requires a sample cloud profile and cloud
provider configuration file in the integration test file directory located at
tests/integration/files/conf/cloud.*.d/
.
The following is an example of the default profile configuration file for Digital
Ocean, located at: tests/integration/files/conf/cloud.profiles.d/digitalocean.conf
:
digitalocean-test:
provider: digitalocean-config
image: Ubuntu 14.04 x64
size: 512MB
Each cloud provider requires different configuration credentials. Therefore, sensitive information such as API keys or passwords should be omitted from the cloud provider configuration file and replaced with an empty string. The necessary credentials can be provided by the user by editing the provider configuration file before running the tests.
The following is an example of the default provider configuration file for Digital
Ocean, located at: tests/integration/files/conf/cloud.providers.d/digitalocean.conf
:
digitalocean-config:
driver: digitalocean
client_key: ''
api_key: ''
location: New York 1
In addition to providing the necessary cloud profile and provider files in the integration
test suite file structure, appropriate checks for if the configuration files exist and
contain valid information are also required in the test class's setUp
function:
from tests.support.case import ShellCase
from tests.support.paths import FILES
class LinodeTest(ShellCase):
"""
Integration tests for the Linode cloud provider in Salt-Cloud
"""
def setUp(self):
"""
Sets up the test requirements
"""
super(LinodeTest, self).setUp()
# check if appropriate cloud provider and profile files are present
profile_str = "linode-config:"
provider = "linode"
providers = self.run_cloud("--list-providers")
if profile_str not in providers:
self.skipTest(
"Configuration file for {0} was not found. Check {0}.conf files "
"in tests/integration/files/conf/cloud.*.d/ to run these tests.".format(
provider
)
)
# check if apikey and password are present
path = os.path.join(FILES, "conf", "cloud.providers.d", provider + ".conf")
config = cloud_providers_config(path)
api = config["linode-config"]["linode"]["apikey"]
password = config["linode-config"]["linode"]["password"]
if api == "" or password == "":
self.skipTest(
"An api key and password must be provided to run these tests. Check "
"tests/integration/files/conf/cloud.providers.d/{0}.conf".format(
provider
)
)
Repeatedly creating and destroying instances on cloud providers can be costly.
Therefore, cloud provider tests are off by default and do not run automatically. To
run the cloud provider tests, the --cloud-provider-tests
flag must be provided:
./tests/runtests.py --cloud-provider-tests
Since cloud provider tests do not run automatically, all provider tests must be
preceded with the @expensiveTest
decorator. The expensive test decorator is
necessary because it signals to the test suite that the
--cloud-provider-tests
flag is required to run the cloud provider tests.
To write a cloud provider test, import, and use the expensiveTest decorator for the test function:
from tests.support.helpers import expensiveTest
@expensiveTest
def test_instance(self):
"""
Test creating an instance on Linode
"""
name = "linode-testing"
# create the instance
instance = self.run_cloud("-p linode-test {0}".format(name))
str = " {0}".format(name)
# check if instance with salt installed returned as expected
try:
self.assertIn(str, instance)
except AssertionError:
self.run_cloud("-d {0} --assume-yes".format(name))
raise
# delete the instance
delete = self.run_cloud("-d {0} --assume-yes".format(name))
str = " True"
try:
self.assertIn(str, delete)
except AssertionError:
raise
If the corresponding Salt directory does not exist within
tests/integration
, the new directory must be created along with the
appropriate test file to maintain Salt's testing directory structure.
In order for Salt's test suite to recognize tests within the newly
created directory, options to run the new integration tests must be added to
tests/runtests.py
. Examples of the necessary options that must be added
can be found here: tests/runtests.py. The functions that need to be
edited are setup_additional_options
, validate_options
, and
run_integration_tests
.