mirror of
https://github.com/tbsdtv/linux_media.git
synced 2025-07-23 12:43:29 +02:00
Merge tag 'linux-kselftest-kunit-5.12-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/shuah/linux-kselftest
Pull KUnit updates from Shuah Khan - support for filtering test suites using glob from Daniel Latypov. "kunit_filter.glob" command line option is passed to the UML kernel, which currently only supports filtering by suite name. This support allows running different subsets of tests, e.g. $ ./tools/testing/kunit/kunit.py build $ ./tools/testing/kunit/kunit.py exec 'list*' $ ./tools/testing/kunit/kunit.py exec 'kunit*' - several fixes and cleanups also from Daniel Latypov. * tag 'linux-kselftest-kunit-5.12-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/shuah/linux-kselftest: kunit: tool: fix unintentional statefulness in run_kernel() kunit: tool: add support for filtering suites by glob kunit: add kunit.filter_glob cmdline option to filter suites kunit: don't show `1 == 1` in failed assertion messages kunit: make kunit_tool accept optional path to .kunitconfig fragment Documentation: kunit: add tips.rst for small examples KUnit: Docs: make start.rst example Kconfig follow style.rst kunit: tool: simplify kconfig is_subset_of() logic minor: kunit: tool: fix unit test so it can run from non-root dir kunit: tool: use `with open()` in unit test kunit: tool: stop using bare asserts in unit test kunit: tool: fix unit test cleanup handling
This commit is contained in:
@@ -13,6 +13,7 @@ KUnit - Unit Testing for the Linux Kernel
|
|||||||
api/index
|
api/index
|
||||||
style
|
style
|
||||||
faq
|
faq
|
||||||
|
tips
|
||||||
|
|
||||||
What is KUnit?
|
What is KUnit?
|
||||||
==============
|
==============
|
||||||
@@ -88,6 +89,7 @@ How do I use it?
|
|||||||
================
|
================
|
||||||
|
|
||||||
* :doc:`start` - for new users of KUnit
|
* :doc:`start` - for new users of KUnit
|
||||||
|
* :doc:`tips` - for short examples of best practices
|
||||||
* :doc:`usage` - for a more detailed explanation of KUnit features
|
* :doc:`usage` - for a more detailed explanation of KUnit features
|
||||||
* :doc:`api/index` - for the list of KUnit APIs used for testing
|
* :doc:`api/index` - for the list of KUnit APIs used for testing
|
||||||
* :doc:`kunit-tool` - for more information on the kunit_tool helper script
|
* :doc:`kunit-tool` - for more information on the kunit_tool helper script
|
||||||
|
@@ -196,8 +196,9 @@ Now add the following to ``drivers/misc/Kconfig``:
|
|||||||
.. code-block:: kconfig
|
.. code-block:: kconfig
|
||||||
|
|
||||||
config MISC_EXAMPLE_TEST
|
config MISC_EXAMPLE_TEST
|
||||||
bool "Test for my example"
|
tristate "Test for my example" if !KUNIT_ALL_TESTS
|
||||||
depends on MISC_EXAMPLE && KUNIT=y
|
depends on MISC_EXAMPLE && KUNIT=y
|
||||||
|
default KUNIT_ALL_TESTS
|
||||||
|
|
||||||
and the following to ``drivers/misc/Makefile``:
|
and the following to ``drivers/misc/Makefile``:
|
||||||
|
|
||||||
@@ -233,5 +234,7 @@ Congrats! You just wrote your first KUnit test!
|
|||||||
|
|
||||||
Next Steps
|
Next Steps
|
||||||
==========
|
==========
|
||||||
* Check out the :doc:`usage` page for a more
|
* Check out the :doc:`tips` page for tips on
|
||||||
|
writing idiomatic KUnit tests.
|
||||||
|
* Optional: see the :doc:`usage` page for a more
|
||||||
in-depth explanation of KUnit.
|
in-depth explanation of KUnit.
|
||||||
|
115
Documentation/dev-tools/kunit/tips.rst
Normal file
115
Documentation/dev-tools/kunit/tips.rst
Normal file
@@ -0,0 +1,115 @@
|
|||||||
|
.. SPDX-License-Identifier: GPL-2.0
|
||||||
|
|
||||||
|
============================
|
||||||
|
Tips For Writing KUnit Tests
|
||||||
|
============================
|
||||||
|
|
||||||
|
Exiting early on failed expectations
|
||||||
|
------------------------------------
|
||||||
|
|
||||||
|
``KUNIT_EXPECT_EQ`` and friends will mark the test as failed and continue
|
||||||
|
execution. In some cases, it's unsafe to continue and you can use the
|
||||||
|
``KUNIT_ASSERT`` variant to exit on failure.
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
void example_test_user_alloc_function(struct kunit *test)
|
||||||
|
{
|
||||||
|
void *object = alloc_some_object_for_me();
|
||||||
|
|
||||||
|
/* Make sure we got a valid pointer back. */
|
||||||
|
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, object);
|
||||||
|
do_something_with_object(object);
|
||||||
|
}
|
||||||
|
|
||||||
|
Allocating memory
|
||||||
|
-----------------
|
||||||
|
|
||||||
|
Where you would use ``kzalloc``, you should prefer ``kunit_kzalloc`` instead.
|
||||||
|
KUnit will ensure the memory is freed once the test completes.
|
||||||
|
|
||||||
|
This is particularly useful since it lets you use the ``KUNIT_ASSERT_EQ``
|
||||||
|
macros to exit early from a test without having to worry about remembering to
|
||||||
|
call ``kfree``.
|
||||||
|
|
||||||
|
Example:
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
void example_test_allocation(struct kunit *test)
|
||||||
|
{
|
||||||
|
char *buffer = kunit_kzalloc(test, 16, GFP_KERNEL);
|
||||||
|
/* Ensure allocation succeeded. */
|
||||||
|
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, buffer);
|
||||||
|
|
||||||
|
KUNIT_ASSERT_STREQ(test, buffer, "");
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
Testing static functions
|
||||||
|
------------------------
|
||||||
|
|
||||||
|
If you don't want to expose functions or variables just for testing, one option
|
||||||
|
is to conditionally ``#include`` the test file at the end of your .c file, e.g.
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
/* In my_file.c */
|
||||||
|
|
||||||
|
static int do_interesting_thing();
|
||||||
|
|
||||||
|
#ifdef CONFIG_MY_KUNIT_TEST
|
||||||
|
#include "my_kunit_test.c"
|
||||||
|
#endif
|
||||||
|
|
||||||
|
Injecting test-only code
|
||||||
|
------------------------
|
||||||
|
|
||||||
|
Similarly to the above, it can be useful to add test-specific logic.
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
/* In my_file.h */
|
||||||
|
|
||||||
|
#ifdef CONFIG_MY_KUNIT_TEST
|
||||||
|
/* Defined in my_kunit_test.c */
|
||||||
|
void test_only_hook(void);
|
||||||
|
#else
|
||||||
|
void test_only_hook(void) { }
|
||||||
|
#endif
|
||||||
|
|
||||||
|
TODO(dlatypov@google.com): add an example of using ``current->kunit_test`` in
|
||||||
|
such a hook when it's not only updated for ``CONFIG_KASAN=y``.
|
||||||
|
|
||||||
|
Customizing error messages
|
||||||
|
--------------------------
|
||||||
|
|
||||||
|
Each of the ``KUNIT_EXPECT`` and ``KUNIT_ASSERT`` macros have a ``_MSG`` variant.
|
||||||
|
These take a format string and arguments to provide additional context to the automatically generated error messages.
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
char some_str[41];
|
||||||
|
generate_sha1_hex_string(some_str);
|
||||||
|
|
||||||
|
/* Before. Not easy to tell why the test failed. */
|
||||||
|
KUNIT_EXPECT_EQ(test, strlen(some_str), 40);
|
||||||
|
|
||||||
|
/* After. Now we see the offending string. */
|
||||||
|
KUNIT_EXPECT_EQ_MSG(test, strlen(some_str), 40, "some_str='%s'", some_str);
|
||||||
|
|
||||||
|
Alternatively, one can take full control over the error message by using ``KUNIT_FAIL()``, e.g.
|
||||||
|
|
||||||
|
.. code-block:: c
|
||||||
|
|
||||||
|
/* Before */
|
||||||
|
KUNIT_EXPECT_EQ(test, some_setup_function(), 0);
|
||||||
|
|
||||||
|
/* After: full control over the failure message. */
|
||||||
|
if (some_setup_function())
|
||||||
|
KUNIT_FAIL(test, "Failed to setup thing for testing");
|
||||||
|
|
||||||
|
Next Steps
|
||||||
|
==========
|
||||||
|
* Optional: see the :doc:`usage` page for a more
|
||||||
|
in-depth explanation of KUnit.
|
@@ -4,6 +4,7 @@
|
|||||||
|
|
||||||
menuconfig KUNIT
|
menuconfig KUNIT
|
||||||
tristate "KUnit - Enable support for unit tests"
|
tristate "KUnit - Enable support for unit tests"
|
||||||
|
select GLOB if KUNIT=y
|
||||||
help
|
help
|
||||||
Enables support for kernel unit tests (KUnit), a lightweight unit
|
Enables support for kernel unit tests (KUnit), a lightweight unit
|
||||||
testing and mocking framework for the Linux kernel. These tests are
|
testing and mocking framework for the Linux kernel. These tests are
|
||||||
|
@@ -85,6 +85,29 @@ void kunit_ptr_not_err_assert_format(const struct kunit_assert *assert,
|
|||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(kunit_ptr_not_err_assert_format);
|
EXPORT_SYMBOL_GPL(kunit_ptr_not_err_assert_format);
|
||||||
|
|
||||||
|
/* Checks if `text` is a literal representing `value`, e.g. "5" and 5 */
|
||||||
|
static bool is_literal(struct kunit *test, const char *text, long long value,
|
||||||
|
gfp_t gfp)
|
||||||
|
{
|
||||||
|
char *buffer;
|
||||||
|
int len;
|
||||||
|
bool ret;
|
||||||
|
|
||||||
|
len = snprintf(NULL, 0, "%lld", value);
|
||||||
|
if (strlen(text) != len)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
buffer = kunit_kmalloc(test, len+1, gfp);
|
||||||
|
if (!buffer)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
snprintf(buffer, len+1, "%lld", value);
|
||||||
|
ret = strncmp(buffer, text, len) == 0;
|
||||||
|
|
||||||
|
kunit_kfree(test, buffer);
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
void kunit_binary_assert_format(const struct kunit_assert *assert,
|
void kunit_binary_assert_format(const struct kunit_assert *assert,
|
||||||
struct string_stream *stream)
|
struct string_stream *stream)
|
||||||
{
|
{
|
||||||
@@ -97,12 +120,16 @@ void kunit_binary_assert_format(const struct kunit_assert *assert,
|
|||||||
binary_assert->left_text,
|
binary_assert->left_text,
|
||||||
binary_assert->operation,
|
binary_assert->operation,
|
||||||
binary_assert->right_text);
|
binary_assert->right_text);
|
||||||
string_stream_add(stream, KUNIT_SUBSUBTEST_INDENT "%s == %lld\n",
|
if (!is_literal(stream->test, binary_assert->left_text,
|
||||||
binary_assert->left_text,
|
binary_assert->left_value, stream->gfp))
|
||||||
binary_assert->left_value);
|
string_stream_add(stream, KUNIT_SUBSUBTEST_INDENT "%s == %lld\n",
|
||||||
string_stream_add(stream, KUNIT_SUBSUBTEST_INDENT "%s == %lld",
|
binary_assert->left_text,
|
||||||
binary_assert->right_text,
|
binary_assert->left_value);
|
||||||
binary_assert->right_value);
|
if (!is_literal(stream->test, binary_assert->right_text,
|
||||||
|
binary_assert->right_value, stream->gfp))
|
||||||
|
string_stream_add(stream, KUNIT_SUBSUBTEST_INDENT "%s == %lld",
|
||||||
|
binary_assert->right_text,
|
||||||
|
binary_assert->right_value);
|
||||||
kunit_assert_print_msg(assert, stream);
|
kunit_assert_print_msg(assert, stream);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(kunit_binary_assert_format);
|
EXPORT_SYMBOL_GPL(kunit_binary_assert_format);
|
||||||
|
@@ -1,6 +1,8 @@
|
|||||||
// SPDX-License-Identifier: GPL-2.0
|
// SPDX-License-Identifier: GPL-2.0
|
||||||
|
|
||||||
#include <kunit/test.h>
|
#include <kunit/test.h>
|
||||||
|
#include <linux/glob.h>
|
||||||
|
#include <linux/moduleparam.h>
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* These symbols point to the .kunit_test_suites section and are defined in
|
* These symbols point to the .kunit_test_suites section and are defined in
|
||||||
@@ -11,14 +13,81 @@ extern struct kunit_suite * const * const __kunit_suites_end[];
|
|||||||
|
|
||||||
#if IS_BUILTIN(CONFIG_KUNIT)
|
#if IS_BUILTIN(CONFIG_KUNIT)
|
||||||
|
|
||||||
static void kunit_print_tap_header(void)
|
static char *filter_glob;
|
||||||
|
module_param(filter_glob, charp, 0);
|
||||||
|
MODULE_PARM_DESC(filter_glob,
|
||||||
|
"Filter which KUnit test suites run at boot-time, e.g. list*");
|
||||||
|
|
||||||
|
static struct kunit_suite * const *
|
||||||
|
kunit_filter_subsuite(struct kunit_suite * const * const subsuite)
|
||||||
|
{
|
||||||
|
int i, n = 0;
|
||||||
|
struct kunit_suite **filtered;
|
||||||
|
|
||||||
|
n = 0;
|
||||||
|
for (i = 0; subsuite[i] != NULL; ++i) {
|
||||||
|
if (glob_match(filter_glob, subsuite[i]->name))
|
||||||
|
++n;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (n == 0)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
filtered = kmalloc_array(n + 1, sizeof(*filtered), GFP_KERNEL);
|
||||||
|
if (!filtered)
|
||||||
|
return NULL;
|
||||||
|
|
||||||
|
n = 0;
|
||||||
|
for (i = 0; subsuite[i] != NULL; ++i) {
|
||||||
|
if (glob_match(filter_glob, subsuite[i]->name))
|
||||||
|
filtered[n++] = subsuite[i];
|
||||||
|
}
|
||||||
|
filtered[n] = NULL;
|
||||||
|
|
||||||
|
return filtered;
|
||||||
|
}
|
||||||
|
|
||||||
|
struct suite_set {
|
||||||
|
struct kunit_suite * const * const *start;
|
||||||
|
struct kunit_suite * const * const *end;
|
||||||
|
};
|
||||||
|
|
||||||
|
static struct suite_set kunit_filter_suites(void)
|
||||||
|
{
|
||||||
|
int i;
|
||||||
|
struct kunit_suite * const **copy, * const *filtered_subsuite;
|
||||||
|
struct suite_set filtered;
|
||||||
|
|
||||||
|
const size_t max = __kunit_suites_end - __kunit_suites_start;
|
||||||
|
|
||||||
|
if (!filter_glob) {
|
||||||
|
filtered.start = __kunit_suites_start;
|
||||||
|
filtered.end = __kunit_suites_end;
|
||||||
|
return filtered;
|
||||||
|
}
|
||||||
|
|
||||||
|
copy = kmalloc_array(max, sizeof(*filtered.start), GFP_KERNEL);
|
||||||
|
filtered.start = copy;
|
||||||
|
if (!copy) { /* won't be able to run anything, return an empty set */
|
||||||
|
filtered.end = copy;
|
||||||
|
return filtered;
|
||||||
|
}
|
||||||
|
|
||||||
|
for (i = 0; i < max; ++i) {
|
||||||
|
filtered_subsuite = kunit_filter_subsuite(__kunit_suites_start[i]);
|
||||||
|
if (filtered_subsuite)
|
||||||
|
*copy++ = filtered_subsuite;
|
||||||
|
}
|
||||||
|
filtered.end = copy;
|
||||||
|
return filtered;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void kunit_print_tap_header(struct suite_set *suite_set)
|
||||||
{
|
{
|
||||||
struct kunit_suite * const * const *suites, * const *subsuite;
|
struct kunit_suite * const * const *suites, * const *subsuite;
|
||||||
int num_of_suites = 0;
|
int num_of_suites = 0;
|
||||||
|
|
||||||
for (suites = __kunit_suites_start;
|
for (suites = suite_set->start; suites < suite_set->end; suites++)
|
||||||
suites < __kunit_suites_end;
|
|
||||||
suites++)
|
|
||||||
for (subsuite = *suites; *subsuite != NULL; subsuite++)
|
for (subsuite = *suites; *subsuite != NULL; subsuite++)
|
||||||
num_of_suites++;
|
num_of_suites++;
|
||||||
|
|
||||||
@@ -30,12 +99,18 @@ int kunit_run_all_tests(void)
|
|||||||
{
|
{
|
||||||
struct kunit_suite * const * const *suites;
|
struct kunit_suite * const * const *suites;
|
||||||
|
|
||||||
kunit_print_tap_header();
|
struct suite_set suite_set = kunit_filter_suites();
|
||||||
|
|
||||||
for (suites = __kunit_suites_start;
|
kunit_print_tap_header(&suite_set);
|
||||||
suites < __kunit_suites_end;
|
|
||||||
suites++)
|
for (suites = suite_set.start; suites < suite_set.end; suites++)
|
||||||
__kunit_test_suites_init(*suites);
|
__kunit_test_suites_init(*suites);
|
||||||
|
|
||||||
|
if (filter_glob) { /* a copy was made of each array */
|
||||||
|
for (suites = suite_set.start; suites < suite_set.end; suites++)
|
||||||
|
kfree(*suites);
|
||||||
|
kfree(suite_set.start);
|
||||||
|
}
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
@@ -28,12 +28,12 @@ KunitBuildRequest = namedtuple('KunitBuildRequest',
|
|||||||
['jobs', 'build_dir', 'alltests',
|
['jobs', 'build_dir', 'alltests',
|
||||||
'make_options'])
|
'make_options'])
|
||||||
KunitExecRequest = namedtuple('KunitExecRequest',
|
KunitExecRequest = namedtuple('KunitExecRequest',
|
||||||
['timeout', 'build_dir', 'alltests'])
|
['timeout', 'build_dir', 'alltests', 'filter_glob'])
|
||||||
KunitParseRequest = namedtuple('KunitParseRequest',
|
KunitParseRequest = namedtuple('KunitParseRequest',
|
||||||
['raw_output', 'input_data', 'build_dir', 'json'])
|
['raw_output', 'input_data', 'build_dir', 'json'])
|
||||||
KunitRequest = namedtuple('KunitRequest', ['raw_output','timeout', 'jobs',
|
KunitRequest = namedtuple('KunitRequest', ['raw_output','timeout', 'jobs',
|
||||||
'build_dir', 'alltests', 'json',
|
'build_dir', 'alltests', 'filter_glob',
|
||||||
'make_options'])
|
'json', 'make_options'])
|
||||||
|
|
||||||
KernelDirectoryPath = sys.argv[0].split('tools/testing/kunit/')[0]
|
KernelDirectoryPath = sys.argv[0].split('tools/testing/kunit/')[0]
|
||||||
|
|
||||||
@@ -93,6 +93,7 @@ def exec_tests(linux: kunit_kernel.LinuxSourceTree,
|
|||||||
test_start = time.time()
|
test_start = time.time()
|
||||||
result = linux.run_kernel(
|
result = linux.run_kernel(
|
||||||
timeout=None if request.alltests else request.timeout,
|
timeout=None if request.alltests else request.timeout,
|
||||||
|
filter_glob=request.filter_glob,
|
||||||
build_dir=request.build_dir)
|
build_dir=request.build_dir)
|
||||||
|
|
||||||
test_end = time.time()
|
test_end = time.time()
|
||||||
@@ -149,7 +150,7 @@ def run_tests(linux: kunit_kernel.LinuxSourceTree,
|
|||||||
return build_result
|
return build_result
|
||||||
|
|
||||||
exec_request = KunitExecRequest(request.timeout, request.build_dir,
|
exec_request = KunitExecRequest(request.timeout, request.build_dir,
|
||||||
request.alltests)
|
request.alltests, request.filter_glob)
|
||||||
exec_result = exec_tests(linux, exec_request)
|
exec_result = exec_tests(linux, exec_request)
|
||||||
if exec_result.status != KunitStatus.SUCCESS:
|
if exec_result.status != KunitStatus.SUCCESS:
|
||||||
return exec_result
|
return exec_result
|
||||||
@@ -182,6 +183,9 @@ def add_common_opts(parser) -> None:
|
|||||||
parser.add_argument('--alltests',
|
parser.add_argument('--alltests',
|
||||||
help='Run all KUnit tests through allyesconfig',
|
help='Run all KUnit tests through allyesconfig',
|
||||||
action='store_true')
|
action='store_true')
|
||||||
|
parser.add_argument('--kunitconfig',
|
||||||
|
help='Path to Kconfig fragment that enables KUnit tests',
|
||||||
|
metavar='kunitconfig')
|
||||||
|
|
||||||
def add_build_opts(parser) -> None:
|
def add_build_opts(parser) -> None:
|
||||||
parser.add_argument('--jobs',
|
parser.add_argument('--jobs',
|
||||||
@@ -197,6 +201,14 @@ def add_exec_opts(parser) -> None:
|
|||||||
type=int,
|
type=int,
|
||||||
default=300,
|
default=300,
|
||||||
metavar='timeout')
|
metavar='timeout')
|
||||||
|
parser.add_argument('filter_glob',
|
||||||
|
help='maximum number of seconds to allow for all tests '
|
||||||
|
'to run. This does not include time taken to build the '
|
||||||
|
'tests.',
|
||||||
|
type=str,
|
||||||
|
nargs='?',
|
||||||
|
default='',
|
||||||
|
metavar='filter_glob')
|
||||||
|
|
||||||
def add_parse_opts(parser) -> None:
|
def add_parse_opts(parser) -> None:
|
||||||
parser.add_argument('--raw_output', help='don\'t format output from kernel',
|
parser.add_argument('--raw_output', help='don\'t format output from kernel',
|
||||||
@@ -256,13 +268,14 @@ def main(argv, linux=None):
|
|||||||
os.mkdir(cli_args.build_dir)
|
os.mkdir(cli_args.build_dir)
|
||||||
|
|
||||||
if not linux:
|
if not linux:
|
||||||
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir)
|
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir, kunitconfig_path=cli_args.kunitconfig)
|
||||||
|
|
||||||
request = KunitRequest(cli_args.raw_output,
|
request = KunitRequest(cli_args.raw_output,
|
||||||
cli_args.timeout,
|
cli_args.timeout,
|
||||||
cli_args.jobs,
|
cli_args.jobs,
|
||||||
cli_args.build_dir,
|
cli_args.build_dir,
|
||||||
cli_args.alltests,
|
cli_args.alltests,
|
||||||
|
cli_args.filter_glob,
|
||||||
cli_args.json,
|
cli_args.json,
|
||||||
cli_args.make_options)
|
cli_args.make_options)
|
||||||
result = run_tests(linux, request)
|
result = run_tests(linux, request)
|
||||||
@@ -274,7 +287,7 @@ def main(argv, linux=None):
|
|||||||
os.mkdir(cli_args.build_dir)
|
os.mkdir(cli_args.build_dir)
|
||||||
|
|
||||||
if not linux:
|
if not linux:
|
||||||
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir)
|
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir, kunitconfig_path=cli_args.kunitconfig)
|
||||||
|
|
||||||
request = KunitConfigRequest(cli_args.build_dir,
|
request = KunitConfigRequest(cli_args.build_dir,
|
||||||
cli_args.make_options)
|
cli_args.make_options)
|
||||||
@@ -286,7 +299,7 @@ def main(argv, linux=None):
|
|||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
elif cli_args.subcommand == 'build':
|
elif cli_args.subcommand == 'build':
|
||||||
if not linux:
|
if not linux:
|
||||||
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir)
|
linux = kunit_kernel.LinuxSourceTree(cli_args.build_dir, kunitconfig_path=cli_args.kunitconfig)
|
||||||
|
|
||||||
request = KunitBuildRequest(cli_args.jobs,
|
request = KunitBuildRequest(cli_args.jobs,
|
||||||
cli_args.build_dir,
|
cli_args.build_dir,
|
||||||
@@ -304,7 +317,8 @@ def main(argv, linux=None):
|
|||||||
|
|
||||||
exec_request = KunitExecRequest(cli_args.timeout,
|
exec_request = KunitExecRequest(cli_args.timeout,
|
||||||
cli_args.build_dir,
|
cli_args.build_dir,
|
||||||
cli_args.alltests)
|
cli_args.alltests,
|
||||||
|
cli_args.filter_glob)
|
||||||
exec_result = exec_tests(linux, exec_request)
|
exec_result = exec_tests(linux, exec_request)
|
||||||
parse_request = KunitParseRequest(cli_args.raw_output,
|
parse_request = KunitParseRequest(cli_args.raw_output,
|
||||||
exec_result.result,
|
exec_result.result,
|
||||||
|
@@ -41,15 +41,14 @@ class Kconfig(object):
|
|||||||
self._entries.append(entry)
|
self._entries.append(entry)
|
||||||
|
|
||||||
def is_subset_of(self, other: 'Kconfig') -> bool:
|
def is_subset_of(self, other: 'Kconfig') -> bool:
|
||||||
|
other_dict = {e.name: e.value for e in other.entries()}
|
||||||
for a in self.entries():
|
for a in self.entries():
|
||||||
found = False
|
b = other_dict.get(a.name)
|
||||||
for b in other.entries():
|
if b is None:
|
||||||
if a.name != b.name:
|
if a.value == 'n':
|
||||||
continue
|
continue
|
||||||
if a.value != b.value:
|
return False
|
||||||
return False
|
elif a.value != b:
|
||||||
found = True
|
|
||||||
if a.value != 'n' and found == False:
|
|
||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@@ -123,7 +123,7 @@ def get_outfile_path(build_dir) -> str:
|
|||||||
class LinuxSourceTree(object):
|
class LinuxSourceTree(object):
|
||||||
"""Represents a Linux kernel source tree with KUnit tests."""
|
"""Represents a Linux kernel source tree with KUnit tests."""
|
||||||
|
|
||||||
def __init__(self, build_dir: str, load_config=True, defconfig=DEFAULT_KUNITCONFIG_PATH) -> None:
|
def __init__(self, build_dir: str, load_config=True, kunitconfig_path='') -> None:
|
||||||
signal.signal(signal.SIGINT, self.signal_handler)
|
signal.signal(signal.SIGINT, self.signal_handler)
|
||||||
|
|
||||||
self._ops = LinuxSourceTreeOperations()
|
self._ops = LinuxSourceTreeOperations()
|
||||||
@@ -131,9 +131,13 @@ class LinuxSourceTree(object):
|
|||||||
if not load_config:
|
if not load_config:
|
||||||
return
|
return
|
||||||
|
|
||||||
kunitconfig_path = get_kunitconfig_path(build_dir)
|
if kunitconfig_path:
|
||||||
if not os.path.exists(kunitconfig_path):
|
if not os.path.exists(kunitconfig_path):
|
||||||
shutil.copyfile(defconfig, kunitconfig_path)
|
raise ConfigError(f'Specified kunitconfig ({kunitconfig_path}) does not exist')
|
||||||
|
else:
|
||||||
|
kunitconfig_path = get_kunitconfig_path(build_dir)
|
||||||
|
if not os.path.exists(kunitconfig_path):
|
||||||
|
shutil.copyfile(DEFAULT_KUNITCONFIG_PATH, kunitconfig_path)
|
||||||
|
|
||||||
self._kconfig = kunit_config.Kconfig()
|
self._kconfig = kunit_config.Kconfig()
|
||||||
self._kconfig.read_from_file(kunitconfig_path)
|
self._kconfig.read_from_file(kunitconfig_path)
|
||||||
@@ -199,8 +203,12 @@ class LinuxSourceTree(object):
|
|||||||
return False
|
return False
|
||||||
return self.validate_config(build_dir)
|
return self.validate_config(build_dir)
|
||||||
|
|
||||||
def run_kernel(self, args=[], build_dir='', timeout=None) -> Iterator[str]:
|
def run_kernel(self, args=None, build_dir='', filter_glob='', timeout=None) -> Iterator[str]:
|
||||||
|
if not args:
|
||||||
|
args = []
|
||||||
args.extend(['mem=1G', 'console=tty'])
|
args.extend(['mem=1G', 'console=tty'])
|
||||||
|
if filter_glob:
|
||||||
|
args.append('kunit.filter_glob='+filter_glob)
|
||||||
self._ops.linux_bin(args, timeout, build_dir)
|
self._ops.linux_bin(args, timeout, build_dir)
|
||||||
outfile = get_outfile_path(build_dir)
|
outfile = get_outfile_path(build_dir)
|
||||||
subprocess.call(['stty', 'sane'])
|
subprocess.call(['stty', 'sane'])
|
||||||
|
@@ -12,6 +12,7 @@ from unittest import mock
|
|||||||
import tempfile, shutil # Handling test_tmpdir
|
import tempfile, shutil # Handling test_tmpdir
|
||||||
|
|
||||||
import json
|
import json
|
||||||
|
import signal
|
||||||
import os
|
import os
|
||||||
|
|
||||||
import kunit_config
|
import kunit_config
|
||||||
@@ -21,16 +22,18 @@ import kunit_json
|
|||||||
import kunit
|
import kunit
|
||||||
|
|
||||||
test_tmpdir = ''
|
test_tmpdir = ''
|
||||||
|
abs_test_data_dir = ''
|
||||||
|
|
||||||
def setUpModule():
|
def setUpModule():
|
||||||
global test_tmpdir
|
global test_tmpdir, abs_test_data_dir
|
||||||
test_tmpdir = tempfile.mkdtemp()
|
test_tmpdir = tempfile.mkdtemp()
|
||||||
|
abs_test_data_dir = os.path.abspath(os.path.join(os.path.dirname(__file__), 'test_data'))
|
||||||
|
|
||||||
def tearDownModule():
|
def tearDownModule():
|
||||||
shutil.rmtree(test_tmpdir)
|
shutil.rmtree(test_tmpdir)
|
||||||
|
|
||||||
def get_absolute_path(path):
|
def test_data_path(path):
|
||||||
return os.path.join(os.path.dirname(__file__), path)
|
return os.path.join(abs_test_data_dir, path)
|
||||||
|
|
||||||
class KconfigTest(unittest.TestCase):
|
class KconfigTest(unittest.TestCase):
|
||||||
|
|
||||||
@@ -46,8 +49,7 @@ class KconfigTest(unittest.TestCase):
|
|||||||
|
|
||||||
def test_read_from_file(self):
|
def test_read_from_file(self):
|
||||||
kconfig = kunit_config.Kconfig()
|
kconfig = kunit_config.Kconfig()
|
||||||
kconfig_path = get_absolute_path(
|
kconfig_path = test_data_path('test_read_from_file.kconfig')
|
||||||
'test_data/test_read_from_file.kconfig')
|
|
||||||
|
|
||||||
kconfig.read_from_file(kconfig_path)
|
kconfig.read_from_file(kconfig_path)
|
||||||
|
|
||||||
@@ -98,21 +100,18 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
str(needle) + '" not found in "' + str(haystack) + '"!')
|
str(needle) + '" not found in "' + str(haystack) + '"!')
|
||||||
|
|
||||||
def test_output_isolated_correctly(self):
|
def test_output_isolated_correctly(self):
|
||||||
log_path = get_absolute_path(
|
log_path = test_data_path('test_output_isolated_correctly.log')
|
||||||
'test_data/test_output_isolated_correctly.log')
|
with open(log_path) as file:
|
||||||
file = open(log_path)
|
result = kunit_parser.isolate_kunit_output(file.readlines())
|
||||||
result = kunit_parser.isolate_kunit_output(file.readlines())
|
|
||||||
self.assertContains('TAP version 14', result)
|
self.assertContains('TAP version 14', result)
|
||||||
self.assertContains(' # Subtest: example', result)
|
self.assertContains(' # Subtest: example', result)
|
||||||
self.assertContains(' 1..2', result)
|
self.assertContains(' 1..2', result)
|
||||||
self.assertContains(' ok 1 - example_simple_test', result)
|
self.assertContains(' ok 1 - example_simple_test', result)
|
||||||
self.assertContains(' ok 2 - example_mock_test', result)
|
self.assertContains(' ok 2 - example_mock_test', result)
|
||||||
self.assertContains('ok 1 - example', result)
|
self.assertContains('ok 1 - example', result)
|
||||||
file.close()
|
|
||||||
|
|
||||||
def test_output_with_prefix_isolated_correctly(self):
|
def test_output_with_prefix_isolated_correctly(self):
|
||||||
log_path = get_absolute_path(
|
log_path = test_data_path('test_pound_sign.log')
|
||||||
'test_data/test_pound_sign.log')
|
|
||||||
with open(log_path) as file:
|
with open(log_path) as file:
|
||||||
result = kunit_parser.isolate_kunit_output(file.readlines())
|
result = kunit_parser.isolate_kunit_output(file.readlines())
|
||||||
self.assertContains('TAP version 14', result)
|
self.assertContains('TAP version 14', result)
|
||||||
@@ -141,61 +140,51 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertContains('ok 3 - string-stream-test', result)
|
self.assertContains('ok 3 - string-stream-test', result)
|
||||||
|
|
||||||
def test_parse_successful_test_log(self):
|
def test_parse_successful_test_log(self):
|
||||||
all_passed_log = get_absolute_path(
|
all_passed_log = test_data_path('test_is_test_passed-all_passed.log')
|
||||||
'test_data/test_is_test_passed-all_passed.log')
|
with open(all_passed_log) as file:
|
||||||
file = open(all_passed_log)
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
kunit_parser.TestStatus.SUCCESS,
|
kunit_parser.TestStatus.SUCCESS,
|
||||||
result.status)
|
result.status)
|
||||||
file.close()
|
|
||||||
|
|
||||||
def test_parse_failed_test_log(self):
|
def test_parse_failed_test_log(self):
|
||||||
failed_log = get_absolute_path(
|
failed_log = test_data_path('test_is_test_passed-failure.log')
|
||||||
'test_data/test_is_test_passed-failure.log')
|
with open(failed_log) as file:
|
||||||
file = open(failed_log)
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
kunit_parser.TestStatus.FAILURE,
|
kunit_parser.TestStatus.FAILURE,
|
||||||
result.status)
|
result.status)
|
||||||
file.close()
|
|
||||||
|
|
||||||
def test_no_tests(self):
|
def test_no_tests(self):
|
||||||
empty_log = get_absolute_path(
|
empty_log = test_data_path('test_is_test_passed-no_tests_run.log')
|
||||||
'test_data/test_is_test_passed-no_tests_run.log')
|
with open(empty_log) as file:
|
||||||
file = open(empty_log)
|
result = kunit_parser.parse_run_tests(
|
||||||
result = kunit_parser.parse_run_tests(
|
kunit_parser.isolate_kunit_output(file.readlines()))
|
||||||
kunit_parser.isolate_kunit_output(file.readlines()))
|
|
||||||
self.assertEqual(0, len(result.suites))
|
self.assertEqual(0, len(result.suites))
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
kunit_parser.TestStatus.NO_TESTS,
|
kunit_parser.TestStatus.NO_TESTS,
|
||||||
result.status)
|
result.status)
|
||||||
file.close()
|
|
||||||
|
|
||||||
def test_no_kunit_output(self):
|
def test_no_kunit_output(self):
|
||||||
crash_log = get_absolute_path(
|
crash_log = test_data_path('test_insufficient_memory.log')
|
||||||
'test_data/test_insufficient_memory.log')
|
|
||||||
file = open(crash_log)
|
|
||||||
print_mock = mock.patch('builtins.print').start()
|
print_mock = mock.patch('builtins.print').start()
|
||||||
result = kunit_parser.parse_run_tests(
|
with open(crash_log) as file:
|
||||||
kunit_parser.isolate_kunit_output(file.readlines()))
|
result = kunit_parser.parse_run_tests(
|
||||||
|
kunit_parser.isolate_kunit_output(file.readlines()))
|
||||||
print_mock.assert_any_call(StrContains('no tests run!'))
|
print_mock.assert_any_call(StrContains('no tests run!'))
|
||||||
print_mock.stop()
|
print_mock.stop()
|
||||||
file.close()
|
file.close()
|
||||||
|
|
||||||
def test_crashed_test(self):
|
def test_crashed_test(self):
|
||||||
crashed_log = get_absolute_path(
|
crashed_log = test_data_path('test_is_test_passed-crash.log')
|
||||||
'test_data/test_is_test_passed-crash.log')
|
with open(crashed_log) as file:
|
||||||
file = open(crashed_log)
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
kunit_parser.TestStatus.TEST_CRASHED,
|
kunit_parser.TestStatus.TEST_CRASHED,
|
||||||
result.status)
|
result.status)
|
||||||
file.close()
|
|
||||||
|
|
||||||
def test_ignores_prefix_printk_time(self):
|
def test_ignores_prefix_printk_time(self):
|
||||||
prefix_log = get_absolute_path(
|
prefix_log = test_data_path('test_config_printk_time.log')
|
||||||
'test_data/test_config_printk_time.log')
|
|
||||||
with open(prefix_log) as file:
|
with open(prefix_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -204,8 +193,7 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
def test_ignores_multiple_prefixes(self):
|
def test_ignores_multiple_prefixes(self):
|
||||||
prefix_log = get_absolute_path(
|
prefix_log = test_data_path('test_multiple_prefixes.log')
|
||||||
'test_data/test_multiple_prefixes.log')
|
|
||||||
with open(prefix_log) as file:
|
with open(prefix_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -214,8 +202,7 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
def test_prefix_mixed_kernel_output(self):
|
def test_prefix_mixed_kernel_output(self):
|
||||||
mixed_prefix_log = get_absolute_path(
|
mixed_prefix_log = test_data_path('test_interrupted_tap_output.log')
|
||||||
'test_data/test_interrupted_tap_output.log')
|
|
||||||
with open(mixed_prefix_log) as file:
|
with open(mixed_prefix_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -224,7 +211,7 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
def test_prefix_poundsign(self):
|
def test_prefix_poundsign(self):
|
||||||
pound_log = get_absolute_path('test_data/test_pound_sign.log')
|
pound_log = test_data_path('test_pound_sign.log')
|
||||||
with open(pound_log) as file:
|
with open(pound_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -233,7 +220,7 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
def test_kernel_panic_end(self):
|
def test_kernel_panic_end(self):
|
||||||
panic_log = get_absolute_path('test_data/test_kernel_panic_interrupt.log')
|
panic_log = test_data_path('test_kernel_panic_interrupt.log')
|
||||||
with open(panic_log) as file:
|
with open(panic_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -242,7 +229,7 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
def test_pound_no_prefix(self):
|
def test_pound_no_prefix(self):
|
||||||
pound_log = get_absolute_path('test_data/test_pound_no_prefix.log')
|
pound_log = test_data_path('test_pound_no_prefix.log')
|
||||||
with open(pound_log) as file:
|
with open(pound_log) as file:
|
||||||
result = kunit_parser.parse_run_tests(file.readlines())
|
result = kunit_parser.parse_run_tests(file.readlines())
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
@@ -250,10 +237,27 @@ class KUnitParserTest(unittest.TestCase):
|
|||||||
result.status)
|
result.status)
|
||||||
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
self.assertEqual('kunit-resource-test', result.suites[0].name)
|
||||||
|
|
||||||
|
class LinuxSourceTreeTest(unittest.TestCase):
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
mock.patch.object(signal, 'signal').start()
|
||||||
|
self.addCleanup(mock.patch.stopall)
|
||||||
|
|
||||||
|
def test_invalid_kunitconfig(self):
|
||||||
|
with self.assertRaisesRegex(kunit_kernel.ConfigError, 'nonexistent.* does not exist'):
|
||||||
|
kunit_kernel.LinuxSourceTree('', kunitconfig_path='/nonexistent_file')
|
||||||
|
|
||||||
|
def test_valid_kunitconfig(self):
|
||||||
|
with tempfile.NamedTemporaryFile('wt') as kunitconfig:
|
||||||
|
tree = kunit_kernel.LinuxSourceTree('', kunitconfig_path=kunitconfig.name)
|
||||||
|
|
||||||
|
# TODO: add more test cases.
|
||||||
|
|
||||||
|
|
||||||
class KUnitJsonTest(unittest.TestCase):
|
class KUnitJsonTest(unittest.TestCase):
|
||||||
|
|
||||||
def _json_for(self, log_file):
|
def _json_for(self, log_file):
|
||||||
with(open(get_absolute_path(log_file))) as file:
|
with open(test_data_path(log_file)) as file:
|
||||||
test_result = kunit_parser.parse_run_tests(file)
|
test_result = kunit_parser.parse_run_tests(file)
|
||||||
json_obj = kunit_json.get_json_result(
|
json_obj = kunit_json.get_json_result(
|
||||||
test_result=test_result,
|
test_result=test_result,
|
||||||
@@ -263,22 +267,19 @@ class KUnitJsonTest(unittest.TestCase):
|
|||||||
return json.loads(json_obj)
|
return json.loads(json_obj)
|
||||||
|
|
||||||
def test_failed_test_json(self):
|
def test_failed_test_json(self):
|
||||||
result = self._json_for(
|
result = self._json_for('test_is_test_passed-failure.log')
|
||||||
'test_data/test_is_test_passed-failure.log')
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
{'name': 'example_simple_test', 'status': 'FAIL'},
|
{'name': 'example_simple_test', 'status': 'FAIL'},
|
||||||
result["sub_groups"][1]["test_cases"][0])
|
result["sub_groups"][1]["test_cases"][0])
|
||||||
|
|
||||||
def test_crashed_test_json(self):
|
def test_crashed_test_json(self):
|
||||||
result = self._json_for(
|
result = self._json_for('test_is_test_passed-crash.log')
|
||||||
'test_data/test_is_test_passed-crash.log')
|
|
||||||
self.assertEqual(
|
self.assertEqual(
|
||||||
{'name': 'example_simple_test', 'status': 'ERROR'},
|
{'name': 'example_simple_test', 'status': 'ERROR'},
|
||||||
result["sub_groups"][1]["test_cases"][0])
|
result["sub_groups"][1]["test_cases"][0])
|
||||||
|
|
||||||
def test_no_tests_json(self):
|
def test_no_tests_json(self):
|
||||||
result = self._json_for(
|
result = self._json_for('test_is_test_passed-no_tests_run.log')
|
||||||
'test_data/test_is_test_passed-no_tests_run.log')
|
|
||||||
self.assertEqual(0, len(result['sub_groups']))
|
self.assertEqual(0, len(result['sub_groups']))
|
||||||
|
|
||||||
class StrContains(str):
|
class StrContains(str):
|
||||||
@@ -287,106 +288,104 @@ class StrContains(str):
|
|||||||
|
|
||||||
class KUnitMainTest(unittest.TestCase):
|
class KUnitMainTest(unittest.TestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
path = get_absolute_path('test_data/test_is_test_passed-all_passed.log')
|
path = test_data_path('test_is_test_passed-all_passed.log')
|
||||||
file = open(path)
|
with open(path) as file:
|
||||||
all_passed_log = file.readlines()
|
all_passed_log = file.readlines()
|
||||||
self.print_patch = mock.patch('builtins.print')
|
|
||||||
self.print_mock = self.print_patch.start()
|
self.print_mock = mock.patch('builtins.print').start()
|
||||||
|
self.addCleanup(mock.patch.stopall)
|
||||||
|
|
||||||
self.linux_source_mock = mock.Mock()
|
self.linux_source_mock = mock.Mock()
|
||||||
self.linux_source_mock.build_reconfig = mock.Mock(return_value=True)
|
self.linux_source_mock.build_reconfig = mock.Mock(return_value=True)
|
||||||
self.linux_source_mock.build_um_kernel = mock.Mock(return_value=True)
|
self.linux_source_mock.build_um_kernel = mock.Mock(return_value=True)
|
||||||
self.linux_source_mock.run_kernel = mock.Mock(return_value=all_passed_log)
|
self.linux_source_mock.run_kernel = mock.Mock(return_value=all_passed_log)
|
||||||
|
|
||||||
def tearDown(self):
|
|
||||||
self.print_patch.stop()
|
|
||||||
pass
|
|
||||||
|
|
||||||
def test_config_passes_args_pass(self):
|
def test_config_passes_args_pass(self):
|
||||||
kunit.main(['config', '--build_dir=.kunit'], self.linux_source_mock)
|
kunit.main(['config', '--build_dir=.kunit'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 0
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 0)
|
||||||
|
|
||||||
def test_build_passes_args_pass(self):
|
def test_build_passes_args_pass(self):
|
||||||
kunit.main(['build'], self.linux_source_mock)
|
kunit.main(['build'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 0
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 0)
|
||||||
self.linux_source_mock.build_um_kernel.assert_called_once_with(False, 8, '.kunit', None)
|
self.linux_source_mock.build_um_kernel.assert_called_once_with(False, 8, '.kunit', None)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 0
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 0)
|
||||||
|
|
||||||
def test_exec_passes_args_pass(self):
|
def test_exec_passes_args_pass(self):
|
||||||
kunit.main(['exec'], self.linux_source_mock)
|
kunit.main(['exec'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 0
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 0)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 1
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 1)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(build_dir='.kunit', timeout=300)
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
|
build_dir='.kunit', filter_glob='', timeout=300)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
def test_run_passes_args_pass(self):
|
def test_run_passes_args_pass(self):
|
||||||
kunit.main(['run'], self.linux_source_mock)
|
kunit.main(['run'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 1
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 1)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
build_dir='.kunit', timeout=300)
|
build_dir='.kunit', filter_glob='', timeout=300)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
def test_exec_passes_args_fail(self):
|
def test_exec_passes_args_fail(self):
|
||||||
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
||||||
with self.assertRaises(SystemExit) as e:
|
with self.assertRaises(SystemExit) as e:
|
||||||
kunit.main(['exec'], self.linux_source_mock)
|
kunit.main(['exec'], self.linux_source_mock)
|
||||||
assert type(e.exception) == SystemExit
|
self.assertEqual(e.exception.code, 1)
|
||||||
assert e.exception.code == 1
|
|
||||||
|
|
||||||
def test_run_passes_args_fail(self):
|
def test_run_passes_args_fail(self):
|
||||||
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
||||||
with self.assertRaises(SystemExit) as e:
|
with self.assertRaises(SystemExit) as e:
|
||||||
kunit.main(['run'], self.linux_source_mock)
|
kunit.main(['run'], self.linux_source_mock)
|
||||||
assert type(e.exception) == SystemExit
|
self.assertEqual(e.exception.code, 1)
|
||||||
assert e.exception.code == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 1)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 1
|
|
||||||
self.print_mock.assert_any_call(StrContains(' 0 tests run'))
|
self.print_mock.assert_any_call(StrContains(' 0 tests run'))
|
||||||
|
|
||||||
def test_exec_raw_output(self):
|
def test_exec_raw_output(self):
|
||||||
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
||||||
kunit.main(['exec', '--raw_output'], self.linux_source_mock)
|
kunit.main(['exec', '--raw_output'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 1
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 1)
|
||||||
for kall in self.print_mock.call_args_list:
|
for call in self.print_mock.call_args_list:
|
||||||
assert kall != mock.call(StrContains('Testing complete.'))
|
self.assertNotEqual(call, mock.call(StrContains('Testing complete.')))
|
||||||
assert kall != mock.call(StrContains(' 0 tests run'))
|
self.assertNotEqual(call, mock.call(StrContains(' 0 tests run')))
|
||||||
|
|
||||||
def test_run_raw_output(self):
|
def test_run_raw_output(self):
|
||||||
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
self.linux_source_mock.run_kernel = mock.Mock(return_value=[])
|
||||||
kunit.main(['run', '--raw_output'], self.linux_source_mock)
|
kunit.main(['run', '--raw_output'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
assert self.linux_source_mock.run_kernel.call_count == 1
|
self.assertEqual(self.linux_source_mock.run_kernel.call_count, 1)
|
||||||
for kall in self.print_mock.call_args_list:
|
for call in self.print_mock.call_args_list:
|
||||||
assert kall != mock.call(StrContains('Testing complete.'))
|
self.assertNotEqual(call, mock.call(StrContains('Testing complete.')))
|
||||||
assert kall != mock.call(StrContains(' 0 tests run'))
|
self.assertNotEqual(call, mock.call(StrContains(' 0 tests run')))
|
||||||
|
|
||||||
def test_exec_timeout(self):
|
def test_exec_timeout(self):
|
||||||
timeout = 3453
|
timeout = 3453
|
||||||
kunit.main(['exec', '--timeout', str(timeout)], self.linux_source_mock)
|
kunit.main(['exec', '--timeout', str(timeout)], self.linux_source_mock)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(build_dir='.kunit', timeout=timeout)
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
|
build_dir='.kunit', filter_glob='', timeout=timeout)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
def test_run_timeout(self):
|
def test_run_timeout(self):
|
||||||
timeout = 3453
|
timeout = 3453
|
||||||
kunit.main(['run', '--timeout', str(timeout)], self.linux_source_mock)
|
kunit.main(['run', '--timeout', str(timeout)], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
build_dir='.kunit', timeout=timeout)
|
build_dir='.kunit', filter_glob='', timeout=timeout)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
def test_run_builddir(self):
|
def test_run_builddir(self):
|
||||||
build_dir = '.kunit'
|
build_dir = '.kunit'
|
||||||
kunit.main(['run', '--build_dir=.kunit'], self.linux_source_mock)
|
kunit.main(['run', '--build_dir=.kunit'], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
build_dir=build_dir, timeout=300)
|
build_dir=build_dir, filter_glob='', timeout=300)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
def test_config_builddir(self):
|
def test_config_builddir(self):
|
||||||
build_dir = '.kunit'
|
build_dir = '.kunit'
|
||||||
kunit.main(['config', '--build_dir', build_dir], self.linux_source_mock)
|
kunit.main(['config', '--build_dir', build_dir], self.linux_source_mock)
|
||||||
assert self.linux_source_mock.build_reconfig.call_count == 1
|
self.assertEqual(self.linux_source_mock.build_reconfig.call_count, 1)
|
||||||
|
|
||||||
def test_build_builddir(self):
|
def test_build_builddir(self):
|
||||||
build_dir = '.kunit'
|
build_dir = '.kunit'
|
||||||
@@ -396,8 +395,23 @@ class KUnitMainTest(unittest.TestCase):
|
|||||||
def test_exec_builddir(self):
|
def test_exec_builddir(self):
|
||||||
build_dir = '.kunit'
|
build_dir = '.kunit'
|
||||||
kunit.main(['exec', '--build_dir', build_dir], self.linux_source_mock)
|
kunit.main(['exec', '--build_dir', build_dir], self.linux_source_mock)
|
||||||
self.linux_source_mock.run_kernel.assert_called_once_with(build_dir=build_dir, timeout=300)
|
self.linux_source_mock.run_kernel.assert_called_once_with(
|
||||||
|
build_dir=build_dir, filter_glob='', timeout=300)
|
||||||
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
self.print_mock.assert_any_call(StrContains('Testing complete.'))
|
||||||
|
|
||||||
|
@mock.patch.object(kunit_kernel, 'LinuxSourceTree')
|
||||||
|
def test_run_kunitconfig(self, mock_linux_init):
|
||||||
|
mock_linux_init.return_value = self.linux_source_mock
|
||||||
|
kunit.main(['run', '--kunitconfig=mykunitconfig'])
|
||||||
|
# Just verify that we parsed and initialized it correctly here.
|
||||||
|
mock_linux_init.assert_called_once_with('.kunit', kunitconfig_path='mykunitconfig')
|
||||||
|
|
||||||
|
@mock.patch.object(kunit_kernel, 'LinuxSourceTree')
|
||||||
|
def test_config_kunitconfig(self, mock_linux_init):
|
||||||
|
mock_linux_init.return_value = self.linux_source_mock
|
||||||
|
kunit.main(['config', '--kunitconfig=mykunitconfig'])
|
||||||
|
# Just verify that we parsed and initialized it correctly here.
|
||||||
|
mock_linux_init.assert_called_once_with('.kunit', kunitconfig_path='mykunitconfig')
|
||||||
|
|
||||||
if __name__ == '__main__':
|
if __name__ == '__main__':
|
||||||
unittest.main()
|
unittest.main()
|
||||||
|
Reference in New Issue
Block a user