| #! /usr/bin/env python3 |
| # |
| # BitBake Toaster Implementation |
| # |
| # Copyright (C) 2016 Intel Corporation |
| # |
| # SPDX-License-Identifier: GPL-2.0-only |
| # |
| # Tests were part of openembedded-core oe selftest Authored by: Lucian Musat |
| # Ionut Chisanovici, Paul Eggleton and Cristian Iorga |
| |
| import os |
| import pytest |
| |
| from django.db.models import Q |
| |
| from orm.models import Target_Image_File, Target_Installed_Package, Task |
| from orm.models import Package_Dependency, Recipe_Dependency, Build |
| from orm.models import Task_Dependency, Package, Target, Recipe |
| from orm.models import CustomImagePackage |
| |
| from tests.builds.buildtest import BuildTest |
| |
| @pytest.mark.order(4) |
| @pytest.mark.django_db(True) |
| class BuildCoreImageMinimal(BuildTest): |
| """Build core-image-minimal and test the results""" |
| |
| def setUp(self): |
| self.completed_build = self.target_already_built("core-image-minimal") |
| |
| # Check if build name is unique - tc_id=795 |
| def test_Build_Unique_Name(self): |
| all_builds = Build.objects.all().count() |
| distinct_builds = Build.objects.values('id').distinct().count() |
| self.assertEqual(distinct_builds, |
| all_builds, |
| msg='Build name is not unique') |
| |
| # Check if build cooker log path is unique - tc_id=819 |
| def test_Build_Unique_Cooker_Log_Path(self): |
| distinct_path = Build.objects.values( |
| 'cooker_log_path').distinct().count() |
| total_builds = Build.objects.values('id').count() |
| self.assertEqual(distinct_path, |
| total_builds, |
| msg='Build cooker log path is not unique') |
| |
| # Check task order sequence for one build - tc=825 |
| def test_Task_Order_Sequence(self): |
| cnt_err = [] |
| tasks = Task.objects.filter( |
| Q(build=self.completed_build), |
| ~Q(order=None), |
| ~Q(task_name__contains='_setscene') |
| ).values('id', 'order').order_by("order") |
| |
| cnt_tasks = 0 |
| for task in tasks: |
| cnt_tasks += 1 |
| if (task['order'] != cnt_tasks): |
| cnt_err.append(task['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if disk_io matches the difference between EndTimeIO and |
| # StartTimeIO in build stats - tc=828 |
| # def test_Task_Disk_IO_TC828(self): |
| |
| # Check if outcome = 2 (SSTATE) then sstate_result must be 3 (RESTORED) - |
| # tc=832 |
| def test_Task_If_Outcome_2_Sstate_Result_Must_Be_3(self): |
| tasks = Task.objects.filter(outcome=2).values('id', 'sstate_result') |
| cnt_err = [] |
| for task in tasks: |
| if (task['sstate_result'] != 3): |
| cnt_err.append(task['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if outcome = 1 (COVERED) or 3 (EXISTING) then sstate_result must |
| # be 0 (SSTATE_NA) - tc=833 |
| def test_Task_If_Outcome_1_3_Sstate_Result_Must_Be_0(self): |
| tasks = Task.objects.filter( |
| outcome__in=(Task.OUTCOME_COVERED, |
| Task.OUTCOME_PREBUILT)).values('id', |
| 'task_name', |
| 'sstate_result') |
| cnt_err = [] |
| for task in tasks: |
| if (task['sstate_result'] != Task.SSTATE_NA and |
| task['sstate_result'] != Task.SSTATE_MISS): |
| cnt_err.append({'id': task['id'], |
| 'name': task['task_name'], |
| 'sstate_result': task['sstate_result']}) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if outcome is 0 (SUCCESS) or 4 (FAILED) then sstate_result must be |
| # 0 (NA), 1 (MISS) or 2 (FAILED) - tc=834 |
| def test_Task_If_Outcome_0_4_Sstate_Result_Must_Be_0_1_2(self): |
| tasks = Task.objects.filter( |
| outcome__in=(0, 4)).values('id', 'sstate_result') |
| cnt_err = [] |
| |
| for task in tasks: |
| if (task['sstate_result'] not in [0, 1, 2]): |
| cnt_err.append(task['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if task_executed = TRUE (1), script_type must be 0 (CODING_NA), 2 |
| # (CODING_PYTHON), 3 (CODING_SHELL) - tc=891 |
| def test_Task_If_Task_Executed_True_Script_Type_0_2_3(self): |
| tasks = Task.objects.filter( |
| task_executed=1).values('id', 'script_type') |
| cnt_err = [] |
| |
| for task in tasks: |
| if (task['script_type'] not in [0, 2, 3]): |
| cnt_err.append(task['id']) |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if task_executed = TRUE (1), outcome must be 0 (SUCCESS) or 4 |
| # (FAILED) - tc=836 |
| def test_Task_If_Task_Executed_True_Outcome_0_4(self): |
| tasks = Task.objects.filter(task_executed=1).values('id', 'outcome') |
| cnt_err = [] |
| |
| for task in tasks: |
| if (task['outcome'] not in [0, 4]): |
| cnt_err.append(task['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if task_executed = FALSE (0), script_type must be 0 - tc=890 |
| def test_Task_If_Task_Executed_False_Script_Type_0(self): |
| tasks = Task.objects.filter( |
| task_executed=0).values('id', 'script_type') |
| cnt_err = [] |
| |
| for task in tasks: |
| if (task['script_type'] != 0): |
| cnt_err.append(task['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Check if task_executed = FALSE (0) and build outcome = SUCCEEDED (0), |
| # task outcome must be 1 (COVERED), 2 (CACHED), 3 (PREBUILT), 5 (EMPTY) - |
| # tc=837 |
| def test_Task_If_Task_Executed_False_Outcome_1_2_3_5(self): |
| builds = Build.objects.filter(outcome=0).values('id') |
| cnt_err = [] |
| for build in builds: |
| tasks = Task.objects.filter( |
| build=build['id'], task_executed=0).values('id', 'outcome') |
| for task in tasks: |
| if (task['outcome'] not in [1, 2, 3, 5]): |
| cnt_err.append(task['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task id: %s' % cnt_err) |
| |
| # Key verification - tc=888 |
| def test_Target_Installed_Package(self): |
| rows = Target_Installed_Package.objects.values('id', |
| 'target_id', |
| 'package_id') |
| cnt_err = [] |
| |
| for row in rows: |
| target = Target.objects.filter(id=row['target_id']).values('id') |
| package = Package.objects.filter(id=row['package_id']).values('id') |
| if (not target or not package): |
| cnt_err.append(row['id']) |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for target installed package id: %s' % |
| cnt_err) |
| |
| # Key verification - tc=889 |
| def test_Task_Dependency(self): |
| rows = Task_Dependency.objects.values('id', |
| 'task_id', |
| 'depends_on_id') |
| cnt_err = [] |
| for row in rows: |
| task_id = Task.objects.filter(id=row['task_id']).values('id') |
| depends_on_id = Task.objects.filter( |
| id=row['depends_on_id']).values('id') |
| if (not task_id or not depends_on_id): |
| cnt_err.append(row['id']) |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg='Errors for task dependency id: %s' % cnt_err) |
| |
| # Check if build target file_name is populated only if is_image=true AND |
| # orm_build.outcome=0 then if the file exists and its size matches |
| # the file_size value. Need to add the tc in the test run |
| def test_Target_File_Name_Populated(self): |
| cnt_err = [] |
| builds = Build.objects.filter(outcome=0).values('id') |
| for build in builds: |
| targets = Target.objects.filter( |
| build_id=build['id'], is_image=1).values('id') |
| for target in targets: |
| target_files = Target_Image_File.objects.filter( |
| target_id=target['id']).values('id', |
| 'file_name', |
| 'file_size') |
| for file_info in target_files: |
| target_id = file_info['id'] |
| target_file_name = file_info['file_name'] |
| target_file_size = file_info['file_size'] |
| if (not target_file_name or not target_file_size): |
| cnt_err.append(target_id) |
| else: |
| if (not os.path.exists(target_file_name)): |
| cnt_err.append(target_id) |
| else: |
| if (os.path.getsize(target_file_name) != |
| target_file_size): |
| cnt_err.append(target_id) |
| self.assertEqual(len(cnt_err), 0, |
| msg='Errors for target image file id: %s' % |
| cnt_err) |
| |
| # Key verification - tc=884 |
| def test_Package_Dependency(self): |
| cnt_err = [] |
| deps = Package_Dependency.objects.values( |
| 'id', 'package_id', 'depends_on_id') |
| for dep in deps: |
| if (dep['package_id'] == dep['depends_on_id']): |
| cnt_err.append(dep['id']) |
| self.assertEqual(len(cnt_err), 0, |
| msg='Errors for package dependency id: %s' % cnt_err) |
| |
| # Recipe key verification, recipe name does not depends on a recipe having |
| # the same name - tc=883 |
| def test_Recipe_Dependency(self): |
| deps = Recipe_Dependency.objects.values( |
| 'id', 'recipe_id', 'depends_on_id') |
| cnt_err = [] |
| for dep in deps: |
| if (not dep['recipe_id'] or not dep['depends_on_id']): |
| cnt_err.append(dep['id']) |
| else: |
| name = Recipe.objects.filter( |
| id=dep['recipe_id']).values('name') |
| dep_name = Recipe.objects.filter( |
| id=dep['depends_on_id']).values('name') |
| if (name == dep_name): |
| cnt_err.append(dep['id']) |
| self.assertEqual(len(cnt_err), 0, |
| msg='Errors for recipe dependency id: %s' % cnt_err) |
| |
| # Check if package name does not start with a number (0-9) - tc=846 |
| def test_Package_Name_For_Number(self): |
| packages = Package.objects.filter(~Q(size=-1)).values('id', 'name') |
| cnt_err = [] |
| for package in packages: |
| if (package['name'][0].isdigit() is True): |
| cnt_err.append(package['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) |
| |
| # Check if package version starts with a number (0-9) - tc=847 |
| def test_Package_Version_Starts_With_Number(self): |
| packages = Package.objects.filter( |
| ~Q(size=-1)).values('id', 'version') |
| cnt_err = [] |
| for package in packages: |
| if (package['version'][0].isdigit() is False): |
| cnt_err.append(package['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) |
| |
| # Check if package revision starts with 'r' - tc=848 |
| def test_Package_Revision_Starts_With_r(self): |
| packages = Package.objects.filter( |
| ~Q(size=-1)).values('id', 'revision') |
| cnt_err = [] |
| for package in packages: |
| if (package['revision'][0].startswith("r") is False): |
| cnt_err.append(package['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) |
| |
| # Check the validity of the package build_id |
| # TC must be added in test run |
| def test_Package_Build_Id(self): |
| packages = Package.objects.filter( |
| ~Q(size=-1)).values('id', 'build_id') |
| cnt_err = [] |
| for package in packages: |
| build_id = Build.objects.filter( |
| id=package['build_id']).values('id') |
| if (not build_id): |
| # They have no build_id but if they are |
| # CustomImagePackage that's expected |
| try: |
| CustomImagePackage.objects.get(pk=package['id']) |
| except CustomImagePackage.DoesNotExist: |
| cnt_err.append(package['id']) |
| |
| self.assertEqual(len(cnt_err), |
| 0, |
| msg="Errors for package id: %s they have no build" |
| "associated with them" % cnt_err) |
| |
| # Check the validity of package recipe_id |
| # TC must be added in test run |
| def test_Package_Recipe_Id(self): |
| packages = Package.objects.filter( |
| ~Q(size=-1)).values('id', 'recipe_id') |
| cnt_err = [] |
| for package in packages: |
| recipe_id = Recipe.objects.filter( |
| id=package['recipe_id']).values('id') |
| if (not recipe_id): |
| cnt_err.append(package['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) |
| |
| # Check if package installed_size field is not null |
| # TC must be aded in test run |
| def test_Package_Installed_Size_Not_NULL(self): |
| packages = Package.objects.filter( |
| installed_size__isnull=True).values('id') |
| cnt_err = [] |
| for package in packages: |
| cnt_err.append(package['id']) |
| self.assertEqual( |
| len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) |
| |
| def test_custom_packages_generated(self): |
| """Test if there is a corresponding generated CustomImagePackage""" |
| """ for each of the packages generated""" |
| missing_packages = [] |
| |
| for package in Package.objects.all(): |
| try: |
| CustomImagePackage.objects.get(name=package.name) |
| except CustomImagePackage.DoesNotExist: |
| missing_packages.append(package.name) |
| |
| self.assertEqual(len(missing_packages), 0, |
| "Some package were created from the build but their" |
| " corresponding CustomImagePackage was not found") |