aboutsummaryrefslogtreecommitdiffstats
path: root/lib/tests/builds/test_core_image_min.py
diff options
context:
space:
mode:
Diffstat (limited to 'lib/tests/builds/test_core_image_min.py')
-rw-r--r--lib/tests/builds/test_core_image_min.py386
1 files changed, 0 insertions, 386 deletions
diff --git a/lib/tests/builds/test_core_image_min.py b/lib/tests/builds/test_core_image_min.py
deleted file mode 100644
index 586f4a8f..00000000
--- a/lib/tests/builds/test_core_image_min.py
+++ /dev/null
@@ -1,386 +0,0 @@
-#! /usr/bin/env python
-# ex:ts=4:sw=4:sts=4:et
-# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
-#
-# BitBake Toaster Implementation
-#
-# Copyright (C) 2016 Intel Corporation
-#
-# This program is free software; you can redistribute it and/or modify
-# it under the terms of the GNU General Public License version 2 as
-# published by the Free Software Foundation.
-#
-# This program is distributed in the hope that it will be useful,
-# but WITHOUT ANY WARRANTY; without even the implied warranty of
-# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
-# GNU General Public License for more details.
-#
-# You should have received a copy of the GNU General Public License along
-# with this program; if not, write to the Free Software Foundation, Inc.,
-# 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
-
-# Tests were part of openembedded-core oe selftest Authored by: Lucian Musat
-# Ionut Chisanovici, Paul Eggleton and Cristian Iorga
-
-import os
-
-from django.db.models import Q
-
-from orm.models import Target_Image_File, Target_Installed_Package, Task
-from orm.models import Package_Dependency, Recipe_Dependency, Build
-from orm.models import Task_Dependency, Package, Target, Recipe
-from orm.models import CustomImagePackage
-
-from tests.builds.buildtest import BuildTest
-
-
-class BuildCoreImageMinimal(BuildTest):
- """Build core-image-minimal and test the results"""
-
- def setUp(self):
- self.completed_build = self.build("core-image-minimal")
-
- # Check if build name is unique - tc_id=795
- def test_Build_Unique_Name(self):
- all_builds = Build.objects.all().count()
- distinct_builds = Build.objects.values('id').distinct().count()
- self.assertEqual(distinct_builds,
- all_builds,
- msg='Build name is not unique')
-
- # Check if build cooker log path is unique - tc_id=819
- def test_Build_Unique_Cooker_Log_Path(self):
- distinct_path = Build.objects.values(
- 'cooker_log_path').distinct().count()
- total_builds = Build.objects.values('id').count()
- self.assertEqual(distinct_path,
- total_builds,
- msg='Build cooker log path is not unique')
-
- # Check if task order is unique for one build - tc=824
- def test_Task_Unique_Order(self):
- total_task_order = Task.objects.filter(
- build=self.built).values('order').count()
- distinct_task_order = Task.objects.filter(
- build=self.completed_build).values('order').distinct().count()
-
- self.assertEqual(total_task_order,
- distinct_task_order,
- msg='Errors task order is not unique')
-
- # Check task order sequence for one build - tc=825
- def test_Task_Order_Sequence(self):
- cnt_err = []
- tasks = Task.objects.filter(
- Q(build=self.completed_build),
- ~Q(order=None),
- ~Q(task_name__contains='_setscene')
- ).values('id', 'order').order_by("order")
-
- cnt_tasks = 0
- for task in tasks:
- cnt_tasks += 1
- if (task['order'] != cnt_tasks):
- cnt_err.append(task['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err)
-
- # Check if disk_io matches the difference between EndTimeIO and
- # StartTimeIO in build stats - tc=828
- # def test_Task_Disk_IO_TC828(self):
-
- # Check if outcome = 2 (SSTATE) then sstate_result must be 3 (RESTORED) -
- # tc=832
- def test_Task_If_Outcome_2_Sstate_Result_Must_Be_3(self):
- tasks = Task.objects.filter(outcome=2).values('id', 'sstate_result')
- cnt_err = []
- for task in tasks:
- if (task['sstate_result'] != 3):
- cnt_err.append(task['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if outcome = 1 (COVERED) or 3 (EXISTING) then sstate_result must
- # be 0 (SSTATE_NA) - tc=833
- def test_Task_If_Outcome_1_3_Sstate_Result_Must_Be_0(self):
- tasks = Task.objects.filter(
- outcome__in=(Task.OUTCOME_COVERED,
- Task.OUTCOME_PREBUILT)).values('id',
- 'task_name',
- 'sstate_result')
- cnt_err = []
-
- for task in tasks:
- if (task['sstate_result'] != Task.SSTATE_NA and
- task['sstate_result'] != Task.SSTATE_MISS):
- cnt_err.append({'id': task['id'],
- 'name': task['task_name'],
- 'sstate_result': task['sstate_result']})
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if outcome is 0 (SUCCESS) or 4 (FAILED) then sstate_result must be
- # 0 (NA), 1 (MISS) or 2 (FAILED) - tc=834
- def test_Task_If_Outcome_0_4_Sstate_Result_Must_Be_0_1_2(self):
- tasks = Task.objects.filter(
- outcome__in=(0, 4)).values('id', 'sstate_result')
- cnt_err = []
-
- for task in tasks:
- if (task['sstate_result'] not in [0, 1, 2]):
- cnt_err.append(task['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if task_executed = TRUE (1), script_type must be 0 (CODING_NA), 2
- # (CODING_PYTHON), 3 (CODING_SHELL) - tc=891
- def test_Task_If_Task_Executed_True_Script_Type_0_2_3(self):
- tasks = Task.objects.filter(
- task_executed=1).values('id', 'script_type')
- cnt_err = []
-
- for task in tasks:
- if (task['script_type'] not in [0, 2, 3]):
- cnt_err.append(task['id'])
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if task_executed = TRUE (1), outcome must be 0 (SUCCESS) or 4
- # (FAILED) - tc=836
- def test_Task_If_Task_Executed_True_Outcome_0_4(self):
- tasks = Task.objects.filter(task_executed=1).values('id', 'outcome')
- cnt_err = []
-
- for task in tasks:
- if (task['outcome'] not in [0, 4]):
- cnt_err.append(task['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if task_executed = FALSE (0), script_type must be 0 - tc=890
- def test_Task_If_Task_Executed_False_Script_Type_0(self):
- tasks = Task.objects.filter(
- task_executed=0).values('id', 'script_type')
- cnt_err = []
-
- for task in tasks:
- if (task['script_type'] != 0):
- cnt_err.append(task['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Check if task_executed = FALSE (0) and build outcome = SUCCEEDED (0),
- # task outcome must be 1 (COVERED), 2 (CACHED), 3 (PREBUILT), 5 (EMPTY) -
- # tc=837
- def test_Task_If_Task_Executed_False_Outcome_1_2_3_5(self):
- builds = Build.objects.filter(outcome=0).values('id')
- cnt_err = []
- for build in builds:
- tasks = Task.objects.filter(
- build=build['id'], task_executed=0).values('id', 'outcome')
- for task in tasks:
- if (task['outcome'] not in [1, 2, 3, 5]):
- cnt_err.append(task['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task id: %s' % cnt_err)
-
- # Key verification - tc=888
- def test_Target_Installed_Package(self):
- rows = Target_Installed_Package.objects.values('id',
- 'target_id',
- 'package_id')
- cnt_err = []
-
- for row in rows:
- target = Target.objects.filter(id=row['target_id']).values('id')
- package = Package.objects.filter(id=row['package_id']).values('id')
- if (not target or not package):
- cnt_err.append(row['id'])
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for target installed package id: %s' %
- cnt_err)
-
- # Key verification - tc=889
- def test_Task_Dependency(self):
- rows = Task_Dependency.objects.values('id',
- 'task_id',
- 'depends_on_id')
- cnt_err = []
- for row in rows:
- task_id = Task.objects.filter(id=row['task_id']).values('id')
- depends_on_id = Task.objects.filter(
- id=row['depends_on_id']).values('id')
- if (not task_id or not depends_on_id):
- cnt_err.append(row['id'])
- self.assertEqual(len(cnt_err),
- 0,
- msg='Errors for task dependency id: %s' % cnt_err)
-
- # Check if build target file_name is populated only if is_image=true AND
- # orm_build.outcome=0 then if the file exists and its size matches
- # the file_size value. Need to add the tc in the test run
- def test_Target_File_Name_Populated(self):
- builds = Build.objects.filter(outcome=0).values('id')
- for build in builds:
- targets = Target.objects.filter(
- build_id=build['id'], is_image=1).values('id')
- for target in targets:
- target_files = Target_Image_File.objects.filter(
- target_id=target['id']).values('id',
- 'file_name',
- 'file_size')
- cnt_err = []
- for file_info in target_files:
- target_id = file_info['id']
- target_file_name = file_info['file_name']
- target_file_size = file_info['file_size']
- if (not target_file_name or not target_file_size):
- cnt_err.append(target_id)
- else:
- if (not os.path.exists(target_file_name)):
- cnt_err.append(target_id)
- else:
- if (os.path.getsize(target_file_name) !=
- target_file_size):
- cnt_err.append(target_id)
- self.assertEqual(len(cnt_err), 0,
- msg='Errors for target image file id: %s' %
- cnt_err)
-
- # Key verification - tc=884
- def test_Package_Dependency(self):
- cnt_err = []
- deps = Package_Dependency.objects.values(
- 'id', 'package_id', 'depends_on_id')
- for dep in deps:
- if (dep['package_id'] == dep['depends_on_id']):
- cnt_err.append(dep['id'])
- self.assertEqual(len(cnt_err), 0,
- msg='Errors for package dependency id: %s' % cnt_err)
-
- # Recipe key verification, recipe name does not depends on a recipe having
- # the same name - tc=883
- def test_Recipe_Dependency(self):
- deps = Recipe_Dependency.objects.values(
- 'id', 'recipe_id', 'depends_on_id')
- cnt_err = []
- for dep in deps:
- if (not dep['recipe_id'] or not dep['depends_on_id']):
- cnt_err.append(dep['id'])
- else:
- name = Recipe.objects.filter(
- id=dep['recipe_id']).values('name')
- dep_name = Recipe.objects.filter(
- id=dep['depends_on_id']).values('name')
- if (name == dep_name):
- cnt_err.append(dep['id'])
- self.assertEqual(len(cnt_err), 0,
- msg='Errors for recipe dependency id: %s' % cnt_err)
-
- # Check if package name does not start with a number (0-9) - tc=846
- def test_Package_Name_For_Number(self):
- packages = Package.objects.filter(~Q(size=-1)).values('id', 'name')
- cnt_err = []
- for package in packages:
- if (package['name'][0].isdigit() is True):
- cnt_err.append(package['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
-
- # Check if package version starts with a number (0-9) - tc=847
- def test_Package_Version_Starts_With_Number(self):
- packages = Package.objects.filter(
- ~Q(size=-1)).values('id', 'version')
- cnt_err = []
- for package in packages:
- if (package['version'][0].isdigit() is False):
- cnt_err.append(package['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
-
- # Check if package revision starts with 'r' - tc=848
- def test_Package_Revision_Starts_With_r(self):
- packages = Package.objects.filter(
- ~Q(size=-1)).values('id', 'revision')
- cnt_err = []
- for package in packages:
- if (package['revision'][0].startswith("r") is False):
- cnt_err.append(package['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
-
- # Check the validity of the package build_id
- # TC must be added in test run
- def test_Package_Build_Id(self):
- packages = Package.objects.filter(
- ~Q(size=-1)).values('id', 'build_id')
- cnt_err = []
- for package in packages:
- build_id = Build.objects.filter(
- id=package['build_id']).values('id')
- if (not build_id):
- # They have no build_id but if they are
- # CustomImagePackage that's expected
- try:
- CustomImagePackage.objects.get(pk=package['id'])
- except CustomImagePackage.DoesNotExist:
- cnt_err.append(package['id'])
-
- self.assertEqual(len(cnt_err),
- 0,
- msg="Errors for package id: %s they have no build"
- "associated with them" % cnt_err)
-
- # Check the validity of package recipe_id
- # TC must be added in test run
- def test_Package_Recipe_Id(self):
- packages = Package.objects.filter(
- ~Q(size=-1)).values('id', 'recipe_id')
- cnt_err = []
- for package in packages:
- recipe_id = Recipe.objects.filter(
- id=package['recipe_id']).values('id')
- if (not recipe_id):
- cnt_err.append(package['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
-
- # Check if package installed_size field is not null
- # TC must be aded in test run
- def test_Package_Installed_Size_Not_NULL(self):
- packages = Package.objects.filter(
- installed_size__isnull=True).values('id')
- cnt_err = []
- for package in packages:
- cnt_err.append(package['id'])
- self.assertEqual(
- len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err)
-
- def test_custom_packages_generated(self):
- """Test if there is a corresponding generated CustomImagePackage"""
- """ for each of the packages generated"""
- missing_packages = []
-
- for package in Package.objects.all():
- try:
- CustomImagePackage.objects.get(name=package.name)
- except CustomImagePackage.DoesNotExist:
- missing_packages.append(package.name)
-
- self.assertEqual(len(missing_packages), 0,
- "Some package were created from the build but their"
- " corresponding CustomImagePackage was not found")