32
32
import ctypes
33
33
import functools
34
34
from colorama import Fore , Back , Style
35
- from prettytable import PrettyTable
35
+ from prettytable import PrettyTable , HEADER
36
36
from copy import copy , deepcopy
37
37
38
38
from time import sleep , time
@@ -765,7 +765,7 @@ def generate_test_summary_by_target(self, test_summary, shuffle_seed=None):
765
765
result_dict [test [TEST_INDEX ]][test [TOOLCHAIN_INDEX ]] = test [RESULT_INDEX ]
766
766
767
767
pt_cols = ["Target" , "Test ID" , "Test Description" ] + unique_target_toolchains
768
- pt = PrettyTable (pt_cols )
768
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
769
769
for col in pt_cols :
770
770
pt .align [col ] = "l"
771
771
pt .padding_width = 1 # One space between column edges and contents (default)
@@ -793,7 +793,7 @@ def generate_test_summary(self, test_summary, shuffle_seed=None):
793
793
result = "Test summary:\n "
794
794
# Pretty table package is used to print results
795
795
pt = PrettyTable (["Result" , "Target" , "Toolchain" , "Test ID" , "Test Description" ,
796
- "Elapsed Time (sec)" , "Timeout (sec)" , "Loops" ])
796
+ "Elapsed Time (sec)" , "Timeout (sec)" , "Loops" ], junction_char = "|" , hrules = HEADER )
797
797
pt .align ["Result" ] = "l" # Left align
798
798
pt .align ["Target" ] = "l" # Left align
799
799
pt .align ["Toolchain" ] = "l" # Left align
@@ -1327,7 +1327,7 @@ def print_muts_configuration_from_json(json_data, join_delim=", ", platform_filt
1327
1327
1328
1328
# Prepare pretty table object to display all MUTs
1329
1329
pt_cols = ["index" ] + muts_info_cols
1330
- pt = PrettyTable (pt_cols )
1330
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
1331
1331
for col in pt_cols :
1332
1332
pt .align [col ] = "l"
1333
1333
@@ -1365,7 +1365,7 @@ def print_test_configuration_from_json(json_data, join_delim=", "):
1365
1365
1366
1366
# Prepare pretty table object to display test specification
1367
1367
pt_cols = ["mcu" ] + sorted (toolchains_info_cols )
1368
- pt = PrettyTable (pt_cols )
1368
+ pt = PrettyTable (pt_cols , junction_char = "|" , hrules = HEADER )
1369
1369
for col in pt_cols :
1370
1370
pt .align [col ] = "l"
1371
1371
@@ -1454,7 +1454,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1454
1454
'duration' ] if cols is None else cols
1455
1455
1456
1456
# All tests status table print
1457
- pt = PrettyTable (test_properties )
1457
+ pt = PrettyTable (test_properties , junction_char = "|" , hrules = HEADER )
1458
1458
for col in test_properties :
1459
1459
pt .align [col ] = "l"
1460
1460
pt .align ['duration' ] = "r"
@@ -1494,7 +1494,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1494
1494
if result_summary and not platform_filter :
1495
1495
# Automation result summary
1496
1496
test_id_cols = ['automated' , 'all' , 'percent [%]' , 'progress' ]
1497
- pt = PrettyTable (test_id_cols )
1497
+ pt = PrettyTable (test_id_cols , junction_char = "|" , hrules = HEADER )
1498
1498
pt .align ['automated' ] = "r"
1499
1499
pt .align ['all' ] = "r"
1500
1500
pt .align ['percent [%]' ] = "r"
@@ -1508,7 +1508,7 @@ def get_avail_tests_summary_table(cols=None, result_summary=True, join_delim=','
1508
1508
1509
1509
# Test automation coverage table print
1510
1510
test_id_cols = ['id' , 'automated' , 'all' , 'percent [%]' , 'progress' ]
1511
- pt = PrettyTable (test_id_cols )
1511
+ pt = PrettyTable (test_id_cols , junction_char = "|" , hrules = HEADER )
1512
1512
pt .align ['id' ] = "l"
1513
1513
pt .align ['automated' ] = "r"
1514
1514
pt .align ['all' ] = "r"
0 commit comments