@@ -57,13 +57,13 @@ def test_command_help_lists_commands(self):
57
57
58
58
def test_run_benchmarks_by_name_or_ordinal (self ):
59
59
benchmarks = ['AngryPhonebook' , '42' ]
60
- self .assertEquals (
60
+ self .assertEqual (
61
61
parse_args (['run' ] + benchmarks ).benchmarks , benchmarks )
62
62
63
63
def test_run_benchmarks_matching_pattern (self ):
64
64
regexes = ['Prefix' , '.*Suffix.*' ]
65
65
filters = ['-f' , regexes [0 ], '-f' , regexes [1 ]]
66
- self .assertEquals (parse_args (['run' ] + filters ).filters , regexes )
66
+ self .assertEqual (parse_args (['run' ] + filters ).filters , regexes )
67
67
68
68
def test_run_benchmarks_and_filters_are_exclusive (self ):
69
69
with captured_output () as (_ , err ):
@@ -76,17 +76,17 @@ def test_run_benchmarks_and_filters_are_exclusive(self):
76
76
77
77
def test_tests_location (self ):
78
78
here = os .path .dirname (os .path .abspath (__file__ ))
79
- self .assertEquals (parse_args (['run' ]).tests , here )
79
+ self .assertEqual (parse_args (['run' ]).tests , here )
80
80
tests = '/benchmarks/are/here'
81
- self .assertEquals (parse_args (['run' , '-t' , tests ]).tests , tests )
81
+ self .assertEqual (parse_args (['run' , '-t' , tests ]).tests , tests )
82
82
83
83
def test_optimization_argument (self ):
84
- self .assertEquals (parse_args (['run' ]).optimization , 'O' )
85
- self .assertEquals (
84
+ self .assertEqual (parse_args (['run' ]).optimization , 'O' )
85
+ self .assertEqual (
86
86
parse_args (['run' , '-o' , 'O' ]).optimization , 'O' )
87
- self .assertEquals (
87
+ self .assertEqual (
88
88
parse_args (['run' , '-o' , 'Onone' ]).optimization , 'Onone' )
89
- self .assertEquals (
89
+ self .assertEqual (
90
90
parse_args (['run' , '-o' , 'Osize' ]).optimization , 'Osize' )
91
91
92
92
with captured_output () as (_ , err ):
@@ -99,9 +99,9 @@ def test_optimization_argument(self):
99
99
err .getvalue ())
100
100
101
101
def test_independent_samples (self ):
102
- self .assertEquals (parse_args (['run' ]).independent_samples , 1 )
103
- self .assertEquals (parse_args (['run' , '-i' , '3' ]).independent_samples ,
104
- 3 )
102
+ self .assertEqual (parse_args (['run' ]).independent_samples , 1 )
103
+ self .assertEqual (parse_args (['run' , '-i' , '3' ]).independent_samples ,
104
+ 3 )
105
105
with captured_output () as (out , err ):
106
106
self .assertRaises (SystemExit ,
107
107
parse_args , ['run' , '-i' , '-3' ])
@@ -112,7 +112,7 @@ def test_independent_samples(self):
112
112
113
113
def test_output_dir (self ):
114
114
self .assertIsNone (parse_args (['run' ]).output_dir )
115
- self .assertEquals (
115
+ self .assertEqual (
116
116
parse_args (['run' , '--output-dir' , '/log' ]).output_dir , '/log' )
117
117
118
118
def test_check_supports_vebose_output (self ):
@@ -169,12 +169,12 @@ def setUp(self):
169
169
self .subprocess_mock = SubprocessMock ()
170
170
171
171
def test_test_harness (self ):
172
- self .assertEquals (
172
+ self .assertEqual (
173
173
BenchmarkDriver (self .args , tests = ['ignored' ]).test_harness ,
174
174
'/benchmarks/Benchmark_O' )
175
175
self .args .tests = '/path'
176
176
self .args .optimization = 'Suffix'
177
- self .assertEquals (
177
+ self .assertEqual (
178
178
BenchmarkDriver (self .args , tests = ['ignored' ]).test_harness ,
179
179
'/path/Benchmark_Suffix' )
180
180
@@ -185,10 +185,10 @@ def test_gets_list_of_precommit_benchmarks(self):
185
185
driver = BenchmarkDriver (
186
186
self .args , _subprocess = self .subprocess_mock )
187
187
self .subprocess_mock .assert_called_all_expected ()
188
- self .assertEquals (driver .tests ,
189
- ['Benchmark1' , 'Benchmark2' ])
190
- self .assertEquals (driver .all_tests ,
191
- ['Benchmark1' , 'Benchmark2' ])
188
+ self .assertEqual (driver .tests ,
189
+ ['Benchmark1' , 'Benchmark2' ])
190
+ self .assertEqual (driver .all_tests ,
191
+ ['Benchmark1' , 'Benchmark2' ])
192
192
self .assertEquals (driver .test_number ['Benchmark1' ], "1" )
193
193
self .assertEquals (driver .test_number ['Benchmark2' ], "2" )
194
194
@@ -207,19 +207,19 @@ def test_gets_list_of_all_benchmarks_when_benchmarks_args_exist(self):
207
207
driver = BenchmarkDriver (
208
208
self .args , _subprocess = self .subprocess_mock )
209
209
self .subprocess_mock .assert_called_all_expected ()
210
- self .assertEquals (driver .tests , ['Benchmark1' , 'Benchmark3' ])
211
- self .assertEquals (driver .all_tests ,
212
- ['Benchmark1' , 'Benchmark2' , 'Benchmark3' ])
210
+ self .assertEqual (driver .tests , ['Benchmark1' , 'Benchmark3' ])
211
+ self .assertEqual (driver .all_tests ,
212
+ ['Benchmark1' , 'Benchmark2' , 'Benchmark3' ])
213
213
214
214
def test_filters_benchmarks_by_pattern (self ):
215
215
self .args .filters = '-f .+3' .split ()
216
216
self .subprocess_mock .expect (* self .list_all_tests )
217
217
driver = BenchmarkDriver (
218
218
self .args , _subprocess = self .subprocess_mock )
219
219
self .subprocess_mock .assert_called_all_expected ()
220
- self .assertEquals (driver .tests , ['Benchmark3' ])
221
- self .assertEquals (driver .all_tests ,
222
- ['Benchmark1' , 'Benchmark2' , 'Benchmark3' ])
220
+ self .assertEqual (driver .tests , ['Benchmark3' ])
221
+ self .assertEqual (driver .all_tests ,
222
+ ['Benchmark1' , 'Benchmark2' , 'Benchmark3' ])
223
223
224
224
def test_log_file (self ):
225
225
"""When swift-repo is set, log is tied to Git branch and revision."""
@@ -230,8 +230,8 @@ def test_log_file(self):
230
230
driver = BenchmarkDriver (
231
231
Stub (output_dir = '/path' , tests = '/bin/' , optimization = 'Suffix' ,
232
232
swift_repo = None ,), tests = ['ignored' ])
233
- self .assertEquals (driver .log_file ,
234
- '/path/Benchmark_Suffix-' + now + '.log' )
233
+ self .assertEqual (driver .log_file ,
234
+ '/path/Benchmark_Suffix-' + now + '.log' )
235
235
236
236
r = '/repo/'
237
237
subprocess_mock = SubprocessMock (responses = [
@@ -243,8 +243,8 @@ def test_log_file(self):
243
243
driver = BenchmarkDriver (
244
244
Stub (output_dir = '/log/' , tests = '' , optimization = 'S' , swift_repo = r ),
245
245
tests = ['ignored' ], _subprocess = subprocess_mock )
246
- self .assertEquals (driver .log_file ,
247
- '/log/branch/Benchmark_S-' + now + '-short_hash.log' )
246
+ self .assertEqual (driver .log_file ,
247
+ '/log/branch/Benchmark_S-' + now + '-short_hash.log' )
248
248
subprocess_mock .assert_called_all_expected ()
249
249
250
250
@@ -332,14 +332,14 @@ def test_run_benchmark_independent_samples(self):
332
332
"""Extract up to 20 measurements from an independent run."""
333
333
self .driver .args .independent_samples = 3
334
334
r = self .driver .run_independent_samples ('b1' )
335
- self .assertEquals (self .subprocess_mock .calls .count (
335
+ self .assertEqual (self .subprocess_mock .calls .count (
336
336
('/benchmarks/Benchmark_O' , 'b1' , '--num-iters=1' , '--memory' ,
337
337
'--quantile=20' , '--delta' )), 3 )
338
- self .assertEquals (r .num_samples , 3 ) # results are merged
338
+ self .assertEqual (r .num_samples , 3 ) # results are merged
339
339
340
340
def test_run_and_log (self ):
341
341
def mock_run (test ):
342
- self .assertEquals (test , 'b1' )
342
+ self .assertEqual (test , 'b1' )
343
343
return PerformanceTestResult (
344
344
'3,b1,5,101,1,1,1,1,888' .split (',' ),
345
345
quantiles = True , delta = True , memory = True )
@@ -352,8 +352,8 @@ def mock_run(test):
352
352
header = '#,TEST,SAMPLES,MIN(μs),Q1(μs),MEDIAN(μs),Q3(μs),MAX(μs),' + \
353
353
'MAX_RSS(B)\n '
354
354
csv_log = '3,b1,5,101,102,103,104,105,888\n '
355
- self .assertEquals (log , None )
356
- self .assertEquals (
355
+ self .assertEqual (log , None )
356
+ self .assertEqual (
357
357
out .getvalue (),
358
358
header +
359
359
csv_log +
@@ -363,8 +363,8 @@ def mock_run(test):
363
363
with captured_output () as (out , _ ):
364
364
log = driver .run_and_log (csv_console = False )
365
365
366
- self .assertEquals (log , header + csv_log )
367
- self .assertEquals (
366
+ self .assertEqual (log , header + csv_log )
367
+ self .assertEqual (
368
368
out .getvalue (),
369
369
' # TEST SAMPLES MIN(μs) Q1(μs)' +
370
370
' MEDIAN(μs) Q3(μs) MAX(μs) MAX_RSS(B)\n ' +
@@ -376,11 +376,11 @@ def mock_run(test):
376
376
def test_log_results (self ):
377
377
"""Create log directory if it doesn't exist and write the log file."""
378
378
def assert_log_written (out , log_file , content ):
379
- self .assertEquals (out .getvalue (),
380
- 'Logging results to: ' + log_file + '\n ' )
379
+ self .assertEqual (out .getvalue (),
380
+ 'Logging results to: ' + log_file + '\n ' )
381
381
with open (log_file , 'rU' ) as f :
382
382
text = f .read ()
383
- self .assertEquals (text , "formatted output" )
383
+ self .assertEqual (text , "formatted output" )
384
384
385
385
try :
386
386
import tempfile # setUp
@@ -408,7 +408,7 @@ def assert_log_written(out, log_file, content):
408
408
def test_deterministing_hashing (self ):
409
409
cmd = ['printenv' , 'SWIFT_DETERMINISTIC_HASHING' ]
410
410
driver = BenchmarkDriver (['no args' ], tests = ['ignored' ])
411
- self .assertEquals (driver ._invoke (cmd ).strip (), '1' )
411
+ self .assertEqual (driver ._invoke (cmd ).strip (), '1' )
412
412
413
413
414
414
class BenchmarkDriverMock (Mock ):
@@ -437,31 +437,31 @@ def test_plain_log_format(self):
437
437
'name' : 'Base.category' , 'level' : logging .DEBUG ,
438
438
'levelname' : 'DEBUG' , 'msg' : 'Hi!' })
439
439
f = LoggingReportFormatter ()
440
- self .assertEquals (f .format (lr ), 'DEBUG category: Hi!' )
440
+ self .assertEqual (f .format (lr ), 'DEBUG category: Hi!' )
441
441
442
442
def test_colored_log_format (self ):
443
443
def record (level , level_name ):
444
444
return logging .makeLogRecord ({
445
445
'name' : 'Base.category' , 'levelno' : level ,
446
446
'levelname' : level_name , 'msg' : 'Hi!' })
447
447
f = LoggingReportFormatter (use_color = True )
448
- self .assertEquals (f .format (record (logging .DEBUG , 'DEBUG' )),
449
- '\x1b [1;39mcategory: Hi!\x1b [1;0m' )
450
- self .assertEquals (f .format (record (logging .INFO , 'INFO' )),
451
- '\x1b [1;32mcategory: Hi!\x1b [1;0m' )
452
- self .assertEquals (f .format (record (logging .WARNING , 'WARNING' )),
453
- '\x1b [1;33mcategory: Hi!\x1b [1;0m' )
454
- self .assertEquals (f .format (record (logging .ERROR , 'ERROR' )),
455
- '\x1b [1;31mcategory: Hi!\x1b [1;0m' )
456
- self .assertEquals (f .format (record (logging .CRITICAL , 'CRITICAL' )),
457
- '\x1b [1;35mcategory: Hi!\x1b [1;0m' )
448
+ self .assertEqual (f .format (record (logging .DEBUG , 'DEBUG' )),
449
+ '\x1b [1;39mcategory: Hi!\x1b [1;0m' )
450
+ self .assertEqual (f .format (record (logging .INFO , 'INFO' )),
451
+ '\x1b [1;32mcategory: Hi!\x1b [1;0m' )
452
+ self .assertEqual (f .format (record (logging .WARNING , 'WARNING' )),
453
+ '\x1b [1;33mcategory: Hi!\x1b [1;0m' )
454
+ self .assertEqual (f .format (record (logging .ERROR , 'ERROR' )),
455
+ '\x1b [1;31mcategory: Hi!\x1b [1;0m' )
456
+ self .assertEqual (f .format (record (logging .CRITICAL , 'CRITICAL' )),
457
+ '\x1b [1;35mcategory: Hi!\x1b [1;0m' )
458
458
459
459
def test_no_prefix_for_base_logging (self ):
460
460
lr = logging .makeLogRecord ({
461
461
'name' : 'Base' , 'level' : logging .INFO ,
462
462
'levelname' : 'INFO' , 'msg' : 'Hi!' })
463
463
f = LoggingReportFormatter ()
464
- self .assertEquals (f .format (lr ), 'INFO Hi!' )
464
+ self .assertEqual (f .format (lr ), 'INFO Hi!' )
465
465
466
466
467
467
class TestMarkdownReportHandler (unittest .TestCase ):
@@ -481,7 +481,7 @@ def record(self, level, category, msg):
481
481
'levelno' : level , 'msg' : msg })
482
482
483
483
def test_init_writes_table_header (self ):
484
- self .assertEquals (self .handler .level , logging .INFO )
484
+ self .assertEqual (self .handler .level , logging .INFO )
485
485
self .assert_contains (['Benchmark Check Report\n ' , '---|---' ])
486
486
487
487
def test_close_writes_final_newlines (self ):
@@ -552,7 +552,7 @@ def test_uses_logging(self):
552
552
with captured_output () as (out , _ ):
553
553
BenchmarkDoctor (self .args , driver )
554
554
self .assert_contains (['Checking tests: B1, B2' ], self .logs ['debug' ])
555
- self .assertEquals (out .getvalue (), '' )
555
+ self .assertEqual (out .getvalue (), '' )
556
556
557
557
def test_supports_verbose_output (self ):
558
558
driver = BenchmarkDriverMock (tests = ['B1' , 'B2' ])
@@ -607,7 +607,7 @@ def test_measure_10_independent_1s_benchmark_series(self):
607
607
'B1 O i1a' , 'B1 O i1b' , 'B1 O i1c' , 'B1 O i1d' , 'B1 O i1e' ,
608
608
'B1 O i2a' , 'B1 O i2b' , 'B1 O i2c' , 'B1 O i2d' , 'B1 O i2e' ],
609
609
measurements .keys ())
610
- self .assertEquals (measurements ['name' ], 'B1' )
610
+ self .assertEqual (measurements ['name' ], 'B1' )
611
611
self .assert_contains (
612
612
['Calibrating num-samples for B1:' ,
613
613
'Runtime 300 μs yields 4096 adjusted samples per second.' ,
0 commit comments