@@ -325,7 +325,7 @@ uop_dealloc(_PyUOpExecutorObject *self) {
325
325
}
326
326
327
327
const char *
328
- _PyUopName (int index )
328
+ _PyUOpName (int index )
329
329
{
330
330
if (index <= MAX_REAL_OPCODE ) {
331
331
return _PyOpcode_OpName [index ];
@@ -347,7 +347,7 @@ uop_item(_PyUOpExecutorObject *self, Py_ssize_t index)
347
347
PyErr_SetNone (PyExc_IndexError );
348
348
return NULL ;
349
349
}
350
- const char * name = _PyUopName (self -> trace [index ].opcode );
350
+ const char * name = _PyUOpName (self -> trace [index ].opcode );
351
351
if (name == NULL ) {
352
352
name = "<nil>" ;
353
353
}
@@ -388,7 +388,7 @@ PyTypeObject _PyUOpExecutor_Type = {
388
388
389
389
/* TO DO -- Generate these tables */
390
390
static const uint16_t
391
- _PyUop_Replacements [OPCODE_METADATA_SIZE ] = {
391
+ _PyUOp_Replacements [OPCODE_METADATA_SIZE ] = {
392
392
[_ITER_JUMP_RANGE ] = _GUARD_NOT_EXHAUSTED_RANGE ,
393
393
[_ITER_JUMP_LIST ] = _GUARD_NOT_EXHAUSTED_LIST ,
394
394
[_ITER_JUMP_TUPLE ] = _GUARD_NOT_EXHAUSTED_TUPLE ,
@@ -451,7 +451,7 @@ translate_bytecode_to_trace(
451
451
#define ADD_TO_TRACE (OPCODE , OPARG , OPERAND , TARGET ) \
452
452
DPRINTF(2, \
453
453
" ADD_TO_TRACE(%s, %d, %" PRIu64 ")\n", \
454
- _PyUopName (OPCODE), \
454
+ _PyUOpName (OPCODE), \
455
455
(OPARG), \
456
456
(uint64_t)(OPERAND)); \
457
457
assert(trace_length < max_length); \
@@ -474,7 +474,7 @@ translate_bytecode_to_trace(
474
474
}
475
475
476
476
// Reserve space for N uops, plus 3 for _SET_IP, _CHECK_VALIDITY and _EXIT_TRACE
477
- #define RESERVE (needed ) RESERVE_RAW((needed) + 3, _PyUopName (opcode))
477
+ #define RESERVE (needed ) RESERVE_RAW((needed) + 3, _PyUOpName (opcode))
478
478
479
479
// Trace stack operations (used by _PUSH_FRAME, _POP_FRAME)
480
480
#define TRACE_STACK_PUSH () \
@@ -546,8 +546,8 @@ translate_bytecode_to_trace(
546
546
uint32_t uopcode = BRANCH_TO_GUARD [opcode - POP_JUMP_IF_FALSE ][jump_likely ];
547
547
_Py_CODEUNIT * next_instr = instr + 1 + _PyOpcode_Caches [_PyOpcode_Deopt [opcode ]];
548
548
DPRINTF (4 , "%s(%d): counter=%x, bitcount=%d, likely=%d, uopcode=%s\n" ,
549
- _PyUopName (opcode ), oparg ,
550
- counter , bitcount , jump_likely , _PyUopName (uopcode ));
549
+ _PyUOpName (opcode ), oparg ,
550
+ counter , bitcount , jump_likely , _PyUOpName (uopcode ));
551
551
ADD_TO_TRACE (uopcode , max_length , 0 , target );
552
552
if (jump_likely ) {
553
553
_Py_CODEUNIT * target_instr = next_instr + oparg ;
@@ -615,8 +615,8 @@ translate_bytecode_to_trace(
615
615
oparg += extras ;
616
616
}
617
617
}
618
- if (_PyUop_Replacements [uop ]) {
619
- uop = _PyUop_Replacements [uop ];
618
+ if (_PyUOp_Replacements [uop ]) {
619
+ uop = _PyUOp_Replacements [uop ];
620
620
if (uop == _FOR_ITER_TIER_TWO ) {
621
621
target += 1 + INLINE_CACHE_ENTRIES_FOR_ITER + oparg + 1 ;
622
622
assert (_PyCode_CODE (code )[target - 1 ].op .code == END_FOR ||
@@ -712,7 +712,7 @@ translate_bytecode_to_trace(
712
712
}
713
713
break ;
714
714
}
715
- DPRINTF (2 , "Unsupported opcode %s\n" , _PyUopName (opcode ));
715
+ DPRINTF (2 , "Unsupported opcode %s\n" , _PyUOpName (opcode ));
716
716
OPT_UNSUPPORTED_OPCODE (opcode );
717
717
goto done ; // Break out of loop
718
718
} // End default
@@ -832,7 +832,7 @@ make_executor_from_uops(_PyUOpInstruction *buffer, _PyBloomFilter *dependencies)
832
832
dest -- ;
833
833
}
834
834
assert (dest == -1 );
835
- executor -> base .execute = _PyUopExecute ;
835
+ executor -> base .execute = _PyUOpExecute ;
836
836
_Py_ExecutorInit ((_PyExecutorObject * )executor , dependencies );
837
837
#ifdef Py_DEBUG
838
838
char * python_lltrace = Py_GETENV ("PYTHON_LLTRACE" );
@@ -845,7 +845,7 @@ make_executor_from_uops(_PyUOpInstruction *buffer, _PyBloomFilter *dependencies)
845
845
for (int i = 0 ; i < length ; i ++ ) {
846
846
printf ("%4d %s(%d, %d, %" PRIu64 ")\n" ,
847
847
i ,
848
- _PyUopName (executor -> trace [i ].opcode ),
848
+ _PyUOpName (executor -> trace [i ].opcode ),
849
849
executor -> trace [i ].oparg ,
850
850
executor -> trace [i ].target ,
851
851
executor -> trace [i ].operand );
@@ -888,11 +888,11 @@ uop_optimize(
888
888
return 1 ;
889
889
}
890
890
891
- /* Dummy execute() function for Uop Executor.
891
+ /* Dummy execute() function for UOp Executor.
892
892
* The actual implementation is inlined in ceval.c,
893
893
* in _PyEval_EvalFrameDefault(). */
894
894
_PyInterpreterFrame *
895
- _PyUopExecute (_PyExecutorObject * executor , _PyInterpreterFrame * frame , PyObject * * stack_pointer )
895
+ _PyUOpExecute (_PyExecutorObject * executor , _PyInterpreterFrame * frame , PyObject * * stack_pointer )
896
896
{
897
897
Py_FatalError ("Tier 2 is now inlined into Tier 1" );
898
898
}
0 commit comments