Skip to content

Commit

Permalink
Add comments pointing to issue NVIDIA#3064
Browse files Browse the repository at this point in the history
  • Loading branch information
rwgk committed Dec 5, 2024
1 parent 29f0e52 commit f3fffa9
Showing 1 changed file with 4 additions and 2 deletions.
6 changes: 4 additions & 2 deletions python/cuda_parallel/cuda/parallel/experimental/_iterators.py
Original file line number Diff line number Diff line change
Expand Up @@ -300,7 +300,7 @@ def __init__(
transform_dereference,
op_return_ntype(numba.types.CPointer(numba.types.char)),
),
# ATTENTION: NOT op_caller here!
# ATTENTION: NOT op_caller here! (see issue #3064)
_ncc(op_abi_name, op, op_return_ntype(it.ntype)),
]

Expand All @@ -318,6 +318,8 @@ def alignment(self):


def cumap(op, it, op_return_ntype):
# TODO(rwgk): Resolve issue #3064

op_return_ntype_ir = _ir_type_given_numba_type(op_return_ntype)
if op_return_ntype_ir is None:
raise RuntimeError(f"Unsupported: {type(op_return_ntype)=}")
Expand Down Expand Up @@ -414,7 +416,7 @@ def transform_advance(it_state_ptr, diff):
source_advance(it_state_ptr, diff) # just a function call

def transform_dereference(it_state_ptr):
# ATTENTION: op_caller here
# ATTENTION: op_caller here (see issue #3064)
return op_caller(source_dereference(it_state_ptr))

return TransformIterator(
Expand Down

0 comments on commit f3fffa9

Please sign in to comment.