pku-liang / FlexTensor

Automatic Schedule Exploration and Optimization Framework for Tensor Computations

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

optimize_block_celluar.py cannot work with '--target cuda'

Light-of-Hers opened this issue · comments

Just tested 4 files: optimize_block_celluar.py, optimize_conv1d.py, optimize_conv2d.py, optimize_conv3d.py

  • optimize_block_celluar.py:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'AssertionError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 78, in build_func
    s, bufs = schedule_with_config(task_key, configs, op_pos=op_pos)
  File "../../../auto_schedule/testing/scheduler.py", line 1711, in schedule_with_config
    template(s, op)
  File "../../../auto_schedule/testing/scheduler.py", line 1263, in _cuda_schedule_split_reorder_fuse
    assert pos < len(spatial_remainder)
AssertionError
op build fail:
<class 'RuntimeError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 82, in build_func
    raise RuntimeError("Invalid %s(%d) kernel"%(task.target, task.dev_id))
RuntimeError: Invalid cuda(0) kernel
op build fail:Invalid cuda(0) kernel
  • optimize_conv1d.py:
<class 'queue.Empty'>
op run fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'queue.Empty'>
op run fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'queue.Empty'>
op build fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'RuntimeError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 82, in build_func
    raise RuntimeError("Invalid %s(%d) kernel"%(task.target, task.dev_id))
RuntimeError: Invalid cuda(0) kernel
op build fail:Invalid cuda(0) kernel
  • optimize_conv3d.py:
<class 'queue.Empty'>
op run fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'RuntimeError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 82, in build_func
    raise RuntimeError("Invalid %s(%d) kernel"%(task.target, task.dev_id))
RuntimeError: Invalid cuda(0) kernel
op build fail:Invalid cuda(0) kernel
<class 'RuntimeError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 82, in build_func
    raise RuntimeError("Invalid %s(%d) kernel"%(task.target, task.dev_id))
RuntimeError: Invalid cuda(0) kernel
op build fail:Invalid cuda(0) kernel
<class 'queue.Empty'>
op run fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'queue.Empty'>
op run fail:
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 1525, in get
    res = self.q.get(block=True, timeout=timeout)
  File "/usr/lib/python3.5/multiprocessing/queues.py", line 105, in get
    raise Empty
queue.Empty
<class 'RuntimeError'>
Traceback (most recent call last):
  File "../../../auto_schedule/testing/scheduler.py", line 161, in exec_func
    res = func(*args, **kwargs)
  File "../../../auto_schedule/testing/scheduler.py", line 82, in build_func
    raise RuntimeError("Invalid %s(%d) kernel"%(task.target, task.dev_id))
RuntimeError: Invalid cuda(0) kernel
op build fail:Invalid cuda(0) kernel

and optimize_conv2d.py didn't output anything even with '--target llvm'

commented

what about using larger timeout?

what about using larger timeout?

timeout seems not the problem.

and now it can work (just not printing the errors occurred while running), but will cost lots of time even with very small size:

> python optimize_block_circulant_matrix.py --target cuda --trials 1 -f 0 -t 1
Optimize block_circulant_matrix shape [4, 4, 2]
[Warning] Directory lib is not empty, but reusing it
warm up [0.007083300000000001, 0.006464, 0.007751500000000001, 0.0065363, 0.0062628, 0.006461500000000001, 0.0064576, 0.0064043, 0.0063421, 0.0063141, 0.0063750000000000005, 0.006407899999999999, 0.011866600000000001, 0.006450900000000001, inf, 0.0065644, 0.0063054, 0.0064706, 0.0065644, inf]
warm up [inf, inf, inf, inf, 0.0064112, 0.006392599999999999, 0.0064315, 0.0064104999999999995, inf, 0.0062635, 0.006321, 0.006320299999999999, 0.0063985000000000005, 0.0063033, 0.0064278, inf, inf, 0.006300300000000001, 0.0062269, inf]
warm up [0.0064101, 0.0064448000000000005, 0.0063631, 0.007027, inf, 0.006403599999999999, inf, inf, 0.0064766, 0.0064311, 0.0063693000000000005, 0.0064363, 0.006487199999999999, 0.0062851, 0.0064302, 0.006335400000000001, 0.0083196, 0.006477500000000001, 0.0065535, 0.0064475]
warm up [inf, inf, inf, 0.006220400000000001, 0.006448199999999999, 0.0064288999999999995, inf, 0.0064307999999999995, inf, inf, 0.0063108999999999995, 0.0064253999999999995, 0.006099299999999999, inf, inf, inf, 0.006558199999999999, 0.0062876, 0.006305, 0.0062185]
warm up [0.008112499999999998, 0.0078879, inf, 0.0062740999999999995, 0.006189600000000001, 0.006297, 0.0064106, 0.0063939, 0.0061207, inf, 0.006435000000000001, 0.0062702, 0.0064082, inf, inf, 0.0063638, inf, 0.006273600000000001, inf, inf]
warm up [0.0063718, 0.0061918, 0.006278000000000001, 0.0065124, 0.0064921, inf, 0.0063358, inf, 0.006427400000000001, 0.006281299999999999, inf, 0.0062742, inf, 0.006248900000000001, 0.0063563000000000005, 0.0064613000000000005, 0.006326299999999999, inf, 0.0063531, 0.006519700000000001]
warm up [0.006386200000000001, inf, 0.0062494, inf, 0.0061993000000000005, inf, inf, inf, 0.0060618, 0.0064849, inf, inf, 0.0062589, 0.006495300000000001, inf, inf, 0.006480700000000001, 0.0064497, 0.0062311, 0.0063677]
warm up [inf, 0.0065378, 0.006537, inf, inf, inf, 0.0063579, inf, inf, 0.0064034, 0.0064146, 0.0063365, 0.008544699999999999, 0.0064088, inf, 0.006318100000000001, inf, 0.006381700000000001, inf, 0.006404999999999999]
warm up [inf, 0.0063349999999999995, inf, inf, inf, 0.0062586, 0.0063199, 0.0063952, 0.0065164, 0.007663199999999999, 0.0061920000000000005, 0.006328899999999999, 0.0062642, 0.0094932, inf, 0.0061995, inf, inf, 0.0062249, inf]
warm up [inf, 0.0062673, inf, 0.0065657, 0.006202999999999999, inf, 0.0063137, 0.0064050999999999995, 0.0062193000000000005, 0.006232, 0.0091248, inf, inf, 0.009152, 0.0062871, 0.009897300000000001, 0.006304199999999999, inf, 0.006332699999999999, inf]
warm up [0.006498, 0.0062131999999999995, 0.006266, 0.0063267, 0.006254699999999999, 0.0064712, inf, 0.0062408, 0.0063660999999999995, 0.0077544, inf, 0.013322400000000002, inf, 0.0064789999999999995, 0.0063253, 0.0062508, inf, 0.006400599999999999, 0.0063792, 0.006376999999999999]
warm up [0.0063227, 0.0062693, inf, inf, 0.006449200000000001, 0.0062986, 0.0065087999999999995, 0.006140999999999999, inf, inf, 0.006313, 0.0071853, 0.006439800000000001, inf, inf, 0.0063146999999999995, inf, 0.0063343, 0.0064743, 0.0063579]
warm up [0.0063525, 0.0063938, inf, 0.006396700000000001, 0.006287, inf, 0.007411699999999999, 0.0064876, inf, 0.0062542000000000006, 0.0072101000000000005, 0.0061522, inf, 0.0064218, 0.006286399999999999, 0.0062363, 0.0064251, 0.0064093, 0.0063159, inf]
warm up [0.008410099999999998, inf, 0.006269800000000001, inf, 0.0063558, inf, 0.0063435, 0.0062767000000000005, 0.006211, 0.0063809, 0.0063004, 0.0063105, inf, 0.006407899999999999, inf, 0.0063117, 0.0073118, inf, 0.0062732, 0.006298499999999999]
warm up [0.006251499999999999, inf, inf, inf, 0.0081808, 0.006171799999999999, 0.0062289, 0.0063598, 0.0063528000000000005, 0.0063484, inf, 0.01064, 0.006321, 0.006316100000000001, 0.0063991000000000004, inf, 0.006380000000000001, inf, 0.0062758, 0.0064681]
warm up [inf, 0.006210000000000001, 0.0063728, inf, inf, 0.0062706, inf, inf, 0.0063704, inf, 0.0062341, 0.006279399999999999, 0.0062827000000000004, 0.0064117, 0.0062369999999999995, 0.0065281, 0.006365800000000001, inf, 0.0063064, inf]
warm up [0.0061773999999999996, 0.0063389, 0.0063579, inf, inf, 0.0065154, 0.0063606, 0.006478900000000001, inf, 0.006403799999999999, inf, 0.006349700000000001, 0.006256599999999999, 0.0062477, inf, inf, 0.006300900000000001, inf, 0.0062264, 0.0064281]
warm up [inf, 0.006420200000000001, inf, 0.006388099999999999, 0.0063001, inf, 0.0063777, inf, 0.0061569, inf, inf, 0.0062355, 0.006297, 0.0064077000000000005, inf, inf, inf, inf, 0.0062155, inf]
warm up [inf, inf, 0.006382199999999999, 0.0063771999999999995, 0.006229500000000001, 0.0064195, 0.006371, inf, inf, 0.0063178, 0.0063653, 0.006546700000000001, inf, 0.0062764, 0.006420200000000001, 0.007281, 0.0064299, 0.0064957999999999995, 0.006444999999999999, 0.006324300000000001]
warm up [0.0061771000000000005, 0.006622200000000001, 0.0063113, inf, inf, 0.0062798, inf, 0.0064188000000000005, inf, 0.0062866, inf, 0.0062929, 0.0062525, 0.0062733, inf, 0.0063481, 0.0062908, inf, 0.006343199999999999, inf]
tune [0.006442, 0.006323199999999999, 0.0063589, 0.0063555999999999994, 0.0063094, 0.0064287, 0.0084313, 0.0063321, 0.006233, 0.0063175, 0.0063054, inf, 0.0062578, 0.0062024]
No. 0 | The best currently 0.0060618 {'split_i_0': 3, 'fuse_spatial': 0, 'split_rv_0': 1, 'split_j_1': 6, 'reorder': 1, 'unroll': 0}
re-evaluate [0.006341899999999999, 0.0061945, 0.0063421, 0.0061644, 0.0061574, 0.0061938, 0.006340900000000001, 0.006220799999999999, 0.0061982, 0.006246099999999999]
warm up [0.0061943]
warm up [0.011500499999999999, inf, 0.022492599999999998, 0.0115172, 0.011916600000000001, 0.0123383, 0.012220199999999999, 0.0115242, 0.0114204, 0.0115723, 0.0115237, 0.011544500000000001, 0.0114637, 0.0114128, 0.0114597, 0.011534399999999998, 0.011541000000000001, 0.011465, 0.0128245, 0.0115488]
warm up [0.0115333, 0.011394699999999999, 0.0115222, 0.0114678, 0.0116342, 0.011524399999999999, 0.0116779, 0.0114273, 0.0112983, 0.011492100000000002, 0.011622200000000001, 0.0113376, 0.0115941, 0.011397, 0.0114708, 0.0115291, 0.0115697, 0.011483499999999999, 0.0117066, 0.0113839]
warm up [0.0113981, 0.0112635, 0.0117511, 0.011498900000000001, 0.0116242, 0.011514300000000002, 0.011450799999999999, 0.015381800000000003, 0.0115315, 0.0115271, 0.0116127, 0.0115222, 0.0125229, 0.0114495, 0.0115294, 0.0115021, 0.0116203, 0.0117141, 0.0124191, 0.0115375]
warm up [0.0116583, 0.011547700000000001, 0.011579800000000001, 0.0115844, 0.0115934, 0.011701399999999999, 0.0114499, 0.011386700000000001, 0.011611400000000001, 0.011308599999999999, 0.0114728, 0.0116201, 0.0114135, 0.0125809, 0.0115216, 0.011618800000000002, 0.01156, 0.0115158, 0.0111957, 0.011623199999999998]
warm up [0.0113278, 0.0114328, 0.0147618, 0.0127592, 0.0114791, 0.0114984, 0.011471700000000001, 0.0115208, 0.011519099999999999, 0.012255499999999999, 0.0114381, 0.011352800000000001, 0.011542700000000001, 0.0122259, 0.0115751, 0.0115245, 0.011482500000000001, 0.012905400000000001, 0.0115303, 0.015014399999999999]
warm up [0.0114268, 0.011610899999999999, 0.0115214, 0.0114587, 0.0142167, 0.011419899999999998, 0.0114874, 0.011509799999999999, 0.0114459, 0.0115511, 0.011282599999999999, 0.011339599999999998, 0.012604900000000002, 0.011475899999999999, 0.011403, 0.0115129, 0.011601599999999998, 0.0115519, 0.013524000000000001, 0.0113635]
warm up [0.0114793, 0.013215399999999999, 0.0113629, 0.0115436, 0.0115125, 0.0132463, 0.011457700000000001, 0.0116055, 0.011580199999999999, 0.0114112, 0.011480800000000001, 0.0129048, 0.011589299999999999, 0.0194007, 0.0115516, 0.0115079, 0.011440800000000001, 0.0114694, 0.0116306, 0.0116335]
warm up [0.0114472, 0.0114211, 0.0112735, 0.0111394, 0.0125611, 0.011395, 0.011654999999999999, 0.0115994, 0.0115591, 0.0114128, 0.011524399999999999, 0.0115822, 0.0114597, 0.0115008, 0.011641, 0.0116263, 0.011534199999999998, 0.015471799999999999, 0.0115268, 0.0117065]
warm up [0.011654099999999999, 0.011567, 0.0115918, 0.011505099999999999, 0.0115801, 0.0113497, 0.0113449, 0.011415699999999999, 0.0114186, 0.011573700000000001, 0.011448, 0.0115635, 0.011463099999999999, 0.011443700000000001, 0.0113925, 0.0114186, 0.011469, 0.0117475, 0.011345000000000001, 0.011388]
warm up [0.0118748, 0.012604500000000001, 0.0114779, 0.0115172, 0.0126965, 0.0113351, 0.0113757, 0.01178, 0.0111924, 0.0115235, 0.011256, 0.011542, 0.011473400000000002, 0.0115742, 0.0121977, 0.0114478, 0.011503, 0.0112642, 0.0159121, 0.0114783]
warm up [0.0148706, 0.0114164, 0.0115958, 0.014794100000000001, 0.0115449, 0.01236, 0.0115706, 0.0114713, 0.011473400000000002, 0.01151, 0.0123642, 0.011489000000000001, 0.0115911, 0.011644, 0.011409, 0.011473500000000001, 0.0114863, 0.0115932, 0.0115272, 0.011482800000000001]
warm up [0.0114968, 0.011498900000000001, 0.0114763, 0.0115116, 0.011514, 0.0113255, 0.011542799999999999, 0.0114728, 0.0115409, 0.0113868, 0.0115981, 0.0114767, 0.0114099, 0.0116642, 0.011417199999999999, 0.0113578, 0.0114883, 0.0114944, 0.0115243, 0.0114912]
warm up [0.0116087, 0.011536, 0.011164, 0.0115313, 0.011340900000000001, 0.0114449, 0.011500699999999999, 0.0113875, 0.0122908, 0.0112992, 0.0113757, 0.0113743, 0.012242099999999999, 0.0114319, 0.0114503, 0.0116017, 0.0114278, 0.0115075, 0.011557600000000001, 0.0116172]
warm up [0.011553800000000001, 0.0114223, 0.011616399999999999, 0.011518899999999999, 0.011501200000000001, 0.011314399999999999, 0.0115691, 0.0112667, 0.0115249, 0.011447699999999998, 0.0114765, 0.0114526, 0.0115122, 0.0115126, 0.0116367, 0.0121897, 0.0114861, 0.0113846, 0.011593299999999999, 0.0113712]
warm up [0.0117272, 0.0114012, 0.0115385, 0.0115648, 0.0114838, 0.0114557, 0.011497199999999999, 0.0115337, 0.0113641, 0.0114012, 0.0113357, 0.0115033, 0.011585699999999999, 0.011422999999999999, 0.0115172, 0.0116538, 0.011584500000000001, 0.0113703, 0.011584700000000002, 0.011586099999999998]
warm up [0.011331800000000001, 0.0113648, 0.011450199999999999, 0.0115866, 0.011609300000000001, 0.0115085, 0.011460399999999999, 0.0113645, 0.011511, 0.0113787, 0.0116293, 0.0115092, 0.011576999999999999, 0.011548200000000002, 0.0114719, 0.0115043, 0.0115031, 0.0115176, 0.0113299, 0.0114383]
warm up [0.0115395, 0.0127575, 0.0115487, 0.011497499999999999, 0.011570500000000001, 0.0112606, 0.011420600000000001, 0.015881199999999998, 0.0115979, 0.011497499999999999, 0.0114878, 0.011523400000000001, 0.011457199999999999, 0.011493499999999999, 0.011422899999999998, 0.0115382, 0.0113413, 0.0114917, 0.011321, 0.0115566]
warm up [0.011595, 0.011412, 0.0114675, 0.011336399999999998, 0.0115919, 0.011567899999999999, 0.011398499999999999, 0.011538, 0.0116166, 0.0114214, 0.0115207, 0.011419899999999998, 0.0115288, 0.0113048, 0.0115864, 0.0115192, 0.0115194, 0.0116142, 0.0116123, 0.0115168]
warm up [0.0115204, 0.0115267, 0.011567, 0.0114662, 0.0114197, 0.0114327, 0.0113663, 0.0114291, 0.0113539, 0.011504400000000001, 0.011306400000000001, 0.0115213, 0.011326, 0.0114752, 0.0114722, 0.011554, 0.0114023, 0.011356400000000001, 0.0116396, 0.011422999999999999]
warm up [0.011430200000000001, 0.011306, 0.0114541, 0.0116057, 0.0115376, 0.011612500000000001, 0.0114974, 0.0115052, 0.0113607, 0.011363400000000001, 0.011529000000000001, 0.011487, 0.0115136, 0.0113495, 0.0112876, 0.0115407, 0.011471, 0.0114644, 0.0115503, 0.0114737]
tune [0.01136, 0.0115447, 0.0114685, 0.0114824, 0.011626999999999998, 0.011369100000000002, 0.012512, 0.0115327, 0.0115275, 0.011429199999999999, 0.0115436, 0.011479600000000001, 0.011467900000000001, 0.0114133, 0.0115128, 0.011461, 0.011464900000000002, 0.011367599999999999, 0.0116429, 0.0116106, 0.0115998, 0.0114078, 0.011388999999999998, 0.0113518, 0.0115277, 0.011441699999999999, 0.0114129, 0.011560100000000002, 0.0113536, 0.0115604]
No. 0 | The best currently 0.0111394 {'unroll': 7, 'split_i_0': 1, 'reorder': 1, 'split_j_1': 6, 'fuse_spatial': 0}
re-evaluate [0.011503, 0.011508800000000001, 0.011608, 0.0116181, 0.015582, 0.0115897, 0.0115874, 0.011503399999999999, 0.0117333, 0.0115564]
warm up [0.0115369]
warm up [0.0115364, 0.0151397, 0.0115419, 0.019299, 0.011515099999999999, 0.0115147, 0.0115082, 0.011355299999999999, 0.011574599999999999, 0.0115365, 0.0116082, 0.0115849, 0.011484799999999998, 0.0158087, 0.011473400000000002, 0.0114339, 0.0135589, 0.0114872, 0.0112671, 0.0115515]
warm up [0.011524999999999999, 0.0115094, 0.011421500000000001, 0.0116122, 0.011549199999999999, 0.011252999999999999, 0.0113669, 0.0114537, 0.0113569, 0.011374200000000001, 0.0113432, 0.011527599999999999, 0.0112494, 0.011324299999999999, 0.0116436, 0.0115185, 0.011574599999999999, 0.011610899999999999, 0.0113317, 0.011493999999999999]
warm up [0.0112526, 0.0114544, 0.011552999999999999, 0.011431799999999999, 0.011566, 0.0115684, 0.0114051, 0.011356600000000001, 0.011426099999999998, 0.011589, 0.011469499999999999, 0.0113726, 0.0114478, 0.0115672, 0.011477000000000001, 0.011470200000000002, 0.011482, 0.0113797, 0.0114838, 0.0116733]
warm up [0.011356000000000002, 0.014284399999999997, 0.01149, 0.0113663, 0.0116365, 0.011595900000000001, 0.011469299999999998, 0.011453799999999998, 0.011365300000000002, 0.0115019, 0.016056800000000003, 0.011621800000000002, 0.011562900000000001, 0.0113546, 0.0113708, 0.0113758, 0.011487, 0.011534599999999999, 0.0115563, 0.0116366]
warm up [0.0114355, 0.0115227, 0.011450899999999998, 0.0113635, 0.0115111, 0.0114423, 0.011386199999999999, 0.0113816, 0.0123339, 0.0113624, 0.011483, 0.0114947, 0.0115552, 0.0147223, 0.011614099999999999, 0.0114591, 0.0114638, 0.011360499999999999, 0.0114373, 0.0113834]
warm up [0.0113568, 0.0115223, 0.0112571, 0.0115489, 0.0115414, 0.0115974, 0.0115153, 0.0152776, 0.011439500000000002, 0.0114576, 0.0113051, 0.011413799999999998, 0.0115678, 0.011434600000000001, 0.0115989, 0.0114998, 0.013962999999999998, 0.0114419, 0.0115345, 0.0114282]
warm up [0.011522, 0.0113229, 0.0114128, 0.0114849, 0.0114527, 0.01158, 0.0113902, 0.0115369, 0.0115663, 0.0171739, 0.011471100000000001, 0.0112623, 0.0115704, 0.0115852, 0.0114871, 0.0115453, 0.0114135, 0.0114156, 0.0114088, 0.011419]
warm up [0.01144, 0.0116348, 0.0114434, 0.0112878, 0.01156, 0.0116043, 0.0115739, 0.011538300000000001, 0.0113702, 0.0147641, 0.0113781, 0.0115275, 0.0114235, 0.0114387, 0.011503399999999999, 0.0114947, 0.011493, 0.0113284, 0.011349900000000001, 0.0115156]
warm up [0.011481799999999999, 0.011291899999999999, 0.011387099999999999, 0.0115471, 0.011545399999999999, 0.0113253, 0.0112644, 0.01132, 0.0114836, 0.0114497, 0.0115133, 0.011541000000000001, 0.0112522, 0.0115195, 0.0113927, 0.011347300000000001, 0.011527999999999998, 0.0114936, 0.011524399999999999, 0.0116087]
warm up [0.0115691, 0.011464900000000002, 0.0114745, 0.011473200000000001, 0.0116399, 0.011601799999999999, 0.0115295, 0.0114801, 0.011493, 0.0114629, 0.0113794, 0.0114452, 0.011571399999999999, 0.011408600000000001, 0.0113888, 0.0113151, 0.011436300000000002, 0.0115784, 0.0114698, 0.0115162]
warm up [0.0115441, 0.011425900000000001, 0.0114249, 0.0113515, 0.0113853, 0.011278400000000001, 0.01141, 0.0112019, 0.011447299999999999, 0.011509499999999999, 0.011519800000000002, 0.0115107, 0.0112862, 0.0115356, 0.0115782, 0.0114337, 0.01144, 0.011503900000000001, 0.011652800000000001, 0.011492700000000002]
warm up [0.0116035, 0.0115309, 0.0114396, 0.0160267, 0.011450700000000001, 0.0114896, 0.0113498, 0.011294100000000001, 0.011558299999999999, 0.0115532, 0.0112942, 0.01142, 0.0123225, 0.0115269, 0.0116017, 0.0116005, 0.011541000000000001, 0.0114501, 0.014915, 0.0115313]
warm up [0.0119462, 0.011448600000000001, 0.0114382, 0.0114384, 0.0116113, 0.01535, 0.0115241, 0.0115529, 0.0113372, 0.011395099999999998, 0.0113044, 0.0115068, 0.0124498, 0.0115059, 0.0115539, 0.0115369, 0.0116501, 0.0114775, 0.0112436, 0.0116285]
warm up [0.0115082, 0.0115965, 0.0111915, 0.0114787, 0.0113429, 0.0116133, 0.0115759, 0.0115294, 0.011281, 0.011419499999999999, 0.0112998, 0.011492500000000001, 0.011561499999999999, 0.0115756, 0.012422600000000002, 0.011502000000000002, 0.0112718, 0.01148, 0.0114769, 0.0121826]
warm up [0.0115257, 0.0115172, 0.0115641, 0.0115464, 0.0114497, 0.011382, 0.0115862, 0.0122358, 0.011521899999999998, 0.0113242, 0.0114885, 0.011396900000000001, 0.0115142, 0.011334, 0.0113848, 0.0114517, 0.0114698, 0.0113964, 0.0146482, 0.0115681]
warm up [0.0115086, 0.0116601, 0.0114141, 0.011340300000000001, 0.0114907, 0.0114594, 0.0114588, 0.011370499999999999, 0.011475899999999999, 0.0114323, 0.0123773, 0.0114826, 0.0127367, 0.011562900000000001, 0.011380800000000002, 0.0114696, 0.011489899999999999, 0.0114019, 0.0114173, 0.011320499999999999]
warm up [0.0115964, 0.0114328, 0.0114885, 0.011623399999999999, 0.0114162, 0.0115886, 0.0115008, 0.0114966, 0.0114544, 0.011373, 0.0115375, 0.011612, 0.011469299999999998, 0.0115104, 0.0115752, 0.011651799999999999, 0.0115388, 0.0115689, 0.011479, 0.0124621]
warm up [0.011322700000000002, 0.0112694, 0.012187300000000002, 0.0117535, 0.0143264, 0.0115169, 0.011315700000000001, 0.0113154, 0.0114203, 0.0114637, 0.011294100000000001, 0.011509599999999998, 0.011348899999999999, 0.012311700000000002, 0.0116084, 0.0114198, 0.011453, 0.0114603, 0.011722799999999998, 0.0113456]
warm up [0.011503399999999999, 0.011470000000000001, 0.0115983, 0.011604, 0.011559900000000001, 0.0113911, 0.0116183, 0.014966, 0.0115236, 0.0114546, 0.0114336, 0.011576000000000001, 0.0113359, 0.011425300000000001, 0.0113755, 0.0139666, 0.011534399999999998, 0.011616900000000001, 0.011325, 0.0113217]
warm up [0.0115041, 0.0113348, 0.011453399999999999, 0.0116374, 0.011352600000000001, 0.0114614, 0.0113532, 0.011317899999999999, 0.0115519, 0.014837399999999999, 0.011455000000000002, 0.011388899999999999, 0.0114763, 0.011356400000000001, 0.011335799999999998, 0.011438799999999999, 0.0113746, 0.011519099999999999, 0.011398499999999999, 0.011542]
tune []
No. 0 | The best currently 0.0111915 {'merge': 0, 'inline': 0}
re-evaluate [0.011395099999999998, 0.0115397, 0.011388200000000001, 0.011325, 0.0115052, 0.0114513, 0.012325199999999998, 0.011347300000000001, 0.0116181, 0.0127767]
warm up [0.0115194]
tune []
No. 1 | The best currently 0.0112671 {'merge': 3, 'inline': 0}
re-evaluate [0.0115996, 0.011560500000000001, 0.0116746, 0.0116094, 0.0113551, 0.011467300000000001, 0.0114994, 0.0115478, 0.011421500000000001, 0.0115082]
warm up [0.0113515]
tune []
No. 2 | The best currently 0.0112942 {'merge': 0, 'inline': 0}
re-evaluate [0.0115762, 0.011411500000000002, 0.0115168, 0.0115485, 0.0114256, 0.0115784, 0.011716899999999999, 0.0124064, 0.0162457, 0.0115833]
warm up [0.011349099999999999]
tune []
No. 3 | The best currently 0.0113217 {'merge': 3, 'inline': 0}
re-evaluate [0.011448700000000001, 0.0113547, 0.0115209, 0.0115766, 0.0114565, 0.0116364, 0.0114797, 0.0124907, 0.0113638, 0.0114947]
warm up [0.0114351]
tune []
No. 4 | The best currently 0.011334 {'merge': 2, 'inline': 0}
Early stop after continuous no trials 5 times
######################################
op schedules:
----------------------------------
spatial [[1, 1, 1, 2], [1, 1, 1, 4]]
reduce [[1, 1, 2]]
fuse [[1, 2, 2]]
unroll [[1500, 0]]
reorder [[1]]
----------------------------------
spatial [[1, 1, 2, 2], [1, 1, 2, 2]]
fuse [[1, 2, 2]]
unroll [[1, 0]]
reorder [[0]]
graph schedules:
merge [[0, 1]]
inline [[0, 0]]
block_circulant_matrix_block_circulant_matrix_(4, 4, 2)_cuda(0):[[{"spatial": [[1, 1, 1, 2], [1, 1, 1, 4]], "reduce": [[1, 1, 2]], "fuse": [[1, 2, 2]], "merge": [], "unroll": [[1500, 0]], "inline": [], "reorder": [[1]]}, {"spatial": [[1, 1, 2, 2], [1, 1, 2, 2]], "reduce": [], "fuse": [[1, 2, 2]], "merge": [], "unroll": [[1, 0]], "inline": [], "reorder": [[0]]}], {"spatial": [], "reduce": [], "fuse": [], "merge": [[0, 1]], "unroll": [], "inline": [[0, 0]], "reorder": []}]
Use 0.011493 ms
Cost 3911.8264739513397 s
commented