mirror of https://github.com/commaai/tinygrad.git
74 lines
2.3 KiB
Python
74 lines
2.3 KiB
Python
#!/usr/bin/env python
|
|
import numpy as np
|
|
import unittest
|
|
from tinygrad.lazy import LazyBuffer
|
|
from tinygrad import Device
|
|
from tinygrad.tensor import Tensor
|
|
from tinygrad.jit import CacheCollector
|
|
|
|
class TestLazyBuffer(unittest.TestCase):
|
|
@unittest.skip("it doesn't work like this anymore")
|
|
def test_fromcpu_buffer_sharing(self):
|
|
a = np.arange(8)
|
|
assert LazyBuffer.fromCPU(a).realized._buf is a
|
|
|
|
def test_fromcpu_shape_tracker(self):
|
|
def helper(a: np.ndarray):
|
|
print(a.shape, a.strides, a.flags.c_contiguous)
|
|
b = LazyBuffer.fromCPU(a)
|
|
#assert b.st.contiguous == a.flags.c_contiguous
|
|
assert b.st.shape == a.shape
|
|
np.testing.assert_equal(a, Tensor(b).numpy())
|
|
|
|
for ndims in range(1, 4):
|
|
a = np.random.randn(*(4,)*ndims).astype(np.float32)
|
|
for stride in [-2, 1, 2]:
|
|
for start in [0, 1]:
|
|
helper(a[(slice(start, None, stride),)*ndims])
|
|
|
|
def test_shuffle_pad_ops_cmpeq(self):
|
|
y = Tensor([1]).cat(Tensor([1]) == 0).numpy()
|
|
z = Tensor([1, 0]).numpy()
|
|
np.testing.assert_allclose(y, z)
|
|
|
|
def test_shuffle_pad_ops_div(self):
|
|
y = Tensor([1]).cat(Tensor([1]).div(Tensor([2.0]))).numpy()
|
|
z = Tensor([1, 0.5]).numpy()
|
|
np.testing.assert_allclose(y, z)
|
|
|
|
def test_shuffle_pad_ops_log(self):
|
|
y = Tensor([1]).cat(Tensor([1]).log()).numpy()
|
|
z = Tensor([1, 0]).numpy()
|
|
np.testing.assert_allclose(y, z)
|
|
|
|
def test_shuffle_pad_ops_exp(self):
|
|
y = Tensor([1]).cat(Tensor([1]).exp()).numpy()
|
|
z = Tensor([1, np.e]).numpy()
|
|
np.testing.assert_allclose(y, z)
|
|
|
|
@unittest.skipUnless(Device.DEFAULT in ["METAL", "CUDA", "GPU"], "Only GPU backends supports cache")
|
|
def test_children_count(self):
|
|
a = Tensor.ones(8,8,8)
|
|
d1 = a.sum((0))
|
|
d2 = a.sum((0)).reshape(32,2) # noqa: F841
|
|
assert len(d1.lazydata.op.src[0].children) == 1
|
|
in1 = d1.reshape(16,4)
|
|
d3 = in1.reshape(8,8)
|
|
assert len(d3.lazydata.op.src[0].children) == 2
|
|
|
|
CacheCollector.start()
|
|
l = Tensor.ones(8,8)
|
|
r = Tensor.ones(8,8)
|
|
dd = d1 + l
|
|
dd.realize()
|
|
de = d3 + r
|
|
de.realize()
|
|
cache = CacheCollector.finish()
|
|
assert len(cache) == 3
|
|
assert cache[0].prg.name.startswith("r_") # Reduce should not merged 2 times.
|
|
assert cache[1].prg.name.startswith("E_")
|
|
assert cache[2].prg.name.startswith("E_")
|
|
|
|
if __name__ == "__main__":
|
|
unittest.main()
|