George Hotz
7e191fbb86
hotfix: don't jitcache with 1 kernel. improvements to hip sniffer
2024-01-03 19:17:08 -08:00
George Hotz
bcc1aa21ac
make disk simpler ( #3002 )
...
* make disk simpler
* upd ops_disk
* works on osx too
* revert ops_hip
2024-01-03 17:46:21 -08:00
George Hotz
9699c8c90b
don't alloc for InterpretedASTRunner ( #2999 )
2024-01-03 17:05:53 -08:00
chenyu
bca0b95ee3
bump shapetracker simplify message to DEBUG >= 5 ( #2998 )
2024-01-03 20:00:36 -05:00
chenyu
74a30431b4
replace `d[a] if a in d else b` with `d.get(a, b)` ( #2997 )
2024-01-03 18:10:25 -05:00
chenyu
74cc6fd3c2
remove AndNode.__floordiv__ special case ( #2996 )
...
* remove AndNode.__floordiv__
AndNode produces a Node that min/max is bounded by [0, 1] so `//` on top of that is almost always 0.
we don't really use that either
* keep the test
2024-01-03 17:44:55 -05:00
George Hotz
a0c7cb2564
hotfix: create weights dir in local tg checkout
2024-01-03 14:14:33 -08:00
George Hotz
fc36a7d669
tinygrad weights
2024-01-03 14:09:28 -08:00
chenyu
1ac4d27869
remove VariableOrNum from Node.substitute arg ( #2995 )
...
having NumNode in var_vals does not change the substitute output
2024-01-03 17:02:25 -05:00
George Hotz
65dc3700b7
hip device is default on supported platforms ( #2993 )
2024-01-03 13:42:13 -08:00
George Hotz
77c98a1543
hotfix: remove weights directory
2024-01-03 13:40:39 -08:00
George Hotz
0be0f2f745
remove stable diffusion test on tinymac
2024-01-03 13:18:24 -08:00
George Hotz
a354ec9dad
Revert "hotfix: HIP is the default device on HIP platforms"
...
This reverts commit b748b569f5
.
2024-01-03 13:16:54 -08:00
George Hotz
b748b569f5
hotfix: HIP is the default device on HIP platforms
2024-01-03 13:13:52 -08:00
George Hotz
753a7ecc05
Hip driver ( #2992 )
...
* start hip driver
* fix hip llama
* make HIP default if we can
* don't change those
2024-01-03 12:53:47 -08:00
George Hotz
f290ca3924
hotfix: save lines in graph
2024-01-03 12:03:42 -08:00
Yixiang Gao
bc4b6e758b
Merge pull request #2981 from g1y5x3/cifar_fp16
...
adjsut div factor to avoid underflow for cifar in fp16
2024-01-03 11:15:42 -08:00
George Hotz
d7d5a487ad
hotfix: all device canonicalize should be done in Tensor
2024-01-03 10:48:04 -08:00
Yixiang Gao
ea3bc2f509
remove wino benchmark for now
2024-01-03 10:46:43 -08:00
Yixiang Gao
5663dd46b6
Merge branch 'master' of github.com:tinygrad/tinygrad into cifar_fp16
2024-01-03 10:11:46 -08:00
chenyu
81b97cd2c6
canonicalize device in LazyBuffer constructor ( #2991 )
...
fixed the multitensor +1 then sum bug
2024-01-03 12:55:25 -05:00
chenyu
db525cf8c2
multitensor failed test case with +1 then sum on DEVICE:0 ( #2990 )
2024-01-03 12:17:11 -05:00
Yixiang Gao
7f1802cd50
update benchmark
2024-01-03 09:09:34 -08:00
George Hotz
5dbaaa7061
hotfix: make multitensor shard contiguous
2024-01-03 08:48:30 -08:00
chenyu
590268fa03
out_tokens -> grouped in linearizer ( #2989 )
...
no more token now
2024-01-03 11:45:28 -05:00
Yixiang Gao
8e1fd6ae9d
test works
2024-01-03 07:22:01 -08:00
Yixiang Gao
4f89f8b73a
make sure the old hyp breaks the test
2024-01-03 07:13:54 -08:00
Yixiang Gao
84eb6dd32a
skip GPU cause opencl on intel can't compile half
2024-01-03 07:07:21 -08:00
Yixiang Gao
73879b50ad
only need to check the min_lr for the nan bug
2024-01-03 07:00:50 -08:00
Yixiang Gao
99f8740c60
running half in CI CPU is slow
2024-01-02 18:44:35 -08:00
Yixiang Gao
781690fd99
how long it takes on CI CPU without the lr scheduler
2024-01-02 18:33:48 -08:00
Yixiang Gao
dd00bcb9c0
fix whitespace
2024-01-02 18:16:33 -08:00
Yixiang Gao
841487cad9
add half test with using hyp from benchmarks
2024-01-02 18:14:30 -08:00
George Hotz
f494b9d463
simple multitensor API ( #2903 )
...
* simple multitensor API
* test multitensor
* mt work
* new api
* copies
* all but data parallel
* allreduce there
* works, but axis sharded
* fix all mt tests
* features/multi
* work
* backprop
* fix tests
* tests passing
* mt progress
* cleanups
* less lines
* tensor cleanup
* save more lines
* mypy passes
* fix tests
* skip for cuda too
* bump download cache
2024-01-02 17:49:44 -08:00
George Hotz
5522ba234b
simplify image functions ( #2987 )
...
* simplify image functions
* line in tensor
2024-01-02 17:35:08 -08:00
chenyu
6e9406c986
one list comprehension in search action ( #2988 )
...
instead of list of list then flatten
2024-01-02 20:29:26 -05:00
chenyu
08a34faea8
pass tuple for strs to startswith ( #2986 )
2024-01-02 19:51:15 -05:00
George Hotz
dbe4a1a914
switch CI to tiny8 ( #2984 )
...
* switch CI to tiny8
* no copyin for disk
* Revert "no copyin for disk"
This reverts commit eb46b7e93da4a650d8125020c38f44d1f8f2c86e.
* rocm 6 broke llama
* rename it
2024-01-02 16:40:25 -08:00
Yixiang Gao
b753d280f7
move hyp out of the train so it can be imported
2024-01-02 15:56:17 -08:00
chenyu
0dd3ca59cd
simpler ModNode.__mod__ and ModNode.__floordiv__ ( #2983 )
...
`gcd(self.b, b) == b` is equivalent to `self.b % b == 0`.
use the same condition and format in __floordiv__ too.
2024-01-02 18:52:42 -05:00
chenyu
c07907e644
grad -> grad_output in mlops for consistency ( #2982 )
2024-01-02 18:03:55 -05:00
Yixiang Gao
54cdba57e7
mend
2024-01-02 14:21:06 -08:00
Yixiang Gao
26303d181b
re-enable half cifar benchmarks
2024-01-02 14:16:35 -08:00
Yixiang Gao
2e4d9ad936
adjsut div factor to avoid underflow
2024-01-02 13:47:13 -08:00
chenyu
ad0d710ec4
merge apply_opt OptOps.LOCAL and OptOps.LASTLOCAL into one block ( #2980 )
...
and other minor apply_opt cleanups
2024-01-02 16:40:10 -05:00
George Hotz
8de160d08e
hotfix: remove dead code, save lines
2024-01-02 12:52:20 -08:00
chenyu
878e869663
simpler SumNode.__mod__ ( #2979 )
...
* simpler SumNode.__mod__
delegate simplification to individual node
* ModNode.__mod__ simplification case
* Revert "ModNode.__mod__ simplification case"
This reverts commit 73a42205a8b81dc56e0c517db49ede9f1b7c8fdd.
2024-01-02 15:09:15 -05:00
chenyu
91ddda244f
minor cleanups in dtype.py ( #2978 )
...
* minor cleanups in dtype.py
* all not
2024-01-02 13:42:37 -05:00
chenyu
ff5399f053
move one last dtype test from test_helpers to test_dtype ( #2975 )
2024-01-02 12:37:56 -05:00
qazal
deb3722aac
refactor workitems ( #2973 )
2024-01-02 09:16:52 -08:00