Commit Graph

3282 Commits

Author SHA1 Message Date
George Hotz 7e191fbb86 hotfix: don't jitcache with 1 kernel. improvements to hip sniffer 2024-01-03 19:17:08 -08:00
George Hotz bcc1aa21ac
make disk simpler (#3002)
* make disk simpler

* upd ops_disk

* works on osx too

* revert ops_hip
2024-01-03 17:46:21 -08:00
George Hotz 9699c8c90b
don't alloc for InterpretedASTRunner (#2999) 2024-01-03 17:05:53 -08:00
chenyu bca0b95ee3
bump shapetracker simplify message to DEBUG >= 5 (#2998) 2024-01-03 20:00:36 -05:00
chenyu 74a30431b4
replace `d[a] if a in d else b` with `d.get(a, b)` (#2997) 2024-01-03 18:10:25 -05:00
chenyu 74cc6fd3c2
remove AndNode.__floordiv__ special case (#2996)
* remove AndNode.__floordiv__

AndNode produces a Node that min/max is bounded by [0, 1] so `//` on top of that is almost always 0.
we don't really use that either

* keep the test
2024-01-03 17:44:55 -05:00
George Hotz a0c7cb2564 hotfix: create weights dir in local tg checkout 2024-01-03 14:14:33 -08:00
George Hotz fc36a7d669 tinygrad weights 2024-01-03 14:09:28 -08:00
chenyu 1ac4d27869
remove VariableOrNum from Node.substitute arg (#2995)
having NumNode in var_vals does not change the substitute output
2024-01-03 17:02:25 -05:00
George Hotz 65dc3700b7
hip device is default on supported platforms (#2993) 2024-01-03 13:42:13 -08:00
George Hotz 77c98a1543 hotfix: remove weights directory 2024-01-03 13:40:39 -08:00
George Hotz 0be0f2f745 remove stable diffusion test on tinymac 2024-01-03 13:18:24 -08:00
George Hotz a354ec9dad Revert "hotfix: HIP is the default device on HIP platforms"
This reverts commit b748b569f5.
2024-01-03 13:16:54 -08:00
George Hotz b748b569f5 hotfix: HIP is the default device on HIP platforms 2024-01-03 13:13:52 -08:00
George Hotz 753a7ecc05
Hip driver (#2992)
* start hip driver

* fix hip llama

* make HIP default if we can

* don't change those
2024-01-03 12:53:47 -08:00
George Hotz f290ca3924 hotfix: save lines in graph 2024-01-03 12:03:42 -08:00
Yixiang Gao bc4b6e758b
Merge pull request #2981 from g1y5x3/cifar_fp16
adjsut div factor to avoid underflow for cifar in fp16
2024-01-03 11:15:42 -08:00
George Hotz d7d5a487ad hotfix: all device canonicalize should be done in Tensor 2024-01-03 10:48:04 -08:00
Yixiang Gao ea3bc2f509 remove wino benchmark for now 2024-01-03 10:46:43 -08:00
Yixiang Gao 5663dd46b6 Merge branch 'master' of github.com:tinygrad/tinygrad into cifar_fp16 2024-01-03 10:11:46 -08:00
chenyu 81b97cd2c6
canonicalize device in LazyBuffer constructor (#2991)
fixed the multitensor +1 then sum bug
2024-01-03 12:55:25 -05:00
chenyu db525cf8c2
multitensor failed test case with +1 then sum on DEVICE:0 (#2990) 2024-01-03 12:17:11 -05:00
Yixiang Gao 7f1802cd50 update benchmark 2024-01-03 09:09:34 -08:00
George Hotz 5dbaaa7061 hotfix: make multitensor shard contiguous 2024-01-03 08:48:30 -08:00
chenyu 590268fa03
out_tokens -> grouped in linearizer (#2989)
no more token now
2024-01-03 11:45:28 -05:00
Yixiang Gao 8e1fd6ae9d test works 2024-01-03 07:22:01 -08:00
Yixiang Gao 4f89f8b73a make sure the old hyp breaks the test 2024-01-03 07:13:54 -08:00
Yixiang Gao 84eb6dd32a skip GPU cause opencl on intel can't compile half 2024-01-03 07:07:21 -08:00
Yixiang Gao 73879b50ad only need to check the min_lr for the nan bug 2024-01-03 07:00:50 -08:00
Yixiang Gao 99f8740c60 running half in CI CPU is slow 2024-01-02 18:44:35 -08:00
Yixiang Gao 781690fd99 how long it takes on CI CPU without the lr scheduler 2024-01-02 18:33:48 -08:00
Yixiang Gao dd00bcb9c0 fix whitespace 2024-01-02 18:16:33 -08:00
Yixiang Gao 841487cad9 add half test with using hyp from benchmarks 2024-01-02 18:14:30 -08:00
George Hotz f494b9d463
simple multitensor API (#2903)
* simple multitensor API

* test multitensor

* mt work

* new api

* copies

* all but data parallel

* allreduce there

* works, but axis sharded

* fix all mt tests

* features/multi

* work

* backprop

* fix tests

* tests passing

* mt progress

* cleanups

* less lines

* tensor cleanup

* save more lines

* mypy passes

* fix tests

* skip for cuda too

* bump download cache
2024-01-02 17:49:44 -08:00
George Hotz 5522ba234b
simplify image functions (#2987)
* simplify image functions

* line in tensor
2024-01-02 17:35:08 -08:00
chenyu 6e9406c986
one list comprehension in search action (#2988)
instead of list of list then flatten
2024-01-02 20:29:26 -05:00
chenyu 08a34faea8
pass tuple for strs to startswith (#2986) 2024-01-02 19:51:15 -05:00
George Hotz dbe4a1a914
switch CI to tiny8 (#2984)
* switch CI to tiny8

* no copyin for disk

* Revert "no copyin for disk"

This reverts commit eb46b7e93da4a650d8125020c38f44d1f8f2c86e.

* rocm 6 broke llama

* rename it
2024-01-02 16:40:25 -08:00
Yixiang Gao b753d280f7 move hyp out of the train so it can be imported 2024-01-02 15:56:17 -08:00
chenyu 0dd3ca59cd
simpler ModNode.__mod__ and ModNode.__floordiv__ (#2983)
`gcd(self.b, b) == b` is equivalent to `self.b % b == 0`.
use the same condition and format in __floordiv__ too.
2024-01-02 18:52:42 -05:00
chenyu c07907e644
grad -> grad_output in mlops for consistency (#2982) 2024-01-02 18:03:55 -05:00
Yixiang Gao 54cdba57e7 mend 2024-01-02 14:21:06 -08:00
Yixiang Gao 26303d181b re-enable half cifar benchmarks 2024-01-02 14:16:35 -08:00
Yixiang Gao 2e4d9ad936 adjsut div factor to avoid underflow 2024-01-02 13:47:13 -08:00
chenyu ad0d710ec4
merge apply_opt OptOps.LOCAL and OptOps.LASTLOCAL into one block (#2980)
and other minor apply_opt cleanups
2024-01-02 16:40:10 -05:00
George Hotz 8de160d08e hotfix: remove dead code, save lines 2024-01-02 12:52:20 -08:00
chenyu 878e869663
simpler SumNode.__mod__ (#2979)
* simpler SumNode.__mod__

delegate simplification to individual node

* ModNode.__mod__ simplification case

* Revert "ModNode.__mod__ simplification case"

This reverts commit 73a42205a8b81dc56e0c517db49ede9f1b7c8fdd.
2024-01-02 15:09:15 -05:00
chenyu 91ddda244f
minor cleanups in dtype.py (#2978)
* minor cleanups in dtype.py

* all not
2024-01-02 13:42:37 -05:00
chenyu ff5399f053
move one last dtype test from test_helpers to test_dtype (#2975) 2024-01-02 12:37:56 -05:00
qazal deb3722aac
refactor workitems (#2973) 2024-01-02 09:16:52 -08:00