George Hotz
c594a0a835
fix flip bug, add new unit tests
2023-03-12 23:55:31 -07:00
George Hotz
a4abcf0969
improve test_example
2023-03-12 22:59:40 -07:00
George Hotz
5577634cf3
tests in pre commit
2023-03-12 22:42:26 -07:00
George Hotz
ce1564b05e
fix shapetracker test
2023-03-12 22:33:25 -07:00
George Hotz
153cce0f7e
tutorial
2023-03-12 22:31:46 -07:00
George Hotz
8d16ebaea7
we have docs:
2023-03-12 19:05:44 -07:00
George Hotz
b512edc9ff
no decorators for image methods. move out RawMallocBuffer. -7 lines
2023-03-12 16:28:45 -07:00
George Hotz
ed9ab6ff03
move image to nn/image.py
2023-03-12 16:21:42 -07:00
George Hotz
fe0e8a306f
jittable llama
2023-03-12 14:15:04 -07:00
George Hotz
dcac618515
stop wasting time with the compiler. tinygrad needs to just jit
2023-03-12 12:08:46 -07:00
George Hotz
46b49d50bd
llvm was using wrong shapetracker
2023-03-12 11:49:03 -07:00
George Hotz
fdde87afda
Revert "Revert "late simplify on st""
...
This reverts commit c8508e359d
.
2023-03-12 11:47:44 -07:00
George Hotz
c8508e359d
Revert "late simplify on st"
...
This reverts commit 606550474c
.
2023-03-12 11:46:10 -07:00
George Hotz
606550474c
late simplify on st
2023-03-12 11:38:56 -07:00
George Hotz
de6f1695a3
only allow exact buffer name
2023-03-12 11:13:36 -07:00
George Hotz
15e0b56e39
compile works ( #688 )
...
* compile works
* runtimes
* line count
* fix custom, to tg dtype
* meh, that's fine with lazy import
2023-03-12 11:01:25 -07:00
Kirill
af7745073f
Add comments to SD ( #686 )
...
* Add explanation for empty lambdas
* Fix my_unpickle if pytorch_lightning is installed
* oops
2023-03-12 10:56:49 -07:00
George Hotz
58d3824cbe
better get_state_dict
2023-03-12 00:10:48 -08:00
George Hotz
046b3952c3
get_state_dict
2023-03-11 23:46:53 -08:00
George Hotz
6c3675c01c
_mmap loads to gpu fast
2023-03-11 23:00:13 -08:00
George Hotz
dc9a6b4bb7
fix float16 in CLANG on linux
2023-03-11 21:51:22 -08:00
George Hotz
803b0aef28
track memory for numpy/torch
2023-03-11 20:39:10 -08:00
George Hotz
37cf6fc4c0
err, external_test_opt.py broke...fusing will have to wait. correctness over speed
2023-03-11 17:54:47 -08:00
George Hotz
305b9f2d21
multistep optim tests passing
2023-03-11 17:49:53 -08:00
George Hotz
61071f881a
fix bug, and add unit test to catch failure
2023-03-11 16:57:25 -08:00
George Hotz
3ec457248c
failing llama test
2023-03-11 16:28:10 -08:00
George Hotz
8aa63847c7
llama: up max tokens to 1000
2023-03-11 13:39:33 -08:00
George Hotz
d30005b645
shapetracker touchups
2023-03-11 12:07:52 -08:00
George Hotz
d41ac5f5f1
touchups
2023-03-11 11:41:34 -08:00
Cyril Roumégous
3f08613a2a
apply flake8 E203 rule ( #684 )
2023-03-11 11:35:16 -08:00
Diogo
784afc6c6f
Eq magic function support ( #683 )
...
* add eq magic func
* changed from eq to __eq__
* ignore type for linter
* mypy doenst like descriptions :(
2023-03-11 10:31:46 -08:00
George Hotz
5ea44cefcc
llama: add lexie personality
2023-03-11 10:23:33 -08:00
George Hotz
c908f911a7
llama defaults to metal on osx
2023-03-11 09:30:13 -08:00
George Hotz
fd65edf595
fix mem_estimate for dtype.itemsize
2023-03-11 09:20:05 -08:00
George Hotz
fe8c05b96f
allow disabling method cache
2023-03-11 08:57:49 -08:00
George Hotz
5e1380df6a
profiling llama + cache is_contiguous
2023-03-11 08:23:21 -08:00
George Hotz
01f39b19dc
move to shapetracker.py
2023-03-11 07:50:07 -08:00
George Hotz
f3ac52aee8
Mypyc ( #680 )
...
* building shapetracker
* default ENABLE_METHOD_CACHE
* symbolic compiles
* improve types
* tensor compiles
* oops, that's a bug
* best of both worlds
* find legit typing bugs
* pad2d can take list or tuple
* sub 200ms when compiled
2023-03-11 07:33:30 -08:00
George Hotz
22905dd657
speedups from llama branch
2023-03-10 22:01:32 -08:00
George Hotz
0b03216cc3
losing lines ( #678 )
...
* losing lines
* FLIP -> STRIDE
* shapetracker refactor
2023-03-10 21:57:05 -08:00
George Hotz
d7cb8e3e56
multithreaded fake_torch_load_zipped
2023-03-10 19:16:27 -08:00
George Hotz
b1206bcb18
third try at torch loading ( #677 )
...
* third try at torch loading
* numpy fixed
* fix enet compile
* load_single_weight supports empty weights
* oops, CPU wasn't the default
* so many bugs
2023-03-10 19:11:29 -08:00
Connor Henderson
8b7a16cf85
Add conv binops_no_rerun test assertions ( #665 )
...
* Add conv binops_no_rerun assertions
* use assert_allclose
* widen tolerance for elu
2023-03-10 19:09:48 -08:00
George Hotz
8bf75a7fdd
fix stable diffusion and CI
2023-03-10 17:48:12 -08:00
George Hotz
c7d17c25d9
ugh, that's getting ugle
2023-03-10 17:41:19 -08:00
George Hotz
4780f9a6df
llama runs (slowly) in master
2023-03-10 17:36:51 -08:00
George Hotz
1826ff6b89
dtypes nice and clean ( #673 )
...
* add dtype class
* dtypes
* buffers are lazy
* dtype is tracked by lazybuffer and GenericShape
* fix types in llvm
* llvm store
* dtype tests
* fix tests maybe
* fix flop counter
* fix CI
* CI fix and check format
* fix dtype and dtype check
* fix custom test
* fix test graph
2023-03-10 16:56:07 -08:00
George Hotz
d26345595d
more llama stuff
2023-03-10 10:48:10 -08:00
George Hotz
442e1bcd5a
typo + EARLY_STOPPING
2023-03-10 10:43:07 -08:00
George Hotz
6142e63a3e
touchups, print GB/s
2023-03-10 10:37:37 -08:00