Tag:
Branch:
Tree:
889bcc73c2
__nightly
acodonomo
aggressivepilot
audible-cancel
auto-package-updates
axmodel
badges
beeps
bigmodeld
biiiiig_model
bolt-lower-torque
burn
capnpy
commatwo_master
devel-staging
fakecam
fbw_model
gwm-driving
honda-altradar
honda-canfd
honda-ci-test
honda-notebook
honda-testing
label-class
lagd-window-size
lilffmpeg
lkas_feedback_animation
locationd-standstill
log-compat-test
log-dashcam-reason
lsm_low_pass
master
master-tici
minplan
minplan2
modeld-misc-support
modeld-misc-support-secretgood
nevada-model
nightly
nightly-dev
no-block
op-diff
package-tinygrad
pr-setup-speedup
pylogger
pypanda
recorded-feedback
recorded-feedback-fixed-timeout
release-tici
release-tizi
release2
release3
release3-staging
rerun-fix
revert-fan
rl-scroller
rl-test
rmkaitai
rmqt
runbiiiig
scenarios
secretgoodopenpilot
secretgoodopenpilot-big
sshane-patch-1
sync
taco
taco2
test-tx-msgs
testing-closet
tmp-jenkins-35405
tmp-jenkins-35820
tmp-jenkins-35821
tmp-jenkins-35908
toyota-ff-pitch
tr10
tr13
tr16
tr_for_cookiemonster
tst-odbc-sl
use-non-delay-action
uv_model_v2
wakeword-clip
wm_model
zeroten
v0.1
v0.10.0
v0.2
v0.2.1
v0.2.2
v0.2.3
v0.2.4
v0.2.5
v0.2.6
v0.2.7
v0.2.8
v0.2.9
v0.3.0
v0.3.1
v0.3.2
v0.3.3
v0.3.4
v0.3.5
v0.4.0.2
v0.4.1
v0.4.2
v0.4.4
v0.4.5
v0.4.6
v0.4.7
v0.5
v0.5.1
v0.5.10
v0.5.11
v0.5.12
v0.5.13
v0.5.2
v0.5.3
v0.5.4
v0.5.5
v0.5.6
v0.5.7
v0.5.8
v0.5.9
v0.6
v0.6.1
v0.6.2
v0.6.3
v0.6.4
v0.6.5
v0.6.6
v0.7
v0.7.1
v0.7.10
v0.7.2
v0.7.3
v0.7.4
v0.7.5
v0.7.6.1
v0.7.7
v0.7.8
v0.7.9
v0.8
v0.8.1
v0.8.10
v0.8.11
v0.8.12
v0.8.13
v0.8.14
v0.8.15
v0.8.16
v0.8.2
v0.8.3
v0.8.4
v0.8.5
v0.8.6
v0.8.7
v0.8.8
v0.8.9
v0.9.0
v0.9.1
v0.9.2
v0.9.3
v0.9.4
v0.9.5
v0.9.6
v0.9.7
v0.9.8
v0.9.9
${ noResults }
3 Commits (889bcc73c23f95e102fdc97dd9ebf4597f46ca6d)
Author | SHA1 | Message | Date |
---|---|---|---|
|
ab83e48ec4
|
Add pre-commit hooks (#1629)
|
5 years ago |
|
78a352a8ca
|
This isn't bigmodel, but there's a lot of good stuff here (#1532)
* bigmodel * more debug print * debugging bigmodel * remove the tanh, debugging * print images/buffers * disassemble the command queues * decompiler * dump the shaders * full disasm * support patching kernel and fixing convolution_horizontal_reduced_reads_1x1 * microbenchmark * 42 GFLOPS, 1 GB/s * gemm benchmark * 75 GFLOPS vs 42 GFLOPS * 115 GFLOPS * oops, never mind * gemm image is slow * this is pretty hopeless * gemm image gets 62 GFLOPS * this is addictive and still a waste of time * cleanup cleanup * that hook was dumb * tabbing * more tabbing Co-authored-by: Comma Device <device@comma.ai> |
5 years ago |
|
302d06ee70 |
thneed saves 45% of a core (#1512)
* thneed runs the model * thneed is doing the hooking * set kernel args * thneeding the bufferS * print the images well * thneeds with better buffers * includes * disasm adreno * parse packets * disasm works * disasm better * more thneeding * much thneeding * much more thneeding * thneed works i think * thneed is patient * thneed works * 7.7% * gpuobj sync * yay, it mallocs now * cleaning it up, Thneed * sync objs and set power * thneed needs inputs and outputs * thneed in modeld * special modeld runs * can't thneed the DSP * test is weird * thneed modeld uses 6.4% CPU * add thneed to release * move to debug * delete some junk from the pr * always track the timestamp * timestamp hacks in thneed * create a new command queue * fix timestamp * pretty much back to what we had, you can't use SNPE with thneed * improve thneed test * disable save log Co-authored-by: Comma Device <device@comma.ai> |
5 years ago |