model refuses to load server side within transformers.js and struggles hard in the browser
#24
by
shoebill-droyd
- opened
its a pretty decent model considering its only 1.7B parameters, however i cannot get it working in nodejs for the life of me, and it has HUGE load times within the browser. why even go through the bother of marking it as transformers.js compatible in this broken state????? just seems misleading to me
Hi there :) You can use a lower quantization (e.g., q4f16 or q4) depending on your needs, which runs at ~80 tokens per second in-browser on an M4 Pro Max.
For Node.js support, we've added external data support in Transformers.js v3.4, which means you can load the larger models if you'd like.