# Introduction ![Running SLMs in web browsers](thumb_small_language_model.jpg) This repository is part of [playbook for experiments on fine tuning small language models](https://ashishware.com/2025/11/16/slm_in_browser/) using LoRA, exporting them to ONNX and running them locally using ONNX compatibale runtime like javascript(node js) and WASM (browser) ### Before you start - Clone the repository https://github.com/code2k13/onnx_javascript_browser_inference/edit/main/README.md - From all files from this repository to the `model_files` directory of the cloned github repository. - Run `npm install` ### To run NodeJS example (NodeJS + onnx-runtime, server side) - Simple run `node app.js` This is what you should see ![NodeJS application showing paraphrasing screen](slm_nodejs.gif) ### To run web browser based demo (WASM based in-browser inference) - Simply access `web.html` from a local server (example `http://localhost:3000/web.html`) This is what you should see ![NodeJS application showing paraphrasing screen](slm_web_wasm.gif) ### Citation ``` @misc{allal2024SmolLM, title={SmolLM - blazingly fast and remarkably powerful}, author={Loubna Ben Allal and Anton Lozhkov and Elie Bakouch and Leandro von Werra and Thomas Wolf}, year={2024}, } ```