picoLLM Inference Engine
Web Quick Start
Platforms
- Chrome & Chromium-based browsers
- Edge
- Firefox
- Safari
Requirements
- Picovoice Account and AccessKey
- Node.js 16+
- npm
Picovoice Account & AccessKey
Signup or Login to Picovoice Console to get your AccessKey.
Make sure to keep your AccessKey secret.
Quick Start
Setup
Install Node.js.
Install the picoLLM Web package:
- Download a
picoLLMmodel file (.pllm) from Picovoice Console.
Usage
- Either create an HTML input tag that accepts the
.pllmmodel file:
or put the model file in a web server or public directory:
- Create a
picoLLMinstance usingPicoLLMWorkerand the model file from above:
Replace ${ACCESS_KEY} with yours obtained from Picovoice Console.
- Generate prompt completion:
- To interrupt completion generation before it has finished:
- Release resources explicitly when done with
picoLLM:
Vision models
To run a VLM such as qwen3-vl-2b-it:
Replace ${PROMPT} with a text prompt. For the image, you will need to get image height and width in number of pixels and the raw pixel values of the image in 8-bit, RGB format.
OCR models
To run an OCR model such as deepseek-ocr-2:
For the image, you will need to get image height and width in number of pixels and the raw pixel values of the image in 8-bit, RGB format.
Embedding models
To run an embedding model such as embeddinggemma-300m:
Replace ${PROMPT} with a text prompt that you want to generate embeddings for.
Demo
For the picoLLM Web SDK, we offer demo applications that demonstrate how to use it to generate text from a prompt or in a chat-based environment.
Setup
Clone the picoLLM repository from GitHub:
Usage
- Install dependencies and run:
- Open http://localhost:5000 to view it in the browser.
For more information on our picoLLM demos for Web or to see a chat-based demo, head over to our GitHub repository.