AI stack built on Node.js and powered by RAG Retrieval-Augmented Generation
The below table is describing the different layer abstractions
| Layer | Node.js Tool/Lib | Notes |
|---|---|---|
| Document loader | fs |
- |
| Document Parser | langchain |
File format supported txt |
| RAG Framework | langchain |
Chain together: Retrieval + Generation |
| RAG Embeddings | Xenova/all-MiniLM-L6-v2 |
Fast 384-dim sentence embeddings |
| LLM | EleutherAI/gpt-neo-125M |
Lightweight Model |
Initally based on chatGPT discussion, 27-06-2025
The current instance is running a generative chat supplying basic response on a specific theme
- Python
3.9, later version are not yet covered by torch prebuilt - Nodejs version >=
18.0(not tested) - English documentation is only supported by LLM
Perform global installation
Get a copy of the current project locally git clone https://github.com/johnvanderton/aijstack then go to this folder and perform the installation script npm run installation
Note, py-env_install is creating a .venv virtual environment and imports the following packages,
- fastapi
- uvicorn
- torch
- transformers
- accelerate
- pydantic
- Dispose your documentation into the
/docfolder. Note: Currently only 'txt' files are supported - Run both
nodejsandmodelscript instances withquick-startornpm run start-dev:all - Send through HTTP POST a JSON message to
http://localhost:3000/generatei.e : {"query" : "What is (your scope) ?"} - Expecting for a HTTP '201' JSON message response (json anwser response : Context... > Question... > Answer...)