Skip to content

Commit

Permalink
[WIP] Updating Extensions & Engines guide
Browse files Browse the repository at this point in the history
  • Loading branch information
imtuyethan committed Jan 7, 2025
1 parent 867a51c commit a90287b
Show file tree
Hide file tree
Showing 11 changed files with 449 additions and 346 deletions.
Binary file added docs/src/pages/docs/_assets/extensions-01.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
10 changes: 6 additions & 4 deletions docs/src/pages/docs/_meta.json
Original file line number Diff line number Diff line change
Expand Up @@ -24,17 +24,19 @@
"threads": "Threads",
"settings": "Settings",
"inference-engines": {
"title": "MODEL PROVIDER",
"title": "ENGINES",
"type": "separator"
},
"built-in": "Built-in Models",
"remote-models": "Remote APIs",
"built-in": "Local Engines",
"remote-models": "Remote Engines",
"extensions-separator": {
"title": "EXTENSIONS",
"type": "separator"
},
"extensions": "Overview",
"installing-extension": "Installing an Extension",
"extensions-settings": "Extensions Settings",
"installing-extension": "Install Extension",
"configure-extensions": "Configure Extensions",
"troubleshooting-separator": {
"title": "TROUBLESHOOTING",
"type": "separator"
Expand Down
3 changes: 0 additions & 3 deletions docs/src/pages/docs/built-in/tensorrt-llm.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -42,9 +42,6 @@ This guide walks you through installing Jan's official [TensorRT-LLM Engine](htt
### Step 1: Install TensorRT-Extension

1. Click the **Gear Icon (⚙️)** on the bottom left of your screen.
<br/>
![Settings](../_assets/settings.png)
<br/>
2. Select the **TensorRT-LLM** under the **Model Provider** section.
<br/>
![Click Tensor](../_assets/tensor.png)
Expand Down
268 changes: 268 additions & 0 deletions docs/src/pages/docs/configure-extensions.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,268 @@
---
title: Configure Extensions
description: Learn about Jan's default extensions and explore how to configure them.
[
Jan,
Customizable Intelligence, LLM,
local AI,
privacy focus,
free and open source,
private and offline,
conversational AI,
no-subscription fee,
large language models,
Jan Extensions,
Extensions,
]
---

# Configure an Extension Settings

To configure an extension settings:

1. Navigate to the `~/jan/data/extensions`.
2. Open the `extensions.json` file
3. Edit the file with options including:

| Option | Description |
| ---------------- | ----------------------------------- |
| `_active` | Enable/disable the extension. |
| `listeners` | Default listener setting. |
| `origin` | Extension file path. |
| `installOptions` | Version and metadata configuration. |
| `name` | Extension name. |
| `productName` | Extension display name. |
| `version` | Extension version. |
| `main` | Main file path. |
| `description` | Extension description. |
| `url` | Extension URL. |

```json title="~/jan/data/extensions/extensions.json"
{
"@janhq/conversational-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-conversational-extension-1.0.0.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/conversational-extension",
"productName": "Conversational",
"version": "1.0.0",
"main": "dist/index.js",
"description": "This extension enables conversations and state persistence via your filesystem",
"url": "extension://@janhq/conversational-extension/dist/index.js"
},
"@janhq/inference-anthropic-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-anthropic-extension-1.0.2.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-anthropic-extension",
"productName": "Anthropic Inference Engine",
"version": "1.0.2",
"main": "dist/index.js",
"description": "This extension enables Anthropic chat completion API calls",
"url": "extension://@janhq/inference-anthropic-extension/dist/index.js"
},
"@janhq/inference-triton-trt-llm-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-triton-trt-llm-extension-1.0.0.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-triton-trt-llm-extension",
"productName": "Triton-TRT-LLM Inference Engine",
"version": "1.0.0",
"main": "dist/index.js",
"description": "This extension enables Nvidia's TensorRT-LLM as an inference engine option",
"url": "extension://@janhq/inference-triton-trt-llm-extension/dist/index.js"
},
"@janhq/inference-mistral-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-mistral-extension-1.0.1.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-mistral-extension",
"productName": "MistralAI Inference Engine",
"version": "1.0.1",
"main": "dist/index.js",
"description": "This extension enables Mistral chat completion API calls",
"url": "extension://@janhq/inference-mistral-extension/dist/index.js"
},
"@janhq/inference-martian-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-martian-extension-1.0.1.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-martian-extension",
"productName": "Martian Inference Engine",
"version": "1.0.1",
"main": "dist/index.js",
"description": "This extension enables Martian chat completion API calls",
"url": "extension://@janhq/inference-martian-extension/dist/index.js"
},
"@janhq/inference-openrouter-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-openrouter-extension-1.0.0.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-openrouter-extension",
"productName": "OpenRouter Inference Engine",
"version": "1.0.0",
"main": "dist/index.js",
"description": "This extension enables Open Router chat completion API calls",
"url": "extension://@janhq/inference-openrouter-extension/dist/index.js"
},
"@janhq/inference-nvidia-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-nvidia-extension-1.0.1.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-nvidia-extension",
"productName": "NVIDIA NIM Inference Engine",
"version": "1.0.1",
"main": "dist/index.js",
"description": "This extension enables NVIDIA chat completion API calls",
"url": "extension://@janhq/inference-nvidia-extension/dist/index.js"
},
"@janhq/inference-groq-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-groq-extension-1.0.1.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-groq-extension",
"productName": "Groq Inference Engine",
"version": "1.0.1",
"main": "dist/index.js",
"description": "This extension enables fast Groq chat completion API calls",
"url": "extension://@janhq/inference-groq-extension/dist/index.js"
},
"@janhq/inference-openai-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-openai-extension-1.0.2.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-openai-extension",
"productName": "OpenAI Inference Engine",
"version": "1.0.2",
"main": "dist/index.js",
"description": "This extension enables OpenAI chat completion API calls",
"url": "extension://@janhq/inference-openai-extension/dist/index.js"
},
"@janhq/inference-cohere-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-cohere-extension-1.0.0.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-cohere-extension",
"productName": "Cohere Inference Engine",
"version": "1.0.0",
"main": "dist/index.js",
"description": "This extension enables Cohere chat completion API calls",
"url": "extension://@janhq/inference-cohere-extension/dist/index.js"
},
"@janhq/model-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-model-extension-1.0.33.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/model-extension",
"productName": "Model Management",
"version": "1.0.33",
"main": "dist/index.js",
"description": "Model Management Extension provides model exploration and seamless downloads",
"url": "extension://@janhq/model-extension/dist/index.js"
},
"@janhq/monitoring-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-monitoring-extension-1.0.10.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/monitoring-extension",
"productName": "System Monitoring",
"version": "1.0.10",
"main": "dist/index.js",
"description": "This extension provides system health and OS level data",
"url": "extension://@janhq/monitoring-extension/dist/index.js"
},
"@janhq/assistant-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-assistant-extension-1.0.1.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/assistant-extension",
"productName": "Jan Assistant",
"version": "1.0.1",
"main": "dist/index.js",
"description": "This extension enables assistants, including Jan, a default assistant that can call all downloaded models",
"url": "extension://@janhq/assistant-extension/dist/index.js"
},
"@janhq/tensorrt-llm-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-tensorrt-llm-extension-0.0.3.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/tensorrt-llm-extension",
"productName": "TensorRT-LLM Inference Engine",
"version": "0.0.3",
"main": "dist/index.js",
"description": "This extension enables Nvidia's TensorRT-LLM for the fastest GPU acceleration. See the [setup guide](https://jan.ai/guides/providers/tensorrt-llm/) for next steps.",
"url": "extension://@janhq/tensorrt-llm-extension/dist/index.js"
},
"@janhq/inference-cortex-extension": {
"_active": true,
"listeners": {},
"origin": "C:\\Users\\ACER\\AppData\\Local\\Programs\\jan\\resources\\app.asar.unpacked\\pre-install\\janhq-inference-cortex-extension-1.0.15.tgz",
"installOptions": {
"version": false,
"fullMetadata": true
},
"name": "@janhq/inference-cortex-extension",
"productName": "Cortex Inference Engine",
"version": "1.0.15",
"main": "dist/index.js",
"description": "This extension embeds cortex.cpp, a lightweight inference engine written in C++. See https://nitro.jan.ai.\nAdditional dependencies could be installed to run without Cuda Toolkit installation.",
"url": "extension://@janhq/inference-cortex-extension/dist/index.js"
}
}
```
15 changes: 15 additions & 0 deletions docs/src/pages/docs/extensions-settings/_meta.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,15 @@
{
"cortex": {
"title": "Cortex",
"href": "/docs/extensions-settings/cortex"
},
"model-management": {
"title": "Model Management",
"href": "/docs/extensions-settings/model-management"
},
"system-monitoring": {
"title": "System Monitoring",
"href": "/docs/extensions-settings/system-monitoring"
}
}

20 changes: 20 additions & 0 deletions docs/src/pages/docs/extensions-settings/cortex.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,20 @@
---
title: Cortex
description: Learn about Jan's default extensions and explore how to configure them.
keywords:
[
Jan,
Customizable Intelligence, LLM,
local AI,
privacy focus,
free and open source,
private and offline,
conversational AI,
no-subscription fee,
large language models,
Jan Extensions,
Extensions,
]
---


55 changes: 55 additions & 0 deletions docs/src/pages/docs/extensions-settings/model-management.mdx
Original file line number Diff line number Diff line change
@@ -0,0 +1,55 @@
---
title: Model Management
description: Learn about Jan's default extensions and explore how to configure them.
keywords:
[
Jan,
Customizable Intelligence, LLM,
local AI,
privacy focus,
free and open source,
private and offline,
conversational AI,
no-subscription fee,
large language models,
Jan Extensions,
Extensions,
]
---

import { Callout } from 'nextra/components'
import { Settings, EllipsisVertical, Plus, FolderOpen, Pencil } from 'lucide-react'

# Model Management

Configure how Jan handles model downloads and management. Access these settings through **Settings** (<Settings width={16} height={16} style={{display:"inline"}}/>) > **Extensions** > **Model Management**:

## Hugging Face Access Token

Access tokens authenticate your identity to Hugging Face Hub for model downloads.



Enter your token in the format: `hf_************************`

<Callout type="info">
Get your Hugging Face token from [Hugging Face Settings](https://huggingface.co/settings/tokens)
</Callout>

## Log Management

### Enable App Logs
Toggle to save logs locally on your computer for:
- Debugging model issues
- Crash reports
- Download troubleshooting

### Log Cleaning Interval
Set automatic log deletion interval in milliseconds:
- Default: 120000 (2 minutes)
- Controls disk space usage
- Prevents log accumulation

<Callout type="warning">
Keep your access tokens secure and never share them. Enable logs temporarily when needed for troubleshooting.
</Callout>
Loading

0 comments on commit a90287b

Please sign in to comment.