AI Toolkit for VScode - Install Locally - Run LLMs Locally in Windows
AI Nuggets
Detailed Instructions from YouTube Video Transcript
Prerequisites
- Make sure you have Visual Studio Code (VS Code) installed.
- If you haven’t used VS Code, download it from code.visualstudio.com.
Installing AI Toolkit Extension in VS Code
- Open VS Code.
- Click on the Extensions icon (second last menu item on the left).
- In the search box, type
AI toolkit
.- Look for “AI toolkit” with the publisher “Microsoft”.
- Click on “Install”. (You may install the pre-release version as the release version might not be available yet.)
Using AI Toolkit Extension in VS Code
- After installation, click on the AI toolkit icon on the left.
- Sign in using GitHub when prompted.
- Select Microsoft Edge when asked to choose a browser.
- Log in with your GitHub username and password for authentication.
- In the terminal, an agent will start, which is used for the model playground.
Exploring Model Playground and Model Catalog
- Click on “Model Playground” in the AI toolkit sidebar.
- To download models, click on “Model Catalog” at the top left.
- Choose a model (e.g., Microsoft’s 53 mini 4K CPU) and click “Download”.
- Once downloaded, click “Load in Playground” to load the model.
Requirements for Running Models
- A GPU is recommended (e.g., 20 GB of VRAM).
- At least 32 GB of memory RAM is suggested.
Interacting with the Model
- Provide context and set parameters like temperature, top P, penalties, and response length.
- Enter your query in the text box (e.g., “What is happiness?”).
- The model will generate a response based on your system’s resources.
- You can monitor the response statistics in the terminal window.
Additional Information
- The models use ONNX (Open Neural Network Exchange) runtime for interoperability.
- You can embed the model in your code or ask it to write code for you.
Setting Up Linux on Windows with WSL (Windows Subsystem for Linux)
- Open PowerShell as an administrator.
- Type the command
WSL --install
to install WSL.Additional Tips
- Some models can run on CPU, and you can try those if you don’t have a GPU.
- You can run models remotely using cloud resources.
- Model fine-tuning will be covered in a subsequent video.
Call to Action
- Subscribe to the channel if you like the content.
- Share the video with your network.
Video Description
- The video description may contain additional information or links related to the content discussed in the video. However, without access to the actual video description, I cannot provide those details here.