LM Studio Setup Guide
LM Studio is a user-friendly desktop application for running local large language models, letting you easily run open-source AI models without using command lines.
What is LM Studio?
LM Studio is a desktop application designed for running large language models locally:
- Graphical interface: Intuitive UI, no command-line knowledge needed
- Model marketplace: Built-in model search and download functionality
- Local server: One-click startup for OpenAI-compatible API service
- Cross-platform: Supports Windows, macOS, Linux
Installing LM Studio
Step 1: Download the Installer
- Visit the LM Studio website
- Click "Download"
- Choose the installer for your system:
- Windows:
.exeinstaller - macOS:
.dmgpackage - Linux:
.AppImagefile
- Windows:
Step 2: Install the Application
Windows:
Run the downloaded .exe file and follow the prompts
macOS:
Open the .dmg file and drag LM Studio to your Applications folder
Linux:
chmod +x LM_Studio-*.AppImage
./LM_Studio-*.AppImage
Downloading and Loading Models
Step 1: Search for Models
- Open LM Studio
- Click the "Search" icon on the left
- Search for models you want, such as:
llama 3.2- Meta's open-source modelqwen 2.5- Alibaba's open-source modelmistral- Efficient open-source model
Step 2: Download a Model
- Select a model from search results
- Choose an appropriate quantization version (recommended):
Q4_K_M: Balance between quality and speedQ5_K_M: Better quality, slightly slowerQ8_0: Highest quality, needs more memory
- Click "Download" to start
Step 3: Load the Model
- After download, click the "Chat" icon on the left
- Select the downloaded model in the model selector at the top
- Wait for the model to load
Starting the Local Server
To connect Chatbox to LM Studio, you need to start the local server:
Step 1: Open Server Settings
- Click the "Local Server" icon on the left (or press
Ctrl/Cmd + L) - You'll see the server control panel
Step 2: Configure the Server
- Make sure a model is selected
- Server port defaults to
1234(usually no need to change) - Click the "Start Server" button
Step 3: Confirm Server is Running
After the server starts, you'll see:
- Status shows
Running - API address displayed:
http://localhost:1234/v1
Configure in Chatbox
Step 1: Open Chatbox Settings
- Open the Chatbox app
- Click the "Settings" entry in the bottom left
- Select "AI Provider" or "Model Settings"
Step 2: Add LM Studio
- Click "Add Provider"
- Select "OpenAI API Compatible" or "LM Studio"
- Configure the following:
- API Host:
http://localhost:1234/v1 - API Key: Leave empty or enter any text (e.g.,
lm-studio)
- API Host:
Step 3: Select a Model
- Model name can be
defaultor the actual model name - Save settings
- Start chatting
Remote Connection Setup (Optional)
If you want to connect from other devices:
Security Warning: Changing
localhostto0.0.0.0exposes the server on your entire network. LM Studio has no built-in authentication by default. Only use this on trusted local networks, and never expose port1234to the public internet. Consider using a VPN or SSH tunnel for remote access.
Step 1: Configure Network Listening
- In LM Studio's "Local Server" settings
- Change
localhostto0.0.0.0 - Restart the server
Step 2: Open Firewall
Ensure port 1234 is open in your firewall (refer to Ollama guide for firewall configuration)
Step 3: Connect
In Chatbox on other devices:
- API Host:
http://your-computer-IP:1234/v1
Recommended Models
| Model | Size | Memory Required | Features |
|---|---|---|---|
| Llama 3.2 3B Q4 | ~2GB | 4GB+ | Fast, good for beginners |
| Llama 3.2 8B Q4 | ~5GB | 8GB+ | Balanced choice |
| Qwen 2.5 7B Q4 | ~4GB | 8GB+ | Excellent Chinese capability |
| Mistral 7B Q4 | ~4GB | 8GB+ | Strong reasoning |
Common Issues
Model Fails to Load
Possible causes:
- Insufficient memory
- Corrupted model file
Solutions:
- Try using a smaller model or lower quantization version
- Re-download the model
Server Won't Start
Solutions:
- Ensure a model is loaded
- Check if the port is in use
- Try a different port
Chatbox Cannot Connect
Troubleshooting:
- Confirm LM Studio server is running
- Check if API address is correct (note the
/v1suffix) - Try accessing
http://localhost:1234/v1/modelsin browser
Slow Response
Optimization tips:
- Use a smaller model
- Use higher quantization (e.g., Q4 instead of Q8)
- Close other resource-intensive programs
Usage Tips
-
Choose appropriate quantization:
- Q4_K_M is the best choice for most scenarios
- If you have enough memory, try Q5 or Q8
-
Keep the server running: When using Chatbox, make sure LM Studio server stays running
-
Try different models: LM Studio's model marketplace has rich choices, feel free to experiment
-
Monitor system resources: Watch memory and CPU/GPU usage when running models