Local Large Language Models and Usage of Domestic AI Compatible with OpenAI ChatGPT Interface
In video translation and dubbing software, large language models can serve as efficient translation channels, significantly improving translation quality by leveraging context.
Currently, most domestic AI interfaces are compatible with OpenAI's technology, so users can operate directly in OpenAI ChatGPT or local large language models. Local deployment using ollama is also possible.
Moonshot AI Usage
- Fill in the API address with
https://api.moonshot.cn/v1
- Fill in the SK with the
API Key
obtained from the Moonshot open platform, which can be obtained from this URL: https://platform.moonshot.cn/console/api-keys - Fill in the model text box with
moonshot-v1-8k,moonshot-v1-32k,moonshot-v1-128k
- Then select the desired model in the model selection area. Keep it after testing without problems.
DeepSeek AI Usage
- Fill in the API address with
https://api.deepseek.com/v1
- Fill in the SK with the
API Key
obtained from the Moonshot open platform, which can be obtained from this URL: https://platform.deepseek.com/api_keys - Fill in the model text box with
deepseek-chat
- Then select
deepseek-chat
in the model selection area. Keep it after testing without problems.
Zhipu AI BigModel Usage
- Fill in the API address with
https://open.bigmodel.cn/api/paas/v4/
- Fill in the SK with the
API Key
obtained from the Moonshot open platform, which can be obtained from this URL: https://www.bigmodel.cn/usercenter/apikeys - Fill in the model text box with
glm-4-plus,glm-4-0520,glm-4 ,glm-4-air,glm-4-airx,glm-4-long , glm-4-flashx ,glm-4-flash
- Then select the desired model in the model selection area. You can choose the free model
glm-4-flash
. Keep it after testing without problems.
Baichuan Intelligent AI Usage
- Fill in the API address with
https://api.baichuan-ai.com/v1
- Fill in the SK with the
API Key
obtained from the Moonshot open platform, which can be obtained from this URL: https://platform.baichuan-ai.com/console/apikey - Fill in the model text box with
Baichuan4,Baichuan3-Turbo,Baichuan3-Turbo-128k,Baichuan2-Turbo
- Then select the desired model in the model selection area. Keep it after testing without problems.
Notes:
Most AI translation channels may limit the number of requests per minute. If an error message indicates exceeding the request frequency, click "Translation Channel ↓" on the main software interface, and in the pop-up window, change the pause seconds to 10. This means waiting 10 seconds after each translation before initiating the next translation request, allowing a maximum of 6 times per minute to prevent exceeding the frequency limit.
If the selected model is not intelligent enough, especially locally deployed models limited by hardware resources, which are usually smaller and unable to accurately return translations in the required format according to instructions, there may be too many blank lines in the translation results. In this case, try using a larger model or open Menu--Tools/Options--Advanced Options--Send complete subtitle content when using AI translation, and uncheck it.
Using ollama for Local Deployment of Tongyi Qianwen Large Language Model
If you have some hands-on ability, you can also use this model for translation after local deployment. This section uses Tongyi Qianwen as an example to introduce deployment and usage.
1. Download the exe and Run Successfully
Open the URL https://ollama.com/download
Click to download. After downloading, double-click to open the installation interface and click Install
to complete.
A black or blue window will automatically pop up after completion. Enter three words ollama run qwen
and press Enter. This will automatically download the Tongyi Qianwen model.
Wait for the model download to finish. No proxy is needed, and the speed is quite fast.
After the model is automatically downloaded, it will run directly. When the progress reaches 100% and displays the "Success" character, it means the model has run successfully. This also indicates that the Tongyi Qianwen large language model installation and deployment are complete, and you can start using it happily. Isn't it super simple?
The default API address is http://localhost:11434
If the window is closed, how do you reopen it? It's also simple. Open the computer's start menu, find "Command Prompt" or "Windows PowerShell" (or directly type
Win key + q key
and input cmd to search), click to open it, and enterollama run qwen
to complete.
2. Use Directly in the Console Command Window
As shown in the figure below, when this interface is displayed, you can actually start typing text in the window to use it.
3. Of Course, This Interface Might Not Be Very User-Friendly, So Let's Get a User-Friendly UI
Open the URL https://chatboxai.app/zh and click download
Double-click after downloading and wait for the interface window to open automatically.
Click "Start Setting". In the pop-up layer, click "Ollama" in the top model and AI model provider, fill in the API domain name address http://localhost:11434
, and select Qwen:latest
in the model drop-down menu, then save and you're done.
The usage interface is displayed after saving. Use your imagination and use it freely.
4. Fill in the API into the Video Translation and Dubbing Software
- Open Menu--Settings--Local Compatible OpenAI Large Model, add a model
,qwen
in the middle text box, as shown below, and then select this model
- Fill in
http://localhost:11434
in the API URL, and fill in SK arbitrarily, for example, 1234
- Test whether it is successful. Save it if successful, and then use it.
5. Other Usable Models
Besides Tongyi Qianwen, many other models can be used. The method is just as simple. You only need three words ollama run model name
.
Open this address https://ollama.com/library to see all model names. If you want to use one, copy the name and then execute ollama run model name
.
Remember how to open the command window? Click the start menu and find Command Prompt
or Windows PowerShell
.
For example, if I want to install the openchat
model
Open Command Prompt
, enter ollama run openchat
, press Enter, and wait until "Success" is displayed.
Notes:
Most AI translation channels may limit the number of requests per minute. If an error message indicates exceeding the request frequency, click "Translation Channel ↓" on the main software interface, and in the pop-up window, change the pause seconds to 10. This means waiting 10 seconds after each translation before initiating the next translation request, allowing a maximum of 6 times per minute to prevent exceeding the frequency limit.
If the selected model is not intelligent enough, especially locally deployed models limited by hardware resources, which are usually smaller and unable to accurately return translations in the required format according to instructions, there may be too many blank lines in the translation results. In this case, try using a larger model or open Menu--Tools/Options--Advanced Options--Send complete subtitle content when using AI translation, and uncheck it.