Переложил и добавил модели по папкам, обновил README
This commit is contained in:
3
.vscode/extensions.json
vendored
3
.vscode/extensions.json
vendored
@@ -1,5 +1,6 @@
|
|||||||
{
|
{
|
||||||
"recommendations": [
|
"recommendations": [
|
||||||
"nr-codetools.localaipilot"
|
"nr-codetools.localaipilot",
|
||||||
|
"continue.continue"
|
||||||
]
|
]
|
||||||
}
|
}
|
||||||
|
|||||||
16
README.md
16
README.md
@@ -1,7 +1,17 @@
|
|||||||
# Local ollama
|
# Local ollama
|
||||||
|
|
||||||
|
1. Запустить `./run.sh`
|
||||||
|
2. Запустить скрипт из любой и поддиректорий для скачивания и запуска модели в терминале
|
||||||
|
|
||||||
|
Для настройки vscode поставить один из плагинов:
|
||||||
|
* [Continue](https://marketplace.visualstudio.com/items?itemName=Continue.continue)
|
||||||
|
* [Local AI Pilot](https://marketplace.visualstudio.com/items?itemName=nr-codetools.localaipilot)
|
||||||
|
|
||||||
|
Плагин должен соединиться с `localhost:11434` и подгрузить доступные модели из контейнера.
|
||||||
|
|
||||||
|
Есть веб-морда по адресу [localhost:9999](http://localhost:9999).
|
||||||
|
|
||||||
|
## Дополнительные материалы
|
||||||
|
|
||||||
* https://habr.com/ru/companies/minerva_media/articles/909130/
|
* https://habr.com/ru/companies/minerva_media/articles/909130/
|
||||||
* https://github.com/nagaraj-real/localaipilot-api/blob/main/README.md#chat-models
|
* https://github.com/nagaraj-real/localaipilot-api/blob/main/README.md#chat-models
|
||||||
* https://huggingface.co/google/gemma-2b
|
|
||||||
* https://huggingface.co/deepseek-ai
|
|
||||||
|
|
||||||
|
|||||||
4
codellama/34b-i
Executable file
4
codellama/34b-i
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/codellama/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run codellama:34b-instruct --verbose
|
||||||
4
codellama/7b-i
Executable file
4
codellama/7b-i
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/codellama/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run codellama:7b-instruct --verbose
|
||||||
4
deepseek-coder/1.3b
Executable file
4
deepseek-coder/1.3b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-coder/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-coder:1.3b --verbose
|
||||||
4
deepseek-coder/33b
Executable file
4
deepseek-coder/33b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-coder/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-coder:33b --verbose
|
||||||
4
deepseek-coder/6.7b
Executable file
4
deepseek-coder/6.7b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-coder/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-coder:6.7b --verbose
|
||||||
4
deepseek-r1/1.5b
Executable file
4
deepseek-r1/1.5b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-r1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-r1:1.5b --verbose
|
||||||
4
deepseek-r1/14b
Executable file
4
deepseek-r1/14b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-r1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-r1:14b --verbose
|
||||||
4
deepseek-r1/32b
Executable file
4
deepseek-r1/32b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-r1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-r1:32b --verbose
|
||||||
4
deepseek-r1/7b
Executable file
4
deepseek-r1/7b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-r1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-r1:7b --verbose
|
||||||
4
deepseek-r1/8b
Executable file
4
deepseek-r1/8b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/deepseek-r1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run deepseek-r1:8b --verbose
|
||||||
4
gemma/2b
Executable file
4
gemma/2b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/gemma/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run gemma:2b --verbose
|
||||||
4
gemma/7b
Executable file
4
gemma/7b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/gemma/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run gemma:7b --verbose
|
||||||
4
llama2/13b
Executable file
4
llama2/13b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama2/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama2:13b --verbose
|
||||||
4
llama2/7b
Executable file
4
llama2/7b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama2/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama2:7b --verbose
|
||||||
4
llama3.1/8b
Executable file
4
llama3.1/8b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama3.1/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama3.1:8b --verbose
|
||||||
4
llama3.2/1b
Executable file
4
llama3.2/1b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama3.2/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama3.2:1b --verbose
|
||||||
4
llama3.2/3b
Executable file
4
llama3.2/3b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama3.2/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama3.2:1b --verbose
|
||||||
4
llama3/8b
Executable file
4
llama3/8b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/llama3/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run llama3:8b --verbose
|
||||||
4
phi3/14b
Executable file
4
phi3/14b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/phi3/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run phi3:14b --verbose
|
||||||
4
phi3/3.8b
Executable file
4
phi3/3.8b
Executable file
@@ -0,0 +1,4 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
# https://ollama.com/library/phi3/tags
|
||||||
|
|
||||||
|
docker exec -it ai-ollama ollama run phi3:3.8b --verbose
|
||||||
@@ -20,3 +20,5 @@ docker run \
|
|||||||
echo
|
echo
|
||||||
echo "Ready, opening http://localhost:9999/"
|
echo "Ready, opening http://localhost:9999/"
|
||||||
echo
|
echo
|
||||||
|
|
||||||
|
open http://localhost:9999/
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull codellama:34b-instruct
|
|
||||||
docker exec -it ai-ollama ollama run codellama:34b-instruct --verbose
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull codellama:7b-instruct
|
|
||||||
docker exec -it ai-ollama ollama run codellama:7b-instruct --verbose
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull deepseek-r1:1.5b
|
|
||||||
docker exec -it ai-ollama ollama run deepseek-r1:1.5b --verbose
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull deepseek-r1:7b
|
|
||||||
docker exec -it ai-ollama ollama run deepseek-r1:7b --verbose
|
|
||||||
@@ -1,5 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull deepseek-coder:6.7b-base
|
|
||||||
docker exec -it ai-ollama ollama run deepseek-coder:6.7b-base --verbose
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
# docker exec -it ai-ollama ollama run llama3
|
|
||||||
docker exec -it ai-ollama ollama pull codegemma:2b
|
|
||||||
docker exec -it ai-ollama ollama pull gemma:2b
|
|
||||||
docker exec -it ai-ollama ollama run gemma:2b --verbose
|
|
||||||
Reference in New Issue
Block a user