1
0

Initial commit

This commit is contained in:
2025-05-24 10:51:09 +08:00
commit e0aeb9f8bc
11 changed files with 74 additions and 0 deletions

2
.gitignore vendored Normal file
View File

@@ -0,0 +1,2 @@
/ollama-data
/webui

5
.vscode/extensions.json vendored Normal file
View File

@@ -0,0 +1,5 @@
{
"recommendations": [
"nr-codetools.localaipilot"
]
}

3
.vscode/settings.json vendored Normal file
View File

@@ -0,0 +1,3 @@
{
"localaipilot.standalone.ollamaCodeModel": "deepseek-coder:6.7b-base"
}

7
README.md Normal file
View File

@@ -0,0 +1,7 @@
# Local ollama
* https://habr.com/ru/companies/minerva_media/articles/909130/
* https://github.com/nagaraj-real/localaipilot-api/blob/main/README.md#chat-models
* https://huggingface.co/google/gemma-2b
* https://huggingface.co/deepseek-ai

4
down.sh Executable file
View File

@@ -0,0 +1,4 @@
#!/bin/bash
# https://habr.com/ru/companies/minerva_media/articles/909130/
docker stop ai-ollama ai-webui

10
ollama.code-workspace Normal file
View File

@@ -0,0 +1,10 @@
{
"folders": [
{
"path": "."
}
],
"settings": {
"localaipilot.standalone.ollamaCodeModel": "deepseek-coder:6.7b-base"
}
}

5
run_ds1.5b.sh Executable file
View File

@@ -0,0 +1,5 @@
#!/bin/bash
# docker exec -it ai-ollama ollama run llama3
docker exec -it ai-ollama ollama pull deepseek-r1:1.5b
docker exec -it ai-ollama ollama run deepseek-r1:1.5b

5
run_ds7b.sh Executable file
View File

@@ -0,0 +1,5 @@
#!/bin/bash
# docker exec -it ai-ollama ollama run llama3
docker exec -it ai-ollama ollama pull deepseek-r1:7b
docker exec -it ai-ollama ollama run deepseek-r1:7b

5
run_dscoder6.7b.sh Executable file
View File

@@ -0,0 +1,5 @@
#!/bin/bash
# docker exec -it ai-ollama ollama run llama3
docker exec -it ai-ollama ollama pull deepseek-coder:6.7b-base
docker exec -it ai-ollama ollama run deepseek-coder:6.7b-base

6
run_gemma2b.sh Executable file
View File

@@ -0,0 +1,6 @@
#!/bin/bash
# docker exec -it ai-ollama ollama run llama3
docker exec -it ai-ollama ollama pull codegemma:2b
docker exec -it ai-ollama ollama pull gemma:2b
docker exec -it ai-ollama ollama run gemma:2b

22
run_ollama.sh Executable file
View File

@@ -0,0 +1,22 @@
#!/bin/bash
docker run \
--detach \
--volume ./ollama-data:/root/.ollama \
--publish 11434:11434 \
--restart unless-stopped \
--name ai-ollama \
ollama/ollama
docker run \
--detach \
--volume ./webui:/app/backend/data \
--publish 9999:8080 \
--restart unless-stopped \
--add-host=host.docker.internal:host-gateway \
--name ai-webui \
ghcr.io/open-webui/open-webui:main
echo
echo "Ready, opening http://localhost:9999/"
echo